var/home/core/zuul-output/0000755000175000017500000000000015111443701014522 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111465252015473 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006126455215111465242017711 0ustar rootrootNov 26 00:23:37 crc systemd[1]: Starting Kubernetes Kubelet... Nov 26 00:23:38 crc restorecon[4753]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 00:23:38 crc restorecon[4753]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 00:23:38 crc restorecon[4753]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 26 00:23:39 crc kubenswrapper[4766]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 26 00:23:39 crc kubenswrapper[4766]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 26 00:23:39 crc kubenswrapper[4766]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 26 00:23:39 crc kubenswrapper[4766]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 26 00:23:39 crc kubenswrapper[4766]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 26 00:23:39 crc kubenswrapper[4766]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.583155 4766 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592054 4766 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592114 4766 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592127 4766 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592136 4766 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592144 4766 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592152 4766 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592160 4766 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592169 4766 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592176 4766 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592184 4766 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592192 4766 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592199 4766 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592207 4766 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592214 4766 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592222 4766 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592230 4766 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592241 4766 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592252 4766 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592261 4766 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592270 4766 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592278 4766 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592286 4766 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592294 4766 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592302 4766 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592310 4766 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592318 4766 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592326 4766 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592333 4766 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592341 4766 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592349 4766 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592357 4766 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592364 4766 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592372 4766 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592381 4766 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592389 4766 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592416 4766 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592424 4766 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592432 4766 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592439 4766 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592449 4766 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592456 4766 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592464 4766 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592471 4766 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592479 4766 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592487 4766 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592494 4766 feature_gate.go:330] unrecognized feature gate: Example Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592502 4766 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592509 4766 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592519 4766 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592530 4766 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592539 4766 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592547 4766 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592557 4766 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592567 4766 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592575 4766 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592584 4766 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592592 4766 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592600 4766 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592608 4766 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592615 4766 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592623 4766 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592633 4766 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592643 4766 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592680 4766 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592691 4766 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592710 4766 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592727 4766 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592738 4766 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592748 4766 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592758 4766 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.592767 4766 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.592923 4766 flags.go:64] FLAG: --address="0.0.0.0" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.592942 4766 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.592956 4766 flags.go:64] FLAG: --anonymous-auth="true" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.592967 4766 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.592979 4766 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.592988 4766 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.592999 4766 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593010 4766 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593021 4766 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593031 4766 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593041 4766 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593053 4766 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593063 4766 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593072 4766 flags.go:64] FLAG: --cgroup-root="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593081 4766 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593090 4766 flags.go:64] FLAG: --client-ca-file="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593100 4766 flags.go:64] FLAG: --cloud-config="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593109 4766 flags.go:64] FLAG: --cloud-provider="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593117 4766 flags.go:64] FLAG: --cluster-dns="[]" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593129 4766 flags.go:64] FLAG: --cluster-domain="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593138 4766 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593147 4766 flags.go:64] FLAG: --config-dir="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593156 4766 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593165 4766 flags.go:64] FLAG: --container-log-max-files="5" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593176 4766 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593186 4766 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593195 4766 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593203 4766 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593212 4766 flags.go:64] FLAG: --contention-profiling="false" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593221 4766 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593231 4766 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593240 4766 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593249 4766 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593260 4766 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593269 4766 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593278 4766 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593287 4766 flags.go:64] FLAG: --enable-load-reader="false" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593296 4766 flags.go:64] FLAG: --enable-server="true" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593305 4766 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593316 4766 flags.go:64] FLAG: --event-burst="100" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593325 4766 flags.go:64] FLAG: --event-qps="50" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593334 4766 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593343 4766 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593352 4766 flags.go:64] FLAG: --eviction-hard="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593362 4766 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593371 4766 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593381 4766 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593391 4766 flags.go:64] FLAG: --eviction-soft="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593401 4766 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593409 4766 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593418 4766 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593427 4766 flags.go:64] FLAG: --experimental-mounter-path="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593436 4766 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593445 4766 flags.go:64] FLAG: --fail-swap-on="true" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593453 4766 flags.go:64] FLAG: --feature-gates="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593463 4766 flags.go:64] FLAG: --file-check-frequency="20s" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593473 4766 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593482 4766 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593491 4766 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593500 4766 flags.go:64] FLAG: --healthz-port="10248" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593509 4766 flags.go:64] FLAG: --help="false" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593518 4766 flags.go:64] FLAG: --hostname-override="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593527 4766 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593536 4766 flags.go:64] FLAG: --http-check-frequency="20s" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593545 4766 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593555 4766 flags.go:64] FLAG: --image-credential-provider-config="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593563 4766 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593572 4766 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593581 4766 flags.go:64] FLAG: --image-service-endpoint="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593590 4766 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593598 4766 flags.go:64] FLAG: --kube-api-burst="100" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593607 4766 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593617 4766 flags.go:64] FLAG: --kube-api-qps="50" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593626 4766 flags.go:64] FLAG: --kube-reserved="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593635 4766 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593643 4766 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593703 4766 flags.go:64] FLAG: --kubelet-cgroups="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593725 4766 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593738 4766 flags.go:64] FLAG: --lock-file="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593747 4766 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593757 4766 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593767 4766 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593793 4766 flags.go:64] FLAG: --log-json-split-stream="false" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593805 4766 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593814 4766 flags.go:64] FLAG: --log-text-split-stream="false" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593824 4766 flags.go:64] FLAG: --logging-format="text" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593833 4766 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593845 4766 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593866 4766 flags.go:64] FLAG: --manifest-url="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593884 4766 flags.go:64] FLAG: --manifest-url-header="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593900 4766 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593912 4766 flags.go:64] FLAG: --max-open-files="1000000" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593926 4766 flags.go:64] FLAG: --max-pods="110" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593937 4766 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593949 4766 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593960 4766 flags.go:64] FLAG: --memory-manager-policy="None" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593971 4766 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593983 4766 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.593994 4766 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594005 4766 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594032 4766 flags.go:64] FLAG: --node-status-max-images="50" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594043 4766 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594053 4766 flags.go:64] FLAG: --oom-score-adj="-999" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594062 4766 flags.go:64] FLAG: --pod-cidr="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594071 4766 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594085 4766 flags.go:64] FLAG: --pod-manifest-path="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594093 4766 flags.go:64] FLAG: --pod-max-pids="-1" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594104 4766 flags.go:64] FLAG: --pods-per-core="0" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594113 4766 flags.go:64] FLAG: --port="10250" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594122 4766 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594131 4766 flags.go:64] FLAG: --provider-id="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594140 4766 flags.go:64] FLAG: --qos-reserved="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594149 4766 flags.go:64] FLAG: --read-only-port="10255" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594158 4766 flags.go:64] FLAG: --register-node="true" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594168 4766 flags.go:64] FLAG: --register-schedulable="true" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594176 4766 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594190 4766 flags.go:64] FLAG: --registry-burst="10" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594198 4766 flags.go:64] FLAG: --registry-qps="5" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594207 4766 flags.go:64] FLAG: --reserved-cpus="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594220 4766 flags.go:64] FLAG: --reserved-memory="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594232 4766 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594242 4766 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594251 4766 flags.go:64] FLAG: --rotate-certificates="false" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594259 4766 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594268 4766 flags.go:64] FLAG: --runonce="false" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594277 4766 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594286 4766 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594295 4766 flags.go:64] FLAG: --seccomp-default="false" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594304 4766 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594313 4766 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594323 4766 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594333 4766 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594344 4766 flags.go:64] FLAG: --storage-driver-password="root" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594355 4766 flags.go:64] FLAG: --storage-driver-secure="false" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594366 4766 flags.go:64] FLAG: --storage-driver-table="stats" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594378 4766 flags.go:64] FLAG: --storage-driver-user="root" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594387 4766 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594397 4766 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594406 4766 flags.go:64] FLAG: --system-cgroups="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594415 4766 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594430 4766 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594438 4766 flags.go:64] FLAG: --tls-cert-file="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594447 4766 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594460 4766 flags.go:64] FLAG: --tls-min-version="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594469 4766 flags.go:64] FLAG: --tls-private-key-file="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594479 4766 flags.go:64] FLAG: --topology-manager-policy="none" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594488 4766 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594497 4766 flags.go:64] FLAG: --topology-manager-scope="container" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594506 4766 flags.go:64] FLAG: --v="2" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594517 4766 flags.go:64] FLAG: --version="false" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594528 4766 flags.go:64] FLAG: --vmodule="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594539 4766 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.594549 4766 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594809 4766 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594821 4766 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594831 4766 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594842 4766 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594853 4766 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594862 4766 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594869 4766 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594878 4766 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594886 4766 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594894 4766 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594903 4766 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594911 4766 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594919 4766 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594927 4766 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594935 4766 feature_gate.go:330] unrecognized feature gate: Example Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594943 4766 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594951 4766 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594961 4766 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594971 4766 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594980 4766 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594988 4766 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.594996 4766 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595004 4766 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595012 4766 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595025 4766 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595033 4766 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595041 4766 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595051 4766 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595061 4766 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595071 4766 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595080 4766 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595088 4766 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595096 4766 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595104 4766 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595113 4766 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595121 4766 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595132 4766 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595140 4766 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595149 4766 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595157 4766 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595165 4766 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595173 4766 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595181 4766 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595189 4766 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595196 4766 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595205 4766 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595212 4766 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595220 4766 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595227 4766 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595236 4766 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595244 4766 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595252 4766 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595260 4766 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595267 4766 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595275 4766 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595284 4766 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595295 4766 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595303 4766 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595311 4766 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595319 4766 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595328 4766 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595336 4766 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595344 4766 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595352 4766 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595361 4766 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595370 4766 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595378 4766 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595387 4766 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595398 4766 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595407 4766 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.595415 4766 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.595441 4766 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.606299 4766 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.606335 4766 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606417 4766 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606427 4766 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606434 4766 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606438 4766 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606443 4766 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606448 4766 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606452 4766 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606457 4766 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606462 4766 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606466 4766 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606473 4766 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606479 4766 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606485 4766 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606491 4766 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606497 4766 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606503 4766 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606509 4766 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606517 4766 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606522 4766 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606527 4766 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606532 4766 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606537 4766 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606541 4766 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606546 4766 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606551 4766 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606555 4766 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606560 4766 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606565 4766 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606601 4766 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606607 4766 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606611 4766 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606616 4766 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606620 4766 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606625 4766 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606631 4766 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606636 4766 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606641 4766 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606645 4766 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606663 4766 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606668 4766 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606672 4766 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606677 4766 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606682 4766 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606688 4766 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606694 4766 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606700 4766 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606706 4766 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606711 4766 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606715 4766 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606721 4766 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606725 4766 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606730 4766 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606734 4766 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606739 4766 feature_gate.go:330] unrecognized feature gate: Example Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606744 4766 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606749 4766 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606753 4766 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606758 4766 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606762 4766 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606769 4766 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606774 4766 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606779 4766 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606785 4766 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606790 4766 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606795 4766 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606799 4766 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606804 4766 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606808 4766 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606813 4766 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606818 4766 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.606833 4766 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.606841 4766 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607026 4766 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607040 4766 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607046 4766 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607053 4766 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607061 4766 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607067 4766 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607072 4766 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607078 4766 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607084 4766 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607089 4766 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607093 4766 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607098 4766 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607103 4766 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607108 4766 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607113 4766 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607118 4766 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607124 4766 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607129 4766 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607134 4766 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607139 4766 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607143 4766 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607148 4766 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607152 4766 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607157 4766 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607161 4766 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607166 4766 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607171 4766 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607176 4766 feature_gate.go:330] unrecognized feature gate: Example Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607181 4766 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607187 4766 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607192 4766 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607197 4766 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607202 4766 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607207 4766 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607214 4766 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607219 4766 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607225 4766 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607230 4766 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607234 4766 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607239 4766 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607244 4766 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607249 4766 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607253 4766 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607258 4766 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607262 4766 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607267 4766 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607271 4766 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607276 4766 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607280 4766 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607285 4766 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607290 4766 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607295 4766 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607301 4766 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607305 4766 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607310 4766 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607314 4766 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607318 4766 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607323 4766 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607327 4766 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607332 4766 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607337 4766 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607341 4766 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607346 4766 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607350 4766 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607355 4766 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607360 4766 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607365 4766 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607370 4766 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607375 4766 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607380 4766 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.607386 4766 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.607394 4766 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.607608 4766 server.go:940] "Client rotation is on, will bootstrap in background" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.611910 4766 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.611997 4766 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.613631 4766 server.go:997] "Starting client certificate rotation" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.613674 4766 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.614709 4766 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-18 04:59:58.030278146 +0000 UTC Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.614769 4766 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 532h36m18.41551095s for next certificate rotation Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.646274 4766 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.649178 4766 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.671525 4766 log.go:25] "Validated CRI v1 runtime API" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.715319 4766 log.go:25] "Validated CRI v1 image API" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.718476 4766 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.725012 4766 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-26-00-19-15-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.725066 4766 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:41 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.741376 4766 manager.go:217] Machine: {Timestamp:2025-11-26 00:23:39.738634062 +0000 UTC m=+0.587404512 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:4cba354f-139e-43ce-b547-af06ad27febb BootID:e5fef464-cee5-4575-b7c5-a728a420eb4a Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:41 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:f3:ba:f3 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:f3:ba:f3 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:ba:95:90 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:10:d3:9a Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:9a:27:f0 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:5d:63:24 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:87:8d:26 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:aa:be:b6:49:74:0c Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:2e:5f:a3:a8:eb:0f Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.741624 4766 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.741779 4766 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.743936 4766 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.744108 4766 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.744143 4766 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.744339 4766 topology_manager.go:138] "Creating topology manager with none policy" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.744352 4766 container_manager_linux.go:303] "Creating device plugin manager" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.744901 4766 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.744932 4766 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.745102 4766 state_mem.go:36] "Initialized new in-memory state store" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.745443 4766 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.748583 4766 kubelet.go:418] "Attempting to sync node with API server" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.748605 4766 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.748619 4766 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.748631 4766 kubelet.go:324] "Adding apiserver pod source" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.748645 4766 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.752388 4766 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.753414 4766 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.755778 4766 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Nov 26 00:23:39 crc kubenswrapper[4766]: E1126 00:23:39.755871 4766 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.755864 4766 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Nov 26 00:23:39 crc kubenswrapper[4766]: E1126 00:23:39.755933 4766 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.755976 4766 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.757932 4766 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.757974 4766 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.757988 4766 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.758005 4766 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.758029 4766 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.758042 4766 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.758063 4766 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.758086 4766 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.758100 4766 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.758114 4766 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.758153 4766 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.758167 4766 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.760893 4766 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.761872 4766 server.go:1280] "Started kubelet" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.762096 4766 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.762824 4766 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 26 00:23:39 crc systemd[1]: Started Kubernetes Kubelet. Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.765296 4766 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.766762 4766 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.768472 4766 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.768508 4766 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.768769 4766 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 05:11:08.650866543 +0000 UTC Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.768833 4766 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 26 00:23:39 crc kubenswrapper[4766]: E1126 00:23:39.768834 4766 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.768862 4766 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.768836 4766 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 412h47m28.882035658s for next certificate rotation Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.768849 4766 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.769807 4766 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Nov 26 00:23:39 crc kubenswrapper[4766]: E1126 00:23:39.769928 4766 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.770234 4766 factory.go:55] Registering systemd factory Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.770261 4766 factory.go:221] Registration of the systemd container factory successfully Nov 26 00:23:39 crc kubenswrapper[4766]: E1126 00:23:39.770484 4766 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="200ms" Nov 26 00:23:39 crc kubenswrapper[4766]: E1126 00:23:39.770207 4766 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.32:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b66b80f444146 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-26 00:23:39.761828166 +0000 UTC m=+0.610598636,LastTimestamp:2025-11-26 00:23:39.761828166 +0000 UTC m=+0.610598636,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.773484 4766 factory.go:153] Registering CRI-O factory Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.773502 4766 factory.go:221] Registration of the crio container factory successfully Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.773565 4766 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.773588 4766 factory.go:103] Registering Raw factory Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.773601 4766 manager.go:1196] Started watching for new ooms in manager Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.774565 4766 server.go:460] "Adding debug handlers to kubelet server" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.775501 4766 manager.go:319] Starting recovery of all containers Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.780933 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.780994 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781016 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781033 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781049 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781066 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781082 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781099 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781117 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781135 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781150 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781167 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781184 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781203 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781220 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781236 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781250 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781276 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781293 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781310 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781326 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781346 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781362 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781381 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781397 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781414 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781437 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781455 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781471 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781486 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781503 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781519 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781536 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781551 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781566 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781586 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781605 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781623 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781637 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781674 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781691 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781712 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781731 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781747 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781763 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781778 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781795 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781811 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781828 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781846 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781863 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781878 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781902 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781920 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781938 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781955 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781973 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.781996 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.782013 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.782030 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.782090 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784550 4766 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784603 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784626 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784645 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784682 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784703 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784718 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784733 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784748 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784762 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784776 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784790 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784808 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784822 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784837 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784852 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784869 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784884 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784899 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784916 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784932 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784948 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784963 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784979 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.784995 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785010 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785025 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785040 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785057 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785073 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785089 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785104 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785120 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785136 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785151 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785168 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785184 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785200 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785233 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785252 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785268 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785287 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785302 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785318 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785342 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785363 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785382 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785401 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785418 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785437 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785453 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785478 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785494 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785511 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785528 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785543 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785557 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785571 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785587 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785604 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785620 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785635 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785668 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785686 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785701 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785715 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785730 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785748 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785765 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785788 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785806 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785823 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785837 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785853 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785869 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785891 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785910 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785927 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785942 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785961 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785977 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.785992 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786009 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786027 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786043 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786058 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786073 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786088 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786103 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786117 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786132 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786146 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786164 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786180 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786247 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786269 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786287 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786304 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786321 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786338 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786357 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786394 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786413 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786429 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786444 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786469 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786487 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786503 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786521 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786574 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786592 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786609 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786625 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786640 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786677 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786698 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786716 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786737 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786755 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786774 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786791 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786808 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786824 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786844 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786863 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786880 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786895 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786911 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786929 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786945 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786961 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786978 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.786994 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.787011 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.787030 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.787046 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.787064 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.787079 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.787107 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.787123 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.787142 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.787157 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.787174 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.787195 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.787213 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.787229 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.787247 4766 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.787265 4766 reconstruct.go:97] "Volume reconstruction finished" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.787275 4766 reconciler.go:26] "Reconciler: start to sync state" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.793736 4766 manager.go:324] Recovery completed Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.803832 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.805460 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.805582 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.806802 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.809386 4766 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.809408 4766 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.809431 4766 state_mem.go:36] "Initialized new in-memory state store" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.823615 4766 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.825363 4766 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.825415 4766 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.825463 4766 kubelet.go:2335] "Starting kubelet main sync loop" Nov 26 00:23:39 crc kubenswrapper[4766]: E1126 00:23:39.825529 4766 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 26 00:23:39 crc kubenswrapper[4766]: W1126 00:23:39.829630 4766 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Nov 26 00:23:39 crc kubenswrapper[4766]: E1126 00:23:39.829741 4766 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.831039 4766 policy_none.go:49] "None policy: Start" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.831832 4766 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.831860 4766 state_mem.go:35] "Initializing new in-memory state store" Nov 26 00:23:39 crc kubenswrapper[4766]: E1126 00:23:39.869335 4766 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.881063 4766 manager.go:334] "Starting Device Plugin manager" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.881135 4766 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.881148 4766 server.go:79] "Starting device plugin registration server" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.881553 4766 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.881587 4766 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.881793 4766 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.881873 4766 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.881882 4766 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 26 00:23:39 crc kubenswrapper[4766]: E1126 00:23:39.887673 4766 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.926273 4766 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.926592 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.928074 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.928152 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.928177 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.928452 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.928619 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.928691 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.929845 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.929887 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.929900 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.930006 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.930050 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.930074 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.930239 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.930382 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.930436 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.931837 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.931876 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.931892 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.931874 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.932028 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.932039 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.932052 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.932233 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.932286 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.933256 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.933298 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.933314 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.933410 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.933383 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.933462 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.933526 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.933685 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.933727 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.934610 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.934610 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.934640 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.934719 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.934688 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.934789 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.934968 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.935012 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.936216 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.936245 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.936256 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:39 crc kubenswrapper[4766]: E1126 00:23:39.971072 4766 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="400ms" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.982069 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.983321 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.983375 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.983393 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.983424 4766 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 26 00:23:39 crc kubenswrapper[4766]: E1126 00:23:39.983950 4766 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.32:6443: connect: connection refused" node="crc" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.989276 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.989325 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.989355 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.989377 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.989410 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.989478 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.989537 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.989564 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.989587 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.989604 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.989623 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.989677 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.989713 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.989737 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 00:23:39 crc kubenswrapper[4766]: I1126 00:23:39.989759 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.090924 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.091026 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.091095 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.091145 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.091258 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.091302 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.091314 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.091412 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.091343 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.091465 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.091482 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.091492 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.091566 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.091580 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.091619 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.091600 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.091632 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.091569 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.091499 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.091484 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.091860 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.092020 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.092130 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.092067 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.092141 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.092224 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.092266 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.092290 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.092298 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.092390 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.184993 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.186757 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.186808 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.186829 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.186866 4766 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 26 00:23:40 crc kubenswrapper[4766]: E1126 00:23:40.187401 4766 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.32:6443: connect: connection refused" node="crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.258026 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.264267 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.279597 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.285911 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.290821 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 26 00:23:40 crc kubenswrapper[4766]: W1126 00:23:40.305559 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-2adcafe2d0d131f79ef842aadb0e436b53c1aafc6b843cb2d08ed8004d2318d4 WatchSource:0}: Error finding container 2adcafe2d0d131f79ef842aadb0e436b53c1aafc6b843cb2d08ed8004d2318d4: Status 404 returned error can't find the container with id 2adcafe2d0d131f79ef842aadb0e436b53c1aafc6b843cb2d08ed8004d2318d4 Nov 26 00:23:40 crc kubenswrapper[4766]: W1126 00:23:40.311113 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-e377e43fab7ec3ebdb8e6899c24cecdae0bcfff419109af53fe4be1d8807286d WatchSource:0}: Error finding container e377e43fab7ec3ebdb8e6899c24cecdae0bcfff419109af53fe4be1d8807286d: Status 404 returned error can't find the container with id e377e43fab7ec3ebdb8e6899c24cecdae0bcfff419109af53fe4be1d8807286d Nov 26 00:23:40 crc kubenswrapper[4766]: W1126 00:23:40.318124 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-47c2433ce9461ce907252c6ec19c3df2773cf216e17159e23c2b375c63a88279 WatchSource:0}: Error finding container 47c2433ce9461ce907252c6ec19c3df2773cf216e17159e23c2b375c63a88279: Status 404 returned error can't find the container with id 47c2433ce9461ce907252c6ec19c3df2773cf216e17159e23c2b375c63a88279 Nov 26 00:23:40 crc kubenswrapper[4766]: W1126 00:23:40.319048 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-9ef7f52e9892f09e6e83512cf48543cbd128c63508a866fd4852676b15bcab5b WatchSource:0}: Error finding container 9ef7f52e9892f09e6e83512cf48543cbd128c63508a866fd4852676b15bcab5b: Status 404 returned error can't find the container with id 9ef7f52e9892f09e6e83512cf48543cbd128c63508a866fd4852676b15bcab5b Nov 26 00:23:40 crc kubenswrapper[4766]: W1126 00:23:40.325228 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-49c4ea39c93c19314f3e097f98141334162df516fe2758357f32de72c2311293 WatchSource:0}: Error finding container 49c4ea39c93c19314f3e097f98141334162df516fe2758357f32de72c2311293: Status 404 returned error can't find the container with id 49c4ea39c93c19314f3e097f98141334162df516fe2758357f32de72c2311293 Nov 26 00:23:40 crc kubenswrapper[4766]: E1126 00:23:40.372354 4766 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="800ms" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.588532 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.589789 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.589836 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.589846 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.589876 4766 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 26 00:23:40 crc kubenswrapper[4766]: E1126 00:23:40.590389 4766 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.32:6443: connect: connection refused" node="crc" Nov 26 00:23:40 crc kubenswrapper[4766]: W1126 00:23:40.700410 4766 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Nov 26 00:23:40 crc kubenswrapper[4766]: E1126 00:23:40.700538 4766 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.763001 4766 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.829762 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"9ef7f52e9892f09e6e83512cf48543cbd128c63508a866fd4852676b15bcab5b"} Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.830905 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"47c2433ce9461ce907252c6ec19c3df2773cf216e17159e23c2b375c63a88279"} Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.831843 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e377e43fab7ec3ebdb8e6899c24cecdae0bcfff419109af53fe4be1d8807286d"} Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.832798 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2adcafe2d0d131f79ef842aadb0e436b53c1aafc6b843cb2d08ed8004d2318d4"} Nov 26 00:23:40 crc kubenswrapper[4766]: I1126 00:23:40.833531 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"49c4ea39c93c19314f3e097f98141334162df516fe2758357f32de72c2311293"} Nov 26 00:23:41 crc kubenswrapper[4766]: W1126 00:23:41.148375 4766 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Nov 26 00:23:41 crc kubenswrapper[4766]: E1126 00:23:41.148976 4766 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Nov 26 00:23:41 crc kubenswrapper[4766]: E1126 00:23:41.173345 4766 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="1.6s" Nov 26 00:23:41 crc kubenswrapper[4766]: W1126 00:23:41.280138 4766 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Nov 26 00:23:41 crc kubenswrapper[4766]: E1126 00:23:41.280282 4766 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Nov 26 00:23:41 crc kubenswrapper[4766]: W1126 00:23:41.303446 4766 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Nov 26 00:23:41 crc kubenswrapper[4766]: E1126 00:23:41.303609 4766 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.391368 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.392815 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.392861 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.392872 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.392902 4766 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 26 00:23:41 crc kubenswrapper[4766]: E1126 00:23:41.393363 4766 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.32:6443: connect: connection refused" node="crc" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.763474 4766 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.837594 4766 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72" exitCode=0 Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.837727 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72"} Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.837779 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.839047 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.839096 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.839112 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.840081 4766 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a" exitCode=0 Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.840125 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a"} Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.840258 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.840327 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.841560 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.841576 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.841624 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.841589 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.841681 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.841730 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.842311 4766 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="0624ff399d744217874fd106831a1e945b55b700ad96cf3fc6cc6edf4f42ba0a" exitCode=0 Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.842370 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.842397 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"0624ff399d744217874fd106831a1e945b55b700ad96cf3fc6cc6edf4f42ba0a"} Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.843220 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.843244 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.843253 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.845028 4766 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="5e2b174d3a0b70785380dfbac2eb1062d84fafedff25f8322cc25aea9c362b6e" exitCode=0 Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.845072 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"5e2b174d3a0b70785380dfbac2eb1062d84fafedff25f8322cc25aea9c362b6e"} Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.845109 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.845808 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.845840 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.845850 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.848453 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0"} Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.848514 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e"} Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.848531 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e"} Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.848543 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e"} Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.848570 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.850069 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.850116 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:41 crc kubenswrapper[4766]: I1126 00:23:41.850203 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:42 crc kubenswrapper[4766]: W1126 00:23:42.630010 4766 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Nov 26 00:23:42 crc kubenswrapper[4766]: E1126 00:23:42.630098 4766 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.763388 4766 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Nov 26 00:23:42 crc kubenswrapper[4766]: E1126 00:23:42.774464 4766 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="3.2s" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.852952 4766 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699" exitCode=0 Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.853033 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699"} Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.853076 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.854027 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.854058 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.854069 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.854799 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"a577316cc6eba55a72b6af823345750b50ca7cf138ec4393039ead51d70089b6"} Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.854876 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.855674 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.855718 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.855731 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.857213 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b2dd6961899214f94684d3cb24a3c75f9870bf0fc667ef4e2b5c25fdc4b0e5da"} Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.857239 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a2293d0b6924d276777e52994713be62a0f98ff846ecee3ed7f597835e18a8dc"} Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.857254 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d01950717656ec2766fa80a35e2f7a9136409c2331979c67bffa25b965864be0"} Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.857265 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.858028 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.858062 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.858072 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.862684 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115"} Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.862723 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5"} Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.862734 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319"} Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.862744 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36"} Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.862752 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186"} Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.862755 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.862768 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.863628 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.863664 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.863675 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.863687 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.863687 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.863729 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:42 crc kubenswrapper[4766]: W1126 00:23:42.884251 4766 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Nov 26 00:23:42 crc kubenswrapper[4766]: E1126 00:23:42.884318 4766 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.993907 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.995217 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.995300 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.995327 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:42 crc kubenswrapper[4766]: I1126 00:23:42.995374 4766 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 26 00:23:42 crc kubenswrapper[4766]: E1126 00:23:42.996063 4766 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.32:6443: connect: connection refused" node="crc" Nov 26 00:23:43 crc kubenswrapper[4766]: I1126 00:23:43.024302 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 00:23:43 crc kubenswrapper[4766]: I1126 00:23:43.869444 4766 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379" exitCode=0 Nov 26 00:23:43 crc kubenswrapper[4766]: I1126 00:23:43.869587 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:43 crc kubenswrapper[4766]: I1126 00:23:43.869680 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:43 crc kubenswrapper[4766]: I1126 00:23:43.869640 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379"} Nov 26 00:23:43 crc kubenswrapper[4766]: I1126 00:23:43.869740 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:43 crc kubenswrapper[4766]: I1126 00:23:43.869860 4766 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 00:23:43 crc kubenswrapper[4766]: I1126 00:23:43.869920 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:43 crc kubenswrapper[4766]: I1126 00:23:43.871694 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:43 crc kubenswrapper[4766]: I1126 00:23:43.871729 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:43 crc kubenswrapper[4766]: I1126 00:23:43.871742 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:43 crc kubenswrapper[4766]: I1126 00:23:43.871766 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:43 crc kubenswrapper[4766]: I1126 00:23:43.871804 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:43 crc kubenswrapper[4766]: I1126 00:23:43.871823 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:43 crc kubenswrapper[4766]: I1126 00:23:43.871945 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:43 crc kubenswrapper[4766]: I1126 00:23:43.871970 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:43 crc kubenswrapper[4766]: I1126 00:23:43.871991 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:43 crc kubenswrapper[4766]: I1126 00:23:43.872089 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:43 crc kubenswrapper[4766]: I1126 00:23:43.872116 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:43 crc kubenswrapper[4766]: I1126 00:23:43.872136 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:43 crc kubenswrapper[4766]: I1126 00:23:43.903244 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.686183 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.686368 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.687696 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.687739 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.687752 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.877291 4766 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.877335 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.877848 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.877857 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8"} Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.877889 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99"} Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.877903 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72"} Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.877914 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41"} Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.877926 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3"} Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.877950 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.878109 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.878139 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.878150 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.878533 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.878552 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.878564 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.878633 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.878670 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.878681 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:44 crc kubenswrapper[4766]: I1126 00:23:44.889142 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 26 00:23:45 crc kubenswrapper[4766]: I1126 00:23:45.234744 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:23:45 crc kubenswrapper[4766]: I1126 00:23:45.879358 4766 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 00:23:45 crc kubenswrapper[4766]: I1126 00:23:45.879701 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:45 crc kubenswrapper[4766]: I1126 00:23:45.879438 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:45 crc kubenswrapper[4766]: I1126 00:23:45.880446 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:45 crc kubenswrapper[4766]: I1126 00:23:45.880466 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:45 crc kubenswrapper[4766]: I1126 00:23:45.880482 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:45 crc kubenswrapper[4766]: I1126 00:23:45.880837 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:45 crc kubenswrapper[4766]: I1126 00:23:45.880890 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:45 crc kubenswrapper[4766]: I1126 00:23:45.880907 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:46 crc kubenswrapper[4766]: I1126 00:23:46.196480 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:46 crc kubenswrapper[4766]: I1126 00:23:46.197844 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:46 crc kubenswrapper[4766]: I1126 00:23:46.197879 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:46 crc kubenswrapper[4766]: I1126 00:23:46.197889 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:46 crc kubenswrapper[4766]: I1126 00:23:46.197917 4766 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 26 00:23:46 crc kubenswrapper[4766]: I1126 00:23:46.723868 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 00:23:46 crc kubenswrapper[4766]: I1126 00:23:46.724060 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:46 crc kubenswrapper[4766]: I1126 00:23:46.725277 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:46 crc kubenswrapper[4766]: I1126 00:23:46.725372 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:46 crc kubenswrapper[4766]: I1126 00:23:46.725396 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:46 crc kubenswrapper[4766]: I1126 00:23:46.732896 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 00:23:46 crc kubenswrapper[4766]: I1126 00:23:46.881354 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:46 crc kubenswrapper[4766]: I1126 00:23:46.881412 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:46 crc kubenswrapper[4766]: I1126 00:23:46.882495 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:46 crc kubenswrapper[4766]: I1126 00:23:46.882548 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:46 crc kubenswrapper[4766]: I1126 00:23:46.882571 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:46 crc kubenswrapper[4766]: I1126 00:23:46.882587 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:46 crc kubenswrapper[4766]: I1126 00:23:46.882610 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:46 crc kubenswrapper[4766]: I1126 00:23:46.882621 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:47 crc kubenswrapper[4766]: I1126 00:23:47.441806 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:23:47 crc kubenswrapper[4766]: I1126 00:23:47.442119 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:47 crc kubenswrapper[4766]: I1126 00:23:47.443930 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:47 crc kubenswrapper[4766]: I1126 00:23:47.443969 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:47 crc kubenswrapper[4766]: I1126 00:23:47.443978 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:48 crc kubenswrapper[4766]: I1126 00:23:48.288278 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 00:23:48 crc kubenswrapper[4766]: I1126 00:23:48.288525 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:48 crc kubenswrapper[4766]: I1126 00:23:48.289989 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:48 crc kubenswrapper[4766]: I1126 00:23:48.290025 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:48 crc kubenswrapper[4766]: I1126 00:23:48.290038 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:48 crc kubenswrapper[4766]: I1126 00:23:48.668736 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 26 00:23:48 crc kubenswrapper[4766]: I1126 00:23:48.668938 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:48 crc kubenswrapper[4766]: I1126 00:23:48.670158 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:48 crc kubenswrapper[4766]: I1126 00:23:48.670191 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:48 crc kubenswrapper[4766]: I1126 00:23:48.670200 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:49 crc kubenswrapper[4766]: I1126 00:23:49.140129 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 00:23:49 crc kubenswrapper[4766]: I1126 00:23:49.140335 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:49 crc kubenswrapper[4766]: I1126 00:23:49.141524 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:49 crc kubenswrapper[4766]: I1126 00:23:49.141591 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:49 crc kubenswrapper[4766]: I1126 00:23:49.141618 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:49 crc kubenswrapper[4766]: E1126 00:23:49.888061 4766 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 26 00:23:52 crc kubenswrapper[4766]: I1126 00:23:52.141102 4766 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 26 00:23:52 crc kubenswrapper[4766]: I1126 00:23:52.141256 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 26 00:23:53 crc kubenswrapper[4766]: I1126 00:23:53.486704 4766 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 26 00:23:53 crc kubenswrapper[4766]: I1126 00:23:53.486759 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 26 00:23:53 crc kubenswrapper[4766]: I1126 00:23:53.494277 4766 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 26 00:23:53 crc kubenswrapper[4766]: I1126 00:23:53.494348 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 26 00:23:55 crc kubenswrapper[4766]: I1126 00:23:55.241746 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:23:55 crc kubenswrapper[4766]: I1126 00:23:55.242039 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:55 crc kubenswrapper[4766]: I1126 00:23:55.243821 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:55 crc kubenswrapper[4766]: I1126 00:23:55.243886 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:55 crc kubenswrapper[4766]: I1126 00:23:55.243907 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:55 crc kubenswrapper[4766]: I1126 00:23:55.248259 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:23:55 crc kubenswrapper[4766]: I1126 00:23:55.902297 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:55 crc kubenswrapper[4766]: I1126 00:23:55.903518 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:55 crc kubenswrapper[4766]: I1126 00:23:55.903564 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:55 crc kubenswrapper[4766]: I1126 00:23:55.903583 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.294511 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.294800 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.296367 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.296456 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.296481 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:23:58 crc kubenswrapper[4766]: E1126 00:23:58.482184 4766 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.485682 4766 trace.go:236] Trace[893015531]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (26-Nov-2025 00:23:44.333) (total time: 14151ms): Nov 26 00:23:58 crc kubenswrapper[4766]: Trace[893015531]: ---"Objects listed" error: 14151ms (00:23:58.485) Nov 26 00:23:58 crc kubenswrapper[4766]: Trace[893015531]: [14.151810158s] [14.151810158s] END Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.486369 4766 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.487414 4766 trace.go:236] Trace[1244614778]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (26-Nov-2025 00:23:46.389) (total time: 12097ms): Nov 26 00:23:58 crc kubenswrapper[4766]: Trace[1244614778]: ---"Objects listed" error: 12097ms (00:23:58.487) Nov 26 00:23:58 crc kubenswrapper[4766]: Trace[1244614778]: [12.097466192s] [12.097466192s] END Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.487472 4766 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.487750 4766 trace.go:236] Trace[1896161913]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (26-Nov-2025 00:23:43.817) (total time: 14670ms): Nov 26 00:23:58 crc kubenswrapper[4766]: Trace[1896161913]: ---"Objects listed" error: 14670ms (00:23:58.487) Nov 26 00:23:58 crc kubenswrapper[4766]: Trace[1896161913]: [14.670620307s] [14.670620307s] END Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.487767 4766 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.487752 4766 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.491412 4766 trace.go:236] Trace[1740469102]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (26-Nov-2025 00:23:46.639) (total time: 11851ms): Nov 26 00:23:58 crc kubenswrapper[4766]: Trace[1740469102]: ---"Objects listed" error: 11850ms (00:23:58.489) Nov 26 00:23:58 crc kubenswrapper[4766]: Trace[1740469102]: [11.851919373s] [11.851919373s] END Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.492043 4766 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 26 00:23:58 crc kubenswrapper[4766]: E1126 00:23:58.492146 4766 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.551051 4766 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46876->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.551051 4766 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46868->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.551119 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46876->192.168.126.11:17697: read: connection reset by peer" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.551144 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46868->192.168.126.11:17697: read: connection reset by peer" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.551381 4766 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.551406 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.551569 4766 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.551593 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.713461 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.752034 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.758565 4766 apiserver.go:52] "Watching apiserver" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.763355 4766 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.763987 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.764904 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.764954 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.765068 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 00:23:58 crc kubenswrapper[4766]: E1126 00:23:58.765068 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.765204 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:23:58 crc kubenswrapper[4766]: E1126 00:23:58.765277 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.765316 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.765370 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:23:58 crc kubenswrapper[4766]: E1126 00:23:58.765576 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.767750 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.767824 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.767850 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.767749 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.768107 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.768279 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.768285 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.768287 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.768478 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.769669 4766 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789232 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789279 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789307 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789380 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789403 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789422 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789441 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789459 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789482 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789501 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789522 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789548 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789598 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789598 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789622 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789749 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789782 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789818 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789842 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789871 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789893 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789917 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789940 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789964 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789987 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790014 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790044 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790067 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790091 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790116 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790142 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790167 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790196 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790221 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790244 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790266 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790293 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790314 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790335 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790358 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790380 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790406 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790431 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790456 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790478 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790541 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790572 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790594 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790614 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790639 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790690 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790714 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790735 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790760 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790783 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790806 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790830 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790883 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790907 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790929 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790953 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790978 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791006 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791030 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791053 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791078 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791100 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791123 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791177 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791205 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791237 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791261 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791282 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791305 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791328 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791352 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791373 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791396 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791419 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791441 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791462 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791485 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791512 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791537 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791560 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791585 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789794 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789938 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.789989 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790051 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790177 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791636 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790382 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790570 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790606 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790670 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790777 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790894 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790956 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.790953 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791125 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791143 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791284 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791324 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791404 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791468 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.792212 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.792289 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.792470 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.792543 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.792577 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.792683 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.792688 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.792796 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.792810 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.792856 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.792969 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.793049 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.793111 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.793120 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.793143 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.793345 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.793363 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.793353 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.793378 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.793381 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.793408 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.793546 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.793643 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.793805 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.791611 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.793828 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.793854 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.793903 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.793904 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.793942 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.793953 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.793998 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794033 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794066 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794095 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794114 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794122 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794151 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794169 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794179 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794222 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794244 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794274 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794302 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794631 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794676 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794704 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794735 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794761 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794786 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794810 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794837 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794860 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794887 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794912 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794941 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794968 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794996 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795022 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795046 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795068 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795092 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795139 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795157 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795178 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795195 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795211 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795231 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795249 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795267 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795284 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795301 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795317 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795334 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795351 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795367 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795384 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795400 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795417 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795435 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795450 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795465 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795482 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795497 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795512 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795531 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795547 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795566 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795585 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795602 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795621 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795637 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795686 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795702 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795718 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795734 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795750 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795766 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795782 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795800 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795820 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795836 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795853 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795873 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795889 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795909 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795926 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795942 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795957 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795972 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795986 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796003 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796021 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796037 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796053 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796070 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796086 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796103 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796120 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796138 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796155 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796171 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796187 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796202 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796218 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796234 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796251 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796269 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796284 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796352 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796369 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796387 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796403 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796427 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796444 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796463 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796472 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796522 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796540 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796556 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796576 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796594 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796610 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796667 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796698 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796715 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796735 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796753 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796771 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796791 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796812 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796831 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796863 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796881 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796902 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794356 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796940 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796971 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.797278 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.797387 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.797407 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.797528 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.797539 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.797643 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.797806 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.798729 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.799231 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.799391 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.799570 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.799826 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.799927 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.799948 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.800184 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.800204 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.800353 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.800861 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.800866 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794423 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.795668 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796135 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796305 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796314 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796484 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796613 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796678 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: E1126 00:23:58.801100 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:23:59.301057901 +0000 UTC m=+20.149828321 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.801240 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.801303 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.802051 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.802382 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.802751 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.802855 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.802857 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.802879 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.803825 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.803979 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.805073 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.805197 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.805309 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.805311 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.806753 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.806789 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.806812 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.806996 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.807151 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.807409 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.807672 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.807750 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.807770 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.808103 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.808257 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.808298 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.808312 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.808334 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.808710 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.808694 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.808735 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.808826 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.809350 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.809374 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.796917 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.794379 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.809588 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: E1126 00:23:58.809746 4766 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.809751 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.809770 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.809786 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.809917 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.809927 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: E1126 00:23:58.810583 4766 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.810605 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.811384 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.811428 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: E1126 00:23:58.809850 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 00:23:59.309805969 +0000 UTC m=+20.158576609 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.811665 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.811961 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.811983 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.812015 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.812242 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.812328 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.812358 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.812451 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.812691 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.812968 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: E1126 00:23:58.813347 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 00:23:59.313330617 +0000 UTC m=+20.162101047 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.813419 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.814222 4766 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.814430 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.814881 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.815324 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.815334 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.815426 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.815484 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.815727 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.815768 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.815819 4766 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.815849 4766 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.815867 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.815883 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.815897 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.815912 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.815928 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.815942 4766 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.815954 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.815968 4766 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.815980 4766 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.815994 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816006 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816019 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816031 4766 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816043 4766 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816056 4766 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816069 4766 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816082 4766 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816095 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816107 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816119 4766 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816130 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816143 4766 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816155 4766 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816167 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816179 4766 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816191 4766 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816205 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816218 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816231 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816243 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816257 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816269 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816282 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816295 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816309 4766 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816322 4766 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816337 4766 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816350 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816362 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816375 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816387 4766 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816398 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816409 4766 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816421 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816435 4766 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816447 4766 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816459 4766 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816472 4766 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816485 4766 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816497 4766 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816509 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816521 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816536 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816548 4766 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816559 4766 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816571 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816583 4766 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816597 4766 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816609 4766 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816621 4766 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816634 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816660 4766 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816673 4766 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816687 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816698 4766 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816711 4766 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816724 4766 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816736 4766 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816750 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816761 4766 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816772 4766 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816784 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816840 4766 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816852 4766 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816863 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816876 4766 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816888 4766 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816902 4766 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816915 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816928 4766 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816941 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816956 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816970 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816983 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.816995 4766 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817008 4766 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817021 4766 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817038 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817051 4766 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817064 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817076 4766 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817088 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817100 4766 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817113 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817134 4766 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817148 4766 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817161 4766 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817173 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817186 4766 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817199 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817212 4766 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817224 4766 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817238 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817250 4766 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817261 4766 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817274 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817287 4766 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817300 4766 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817313 4766 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817325 4766 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817337 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817349 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817362 4766 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817374 4766 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817385 4766 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817397 4766 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817411 4766 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817415 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.817791 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.818078 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.818168 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.819142 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.819974 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.820054 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.820235 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.822409 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.822500 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.822782 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.823261 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.823309 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.823332 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.823812 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.823814 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.823929 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: E1126 00:23:58.824924 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 00:23:58 crc kubenswrapper[4766]: E1126 00:23:58.824946 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 00:23:58 crc kubenswrapper[4766]: E1126 00:23:58.824958 4766 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:23:58 crc kubenswrapper[4766]: E1126 00:23:58.825016 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 00:23:59.324996848 +0000 UTC m=+20.173767458 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.825233 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.825600 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:58 crc kubenswrapper[4766]: E1126 00:23:58.827829 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 00:23:58 crc kubenswrapper[4766]: E1126 00:23:58.827858 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 00:23:58 crc kubenswrapper[4766]: E1126 00:23:58.827875 4766 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:23:58 crc kubenswrapper[4766]: E1126 00:23:58.827940 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 00:23:59.327920701 +0000 UTC m=+20.176691131 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.828348 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.828351 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.828637 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.828645 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.828826 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.829100 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.830313 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.833215 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.833454 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.833918 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.833492 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.833931 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.834093 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.834163 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.834245 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.834708 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.834863 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.835247 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.835503 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.835741 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.836724 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.837199 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.837357 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.837466 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.837736 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.839550 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.839641 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.839981 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.839986 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.839986 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.840364 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.840833 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.840914 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.841281 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.841356 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.841558 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.842572 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.842869 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.842929 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.847797 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.848600 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.852116 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.853475 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.860243 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.866460 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.875376 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.884207 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.910994 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.912535 4766 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115" exitCode=255 Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.912588 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115"} Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.919115 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.919192 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.919304 4766 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.919398 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.919540 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.919980 4766 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920012 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920030 4766 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920046 4766 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920058 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920070 4766 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920086 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920099 4766 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920112 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920125 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920140 4766 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920151 4766 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920166 4766 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920179 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920191 4766 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920202 4766 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920214 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920226 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920240 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920254 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920266 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920278 4766 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920289 4766 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920301 4766 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920312 4766 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920324 4766 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920336 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920349 4766 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920361 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920374 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920389 4766 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920405 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920421 4766 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920436 4766 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920450 4766 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920462 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920474 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920491 4766 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920509 4766 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920526 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920540 4766 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920554 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920566 4766 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920578 4766 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920594 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920611 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920627 4766 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920666 4766 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920680 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920695 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920750 4766 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920767 4766 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920782 4766 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920797 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920812 4766 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920828 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920843 4766 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920858 4766 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920898 4766 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.920913 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.921064 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.921085 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.921125 4766 scope.go:117] "RemoveContainer" containerID="34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.921767 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.921824 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.921842 4766 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.921857 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.921871 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.921884 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.921897 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.921910 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.921923 4766 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.921971 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.927589 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.933864 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.938514 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.951071 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.963550 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:58 crc kubenswrapper[4766]: I1126 00:23:58.986013 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.023377 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.033542 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.051727 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.076364 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.083376 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 00:23:59 crc kubenswrapper[4766]: W1126 00:23:59.086445 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-b55f5dd3c2538a619bef73042d3724f85f3e19bf0f476ccc39f8e22042988ba4 WatchSource:0}: Error finding container b55f5dd3c2538a619bef73042d3724f85f3e19bf0f476ccc39f8e22042988ba4: Status 404 returned error can't find the container with id b55f5dd3c2538a619bef73042d3724f85f3e19bf0f476ccc39f8e22042988ba4 Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.089001 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 00:23:59 crc kubenswrapper[4766]: W1126 00:23:59.093546 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-9d10cff4e0094421f10445fb3fa822542166fbb204e6384eee2a7fe782fa87fb WatchSource:0}: Error finding container 9d10cff4e0094421f10445fb3fa822542166fbb204e6384eee2a7fe782fa87fb: Status 404 returned error can't find the container with id 9d10cff4e0094421f10445fb3fa822542166fbb204e6384eee2a7fe782fa87fb Nov 26 00:23:59 crc kubenswrapper[4766]: W1126 00:23:59.132277 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-23647fc91a507d21795ef8d01447d795f6612e93f4cb10a1654598202fb20242 WatchSource:0}: Error finding container 23647fc91a507d21795ef8d01447d795f6612e93f4cb10a1654598202fb20242: Status 404 returned error can't find the container with id 23647fc91a507d21795ef8d01447d795f6612e93f4cb10a1654598202fb20242 Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.146021 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.149871 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.154595 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.157821 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.170401 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.184645 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.204685 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.219503 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.239991 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.259778 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.270805 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.278843 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.288455 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.299888 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.309468 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.318156 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.326576 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.326675 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.326708 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:23:59 crc kubenswrapper[4766]: E1126 00:23:59.326720 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:24:00.326703361 +0000 UTC m=+21.175473791 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.326781 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:23:59 crc kubenswrapper[4766]: E1126 00:23:59.326836 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 00:23:59 crc kubenswrapper[4766]: E1126 00:23:59.326855 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 00:23:59 crc kubenswrapper[4766]: E1126 00:23:59.326858 4766 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 00:23:59 crc kubenswrapper[4766]: E1126 00:23:59.326938 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 00:24:00.326915016 +0000 UTC m=+21.175685526 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 00:23:59 crc kubenswrapper[4766]: E1126 00:23:59.326866 4766 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 00:23:59 crc kubenswrapper[4766]: E1126 00:23:59.326868 4766 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:23:59 crc kubenswrapper[4766]: E1126 00:23:59.327049 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 00:24:00.327035139 +0000 UTC m=+21.175805649 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 00:23:59 crc kubenswrapper[4766]: E1126 00:23:59.327086 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 00:24:00.32707566 +0000 UTC m=+21.175846090 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.328807 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.338808 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.355860 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.369382 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.427762 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:23:59 crc kubenswrapper[4766]: E1126 00:23:59.427938 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 00:23:59 crc kubenswrapper[4766]: E1126 00:23:59.427970 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 00:23:59 crc kubenswrapper[4766]: E1126 00:23:59.427983 4766 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:23:59 crc kubenswrapper[4766]: E1126 00:23:59.428042 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 00:24:00.42802563 +0000 UTC m=+21.276796050 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.830979 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.831595 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.833045 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.833823 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.834994 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.835588 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.836322 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.837482 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.838288 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.839545 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.840204 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.841915 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.842627 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.843454 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.844101 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.844590 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.845236 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.847912 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.848401 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.849242 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.850782 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.851310 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.852495 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.853031 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.854301 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.854857 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.855555 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.856857 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.856876 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.857418 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.858627 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.859282 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.860577 4766 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.860779 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.863122 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.864042 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.864575 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.869458 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.869780 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.870339 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.871079 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.872714 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.873939 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.874433 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.875091 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.876039 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.877059 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.877522 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.878516 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.879216 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.880574 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.881066 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.881977 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.882467 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.883066 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.883115 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.884483 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.884985 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.901372 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.912529 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.916456 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.917896 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b"} Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.918086 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.919342 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62"} Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.919374 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a"} Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.919385 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"23647fc91a507d21795ef8d01447d795f6612e93f4cb10a1654598202fb20242"} Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.920466 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e"} Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.920499 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"9d10cff4e0094421f10445fb3fa822542166fbb204e6384eee2a7fe782fa87fb"} Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.921817 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b55f5dd3c2538a619bef73042d3724f85f3e19bf0f476ccc39f8e22042988ba4"} Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.924512 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: E1126 00:23:59.927825 4766 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.940262 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.949679 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.957021 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.965605 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.977321 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 00:23:59 crc kubenswrapper[4766]: I1126 00:23:59.993129 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:23:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:00 crc kubenswrapper[4766]: I1126 00:24:00.005756 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:00 crc kubenswrapper[4766]: I1126 00:24:00.019592 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:00 crc kubenswrapper[4766]: I1126 00:24:00.035251 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:00 crc kubenswrapper[4766]: I1126 00:24:00.057442 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:00 crc kubenswrapper[4766]: I1126 00:24:00.069979 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:00 crc kubenswrapper[4766]: I1126 00:24:00.335432 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:24:00 crc kubenswrapper[4766]: I1126 00:24:00.335542 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:00 crc kubenswrapper[4766]: I1126 00:24:00.335572 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:00 crc kubenswrapper[4766]: I1126 00:24:00.335601 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:00 crc kubenswrapper[4766]: E1126 00:24:00.335765 4766 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 00:24:00 crc kubenswrapper[4766]: E1126 00:24:00.335818 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:24:02.335783157 +0000 UTC m=+23.184553587 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:24:00 crc kubenswrapper[4766]: E1126 00:24:00.335849 4766 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 00:24:00 crc kubenswrapper[4766]: E1126 00:24:00.335866 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 00:24:02.335855199 +0000 UTC m=+23.184625779 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 00:24:00 crc kubenswrapper[4766]: E1126 00:24:00.335788 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 00:24:00 crc kubenswrapper[4766]: E1126 00:24:00.335977 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 00:24:02.335954752 +0000 UTC m=+23.184725272 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 00:24:00 crc kubenswrapper[4766]: E1126 00:24:00.335996 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 00:24:00 crc kubenswrapper[4766]: E1126 00:24:00.336018 4766 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:24:00 crc kubenswrapper[4766]: E1126 00:24:00.336076 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 00:24:02.336066494 +0000 UTC m=+23.184836914 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:24:00 crc kubenswrapper[4766]: I1126 00:24:00.436856 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:00 crc kubenswrapper[4766]: E1126 00:24:00.437004 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 00:24:00 crc kubenswrapper[4766]: E1126 00:24:00.437018 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 00:24:00 crc kubenswrapper[4766]: E1126 00:24:00.437030 4766 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:24:00 crc kubenswrapper[4766]: E1126 00:24:00.437096 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 00:24:02.437083616 +0000 UTC m=+23.285854046 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:24:00 crc kubenswrapper[4766]: I1126 00:24:00.826469 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:00 crc kubenswrapper[4766]: I1126 00:24:00.826547 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:00 crc kubenswrapper[4766]: I1126 00:24:00.826469 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:00 crc kubenswrapper[4766]: E1126 00:24:00.826620 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:00 crc kubenswrapper[4766]: E1126 00:24:00.826744 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:00 crc kubenswrapper[4766]: E1126 00:24:00.826851 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:01 crc kubenswrapper[4766]: I1126 00:24:01.926448 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03"} Nov 26 00:24:01 crc kubenswrapper[4766]: I1126 00:24:01.942293 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:01 crc kubenswrapper[4766]: I1126 00:24:01.957722 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:01 crc kubenswrapper[4766]: I1126 00:24:01.971161 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:01 crc kubenswrapper[4766]: I1126 00:24:01.985710 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:02 crc kubenswrapper[4766]: I1126 00:24:02.002218 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:02 crc kubenswrapper[4766]: I1126 00:24:02.015977 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:02Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:02 crc kubenswrapper[4766]: I1126 00:24:02.031043 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:02Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:02 crc kubenswrapper[4766]: I1126 00:24:02.052250 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:02Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:02 crc kubenswrapper[4766]: I1126 00:24:02.066450 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:02Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:02 crc kubenswrapper[4766]: I1126 00:24:02.351782 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:24:02 crc kubenswrapper[4766]: I1126 00:24:02.351943 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:02 crc kubenswrapper[4766]: E1126 00:24:02.351973 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:24:06.35193028 +0000 UTC m=+27.200700760 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:24:02 crc kubenswrapper[4766]: I1126 00:24:02.352015 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:02 crc kubenswrapper[4766]: I1126 00:24:02.352107 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:02 crc kubenswrapper[4766]: E1126 00:24:02.352178 4766 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 00:24:02 crc kubenswrapper[4766]: E1126 00:24:02.352265 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 00:24:06.352242987 +0000 UTC m=+27.201013457 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 00:24:02 crc kubenswrapper[4766]: E1126 00:24:02.352317 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 00:24:02 crc kubenswrapper[4766]: E1126 00:24:02.352352 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 00:24:02 crc kubenswrapper[4766]: E1126 00:24:02.352376 4766 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:24:02 crc kubenswrapper[4766]: E1126 00:24:02.352444 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 00:24:06.352422332 +0000 UTC m=+27.201192802 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:24:02 crc kubenswrapper[4766]: E1126 00:24:02.352549 4766 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 00:24:02 crc kubenswrapper[4766]: E1126 00:24:02.352613 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 00:24:06.352594626 +0000 UTC m=+27.201365096 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 00:24:02 crc kubenswrapper[4766]: I1126 00:24:02.452531 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:02 crc kubenswrapper[4766]: E1126 00:24:02.452705 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 00:24:02 crc kubenswrapper[4766]: E1126 00:24:02.452719 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 00:24:02 crc kubenswrapper[4766]: E1126 00:24:02.452730 4766 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:24:02 crc kubenswrapper[4766]: E1126 00:24:02.452808 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 00:24:06.452792037 +0000 UTC m=+27.301562457 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:24:02 crc kubenswrapper[4766]: I1126 00:24:02.826168 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:02 crc kubenswrapper[4766]: I1126 00:24:02.826253 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:02 crc kubenswrapper[4766]: E1126 00:24:02.826302 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:02 crc kubenswrapper[4766]: E1126 00:24:02.826389 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:02 crc kubenswrapper[4766]: I1126 00:24:02.826191 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:02 crc kubenswrapper[4766]: E1126 00:24:02.826564 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.422214 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-dm6rv"] Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.422511 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-dm6rv" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.426809 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.427300 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.428124 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.431899 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.448402 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:03Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.466667 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:03Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.486376 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:03Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.502241 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:03Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.528644 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:03Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.546222 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:03Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.564019 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:03Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.564068 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/95807d5c-2f7e-40f9-9116-a7e4766b85c3-serviceca\") pod \"node-ca-dm6rv\" (UID: \"95807d5c-2f7e-40f9-9116-a7e4766b85c3\") " pod="openshift-image-registry/node-ca-dm6rv" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.564115 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/95807d5c-2f7e-40f9-9116-a7e4766b85c3-host\") pod \"node-ca-dm6rv\" (UID: \"95807d5c-2f7e-40f9-9116-a7e4766b85c3\") " pod="openshift-image-registry/node-ca-dm6rv" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.564145 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rz6w\" (UniqueName: \"kubernetes.io/projected/95807d5c-2f7e-40f9-9116-a7e4766b85c3-kube-api-access-8rz6w\") pod \"node-ca-dm6rv\" (UID: \"95807d5c-2f7e-40f9-9116-a7e4766b85c3\") " pod="openshift-image-registry/node-ca-dm6rv" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.578614 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:03Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.598706 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:03Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.614502 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:03Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.665158 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rz6w\" (UniqueName: \"kubernetes.io/projected/95807d5c-2f7e-40f9-9116-a7e4766b85c3-kube-api-access-8rz6w\") pod \"node-ca-dm6rv\" (UID: \"95807d5c-2f7e-40f9-9116-a7e4766b85c3\") " pod="openshift-image-registry/node-ca-dm6rv" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.665196 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/95807d5c-2f7e-40f9-9116-a7e4766b85c3-serviceca\") pod \"node-ca-dm6rv\" (UID: \"95807d5c-2f7e-40f9-9116-a7e4766b85c3\") " pod="openshift-image-registry/node-ca-dm6rv" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.665227 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/95807d5c-2f7e-40f9-9116-a7e4766b85c3-host\") pod \"node-ca-dm6rv\" (UID: \"95807d5c-2f7e-40f9-9116-a7e4766b85c3\") " pod="openshift-image-registry/node-ca-dm6rv" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.665280 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/95807d5c-2f7e-40f9-9116-a7e4766b85c3-host\") pod \"node-ca-dm6rv\" (UID: \"95807d5c-2f7e-40f9-9116-a7e4766b85c3\") " pod="openshift-image-registry/node-ca-dm6rv" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.673794 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/95807d5c-2f7e-40f9-9116-a7e4766b85c3-serviceca\") pod \"node-ca-dm6rv\" (UID: \"95807d5c-2f7e-40f9-9116-a7e4766b85c3\") " pod="openshift-image-registry/node-ca-dm6rv" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.682556 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rz6w\" (UniqueName: \"kubernetes.io/projected/95807d5c-2f7e-40f9-9116-a7e4766b85c3-kube-api-access-8rz6w\") pod \"node-ca-dm6rv\" (UID: \"95807d5c-2f7e-40f9-9116-a7e4766b85c3\") " pod="openshift-image-registry/node-ca-dm6rv" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.733278 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-dm6rv" Nov 26 00:24:03 crc kubenswrapper[4766]: W1126 00:24:03.743490 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95807d5c_2f7e_40f9_9116_a7e4766b85c3.slice/crio-805a39f31bd42df61c300b95b68eafa7d8d57628880bc3e4a02a1a8e5e3912a8 WatchSource:0}: Error finding container 805a39f31bd42df61c300b95b68eafa7d8d57628880bc3e4a02a1a8e5e3912a8: Status 404 returned error can't find the container with id 805a39f31bd42df61c300b95b68eafa7d8d57628880bc3e4a02a1a8e5e3912a8 Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.838099 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-8s7f9"] Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.838365 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-wf9c2"] Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.838526 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-8s7f9" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.839137 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-b8bk6"] Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.839597 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.839882 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.839904 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-prqrp"] Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.840429 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-prqrp" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.840965 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.841040 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.841924 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.842042 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.842187 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.843598 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.843773 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.843934 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.844033 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.844098 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.844126 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.844299 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.844363 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.846956 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.846967 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.859893 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:03Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.876165 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:03Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.890276 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:03Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.904697 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:03Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.918825 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:03Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.932880 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-dm6rv" event={"ID":"95807d5c-2f7e-40f9-9116-a7e4766b85c3","Type":"ContainerStarted","Data":"805a39f31bd42df61c300b95b68eafa7d8d57628880bc3e4a02a1a8e5e3912a8"} Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.934059 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:03Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.945759 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:03Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.957130 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:03Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.968496 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-host-run-netns\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.968553 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtx6r\" (UniqueName: \"kubernetes.io/projected/561423b1-063f-4968-a1b3-4187bb470b20-kube-api-access-dtx6r\") pod \"node-resolver-8s7f9\" (UID: \"561423b1-063f-4968-a1b3-4187bb470b20\") " pod="openshift-dns/node-resolver-8s7f9" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.968579 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-etc-kubernetes\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.968756 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/63471884-61f7-41d5-8967-e3f81eba90d9-mcd-auth-proxy-config\") pod \"machine-config-daemon-wf9c2\" (UID: \"63471884-61f7-41d5-8967-e3f81eba90d9\") " pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.968815 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/beeb3530-3b3b-40d9-8b0e-1b927c64920a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-b8bk6\" (UID: \"beeb3530-3b3b-40d9-8b0e-1b927c64920a\") " pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.968872 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/234068f7-4097-4e6b-aac3-4bafe425c1f2-multus-daemon-config\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.968924 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/63471884-61f7-41d5-8967-e3f81eba90d9-rootfs\") pod \"machine-config-daemon-wf9c2\" (UID: \"63471884-61f7-41d5-8967-e3f81eba90d9\") " pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.968958 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-host-var-lib-cni-bin\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.968984 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-hostroot\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.969071 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-host-run-multus-certs\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.969173 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kchh\" (UniqueName: \"kubernetes.io/projected/234068f7-4097-4e6b-aac3-4bafe425c1f2-kube-api-access-6kchh\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.969208 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/561423b1-063f-4968-a1b3-4187bb470b20-hosts-file\") pod \"node-resolver-8s7f9\" (UID: \"561423b1-063f-4968-a1b3-4187bb470b20\") " pod="openshift-dns/node-resolver-8s7f9" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.969234 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-system-cni-dir\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.969259 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-host-var-lib-cni-multus\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.969284 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sb9l\" (UniqueName: \"kubernetes.io/projected/63471884-61f7-41d5-8967-e3f81eba90d9-kube-api-access-4sb9l\") pod \"machine-config-daemon-wf9c2\" (UID: \"63471884-61f7-41d5-8967-e3f81eba90d9\") " pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.969307 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/beeb3530-3b3b-40d9-8b0e-1b927c64920a-os-release\") pod \"multus-additional-cni-plugins-b8bk6\" (UID: \"beeb3530-3b3b-40d9-8b0e-1b927c64920a\") " pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.969330 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/beeb3530-3b3b-40d9-8b0e-1b927c64920a-cni-binary-copy\") pod \"multus-additional-cni-plugins-b8bk6\" (UID: \"beeb3530-3b3b-40d9-8b0e-1b927c64920a\") " pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.969364 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55p6x\" (UniqueName: \"kubernetes.io/projected/beeb3530-3b3b-40d9-8b0e-1b927c64920a-kube-api-access-55p6x\") pod \"multus-additional-cni-plugins-b8bk6\" (UID: \"beeb3530-3b3b-40d9-8b0e-1b927c64920a\") " pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.969469 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-os-release\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.969578 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/beeb3530-3b3b-40d9-8b0e-1b927c64920a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-b8bk6\" (UID: \"beeb3530-3b3b-40d9-8b0e-1b927c64920a\") " pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.969610 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-multus-conf-dir\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.969634 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-cnibin\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.969674 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-host-run-k8s-cni-cncf-io\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.969692 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-multus-cni-dir\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.969736 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-multus-socket-dir-parent\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.969768 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-host-var-lib-kubelet\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.969828 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/beeb3530-3b3b-40d9-8b0e-1b927c64920a-system-cni-dir\") pod \"multus-additional-cni-plugins-b8bk6\" (UID: \"beeb3530-3b3b-40d9-8b0e-1b927c64920a\") " pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.969879 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/beeb3530-3b3b-40d9-8b0e-1b927c64920a-cnibin\") pod \"multus-additional-cni-plugins-b8bk6\" (UID: \"beeb3530-3b3b-40d9-8b0e-1b927c64920a\") " pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.969918 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/234068f7-4097-4e6b-aac3-4bafe425c1f2-cni-binary-copy\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.969957 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/63471884-61f7-41d5-8967-e3f81eba90d9-proxy-tls\") pod \"machine-config-daemon-wf9c2\" (UID: \"63471884-61f7-41d5-8967-e3f81eba90d9\") " pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:24:03 crc kubenswrapper[4766]: I1126 00:24:03.979542 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:03Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.028781 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.049214 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.060625 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.070412 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtx6r\" (UniqueName: \"kubernetes.io/projected/561423b1-063f-4968-a1b3-4187bb470b20-kube-api-access-dtx6r\") pod \"node-resolver-8s7f9\" (UID: \"561423b1-063f-4968-a1b3-4187bb470b20\") " pod="openshift-dns/node-resolver-8s7f9" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.070453 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/63471884-61f7-41d5-8967-e3f81eba90d9-mcd-auth-proxy-config\") pod \"machine-config-daemon-wf9c2\" (UID: \"63471884-61f7-41d5-8967-e3f81eba90d9\") " pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.070472 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-etc-kubernetes\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.070490 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/beeb3530-3b3b-40d9-8b0e-1b927c64920a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-b8bk6\" (UID: \"beeb3530-3b3b-40d9-8b0e-1b927c64920a\") " pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.070509 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/234068f7-4097-4e6b-aac3-4bafe425c1f2-multus-daemon-config\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.070533 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/63471884-61f7-41d5-8967-e3f81eba90d9-rootfs\") pod \"machine-config-daemon-wf9c2\" (UID: \"63471884-61f7-41d5-8967-e3f81eba90d9\") " pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.070550 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-host-var-lib-cni-bin\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.070565 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-hostroot\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.070588 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-host-run-multus-certs\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.070603 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kchh\" (UniqueName: \"kubernetes.io/projected/234068f7-4097-4e6b-aac3-4bafe425c1f2-kube-api-access-6kchh\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.070621 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/561423b1-063f-4968-a1b3-4187bb470b20-hosts-file\") pod \"node-resolver-8s7f9\" (UID: \"561423b1-063f-4968-a1b3-4187bb470b20\") " pod="openshift-dns/node-resolver-8s7f9" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.070631 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-etc-kubernetes\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.070696 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-host-var-lib-cni-bin\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.070775 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-host-run-multus-certs\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.070643 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/beeb3530-3b3b-40d9-8b0e-1b927c64920a-os-release\") pod \"multus-additional-cni-plugins-b8bk6\" (UID: \"beeb3530-3b3b-40d9-8b0e-1b927c64920a\") " pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.070795 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/561423b1-063f-4968-a1b3-4187bb470b20-hosts-file\") pod \"node-resolver-8s7f9\" (UID: \"561423b1-063f-4968-a1b3-4187bb470b20\") " pod="openshift-dns/node-resolver-8s7f9" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.070825 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/63471884-61f7-41d5-8967-e3f81eba90d9-rootfs\") pod \"machine-config-daemon-wf9c2\" (UID: \"63471884-61f7-41d5-8967-e3f81eba90d9\") " pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.070849 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/beeb3530-3b3b-40d9-8b0e-1b927c64920a-os-release\") pod \"multus-additional-cni-plugins-b8bk6\" (UID: \"beeb3530-3b3b-40d9-8b0e-1b927c64920a\") " pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.070873 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/beeb3530-3b3b-40d9-8b0e-1b927c64920a-cni-binary-copy\") pod \"multus-additional-cni-plugins-b8bk6\" (UID: \"beeb3530-3b3b-40d9-8b0e-1b927c64920a\") " pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.070845 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-hostroot\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.071328 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55p6x\" (UniqueName: \"kubernetes.io/projected/beeb3530-3b3b-40d9-8b0e-1b927c64920a-kube-api-access-55p6x\") pod \"multus-additional-cni-plugins-b8bk6\" (UID: \"beeb3530-3b3b-40d9-8b0e-1b927c64920a\") " pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.071461 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-system-cni-dir\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.071496 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-host-var-lib-cni-multus\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.071627 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sb9l\" (UniqueName: \"kubernetes.io/projected/63471884-61f7-41d5-8967-e3f81eba90d9-kube-api-access-4sb9l\") pod \"machine-config-daemon-wf9c2\" (UID: \"63471884-61f7-41d5-8967-e3f81eba90d9\") " pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.071694 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/beeb3530-3b3b-40d9-8b0e-1b927c64920a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-b8bk6\" (UID: \"beeb3530-3b3b-40d9-8b0e-1b927c64920a\") " pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.071719 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/234068f7-4097-4e6b-aac3-4bafe425c1f2-multus-daemon-config\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.071895 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-os-release\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.071920 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-cnibin\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.071944 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-host-run-k8s-cni-cncf-io\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.071963 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-multus-conf-dir\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.072065 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/63471884-61f7-41d5-8967-e3f81eba90d9-mcd-auth-proxy-config\") pod \"machine-config-daemon-wf9c2\" (UID: \"63471884-61f7-41d5-8967-e3f81eba90d9\") " pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.072111 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/beeb3530-3b3b-40d9-8b0e-1b927c64920a-cni-binary-copy\") pod \"multus-additional-cni-plugins-b8bk6\" (UID: \"beeb3530-3b3b-40d9-8b0e-1b927c64920a\") " pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.072164 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/beeb3530-3b3b-40d9-8b0e-1b927c64920a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-b8bk6\" (UID: \"beeb3530-3b3b-40d9-8b0e-1b927c64920a\") " pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.072199 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-multus-cni-dir\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.072228 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-multus-socket-dir-parent\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.072251 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-host-var-lib-kubelet\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.072282 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-system-cni-dir\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.072312 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-multus-conf-dir\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.072976 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-os-release\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.073014 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-cnibin\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.073035 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-host-run-k8s-cni-cncf-io\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.073149 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-multus-cni-dir\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.073190 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-multus-socket-dir-parent\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.073228 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-host-var-lib-kubelet\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.073278 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/63471884-61f7-41d5-8967-e3f81eba90d9-proxy-tls\") pod \"machine-config-daemon-wf9c2\" (UID: \"63471884-61f7-41d5-8967-e3f81eba90d9\") " pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.073300 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/beeb3530-3b3b-40d9-8b0e-1b927c64920a-system-cni-dir\") pod \"multus-additional-cni-plugins-b8bk6\" (UID: \"beeb3530-3b3b-40d9-8b0e-1b927c64920a\") " pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.073301 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/beeb3530-3b3b-40d9-8b0e-1b927c64920a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-b8bk6\" (UID: \"beeb3530-3b3b-40d9-8b0e-1b927c64920a\") " pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.073361 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/beeb3530-3b3b-40d9-8b0e-1b927c64920a-system-cni-dir\") pod \"multus-additional-cni-plugins-b8bk6\" (UID: \"beeb3530-3b3b-40d9-8b0e-1b927c64920a\") " pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.073369 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/beeb3530-3b3b-40d9-8b0e-1b927c64920a-cnibin\") pod \"multus-additional-cni-plugins-b8bk6\" (UID: \"beeb3530-3b3b-40d9-8b0e-1b927c64920a\") " pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.073385 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-host-var-lib-cni-multus\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.073408 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/234068f7-4097-4e6b-aac3-4bafe425c1f2-cni-binary-copy\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.073989 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/234068f7-4097-4e6b-aac3-4bafe425c1f2-cni-binary-copy\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.074024 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/beeb3530-3b3b-40d9-8b0e-1b927c64920a-cnibin\") pod \"multus-additional-cni-plugins-b8bk6\" (UID: \"beeb3530-3b3b-40d9-8b0e-1b927c64920a\") " pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.074297 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-host-run-netns\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.074350 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/234068f7-4097-4e6b-aac3-4bafe425c1f2-host-run-netns\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.077496 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/63471884-61f7-41d5-8967-e3f81eba90d9-proxy-tls\") pod \"machine-config-daemon-wf9c2\" (UID: \"63471884-61f7-41d5-8967-e3f81eba90d9\") " pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.079388 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.092313 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtx6r\" (UniqueName: \"kubernetes.io/projected/561423b1-063f-4968-a1b3-4187bb470b20-kube-api-access-dtx6r\") pod \"node-resolver-8s7f9\" (UID: \"561423b1-063f-4968-a1b3-4187bb470b20\") " pod="openshift-dns/node-resolver-8s7f9" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.092396 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sb9l\" (UniqueName: \"kubernetes.io/projected/63471884-61f7-41d5-8967-e3f81eba90d9-kube-api-access-4sb9l\") pod \"machine-config-daemon-wf9c2\" (UID: \"63471884-61f7-41d5-8967-e3f81eba90d9\") " pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.092833 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.095258 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kchh\" (UniqueName: \"kubernetes.io/projected/234068f7-4097-4e6b-aac3-4bafe425c1f2-kube-api-access-6kchh\") pod \"multus-prqrp\" (UID: \"234068f7-4097-4e6b-aac3-4bafe425c1f2\") " pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.102826 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55p6x\" (UniqueName: \"kubernetes.io/projected/beeb3530-3b3b-40d9-8b0e-1b927c64920a-kube-api-access-55p6x\") pod \"multus-additional-cni-plugins-b8bk6\" (UID: \"beeb3530-3b3b-40d9-8b0e-1b927c64920a\") " pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.104254 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.119833 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.135754 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.149993 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.155516 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-8s7f9" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.164996 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" Nov 26 00:24:04 crc kubenswrapper[4766]: W1126 00:24:04.167520 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod561423b1_063f_4968_a1b3_4187bb470b20.slice/crio-a706e6009200662d778652a46c28c02f527c6ae963a011c7b936b80a785f4c1c WatchSource:0}: Error finding container a706e6009200662d778652a46c28c02f527c6ae963a011c7b936b80a785f4c1c: Status 404 returned error can't find the container with id a706e6009200662d778652a46c28c02f527c6ae963a011c7b936b80a785f4c1c Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.169570 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.171179 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.181101 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-prqrp" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.187737 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: W1126 00:24:04.200124 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod234068f7_4097_4e6b_aac3_4bafe425c1f2.slice/crio-7660318bd71c1d6ab6ea9754e0907a4f916ea4a8568cda23ff518d60fa92bec7 WatchSource:0}: Error finding container 7660318bd71c1d6ab6ea9754e0907a4f916ea4a8568cda23ff518d60fa92bec7: Status 404 returned error can't find the container with id 7660318bd71c1d6ab6ea9754e0907a4f916ea4a8568cda23ff518d60fa92bec7 Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.216450 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.220363 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fxnln"] Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.221149 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.223483 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.223714 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.223838 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.223945 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.224032 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.224142 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.226355 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.243597 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.255844 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.279325 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.296622 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.315082 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.341512 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.359632 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.374267 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.380478 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-systemd-units\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.380522 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-var-lib-openvswitch\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.380551 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-run-netns\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.380640 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-run-openvswitch\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.380686 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-node-log\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.380703 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/415fdacb-85c1-4265-89b7-6771a84ffc89-ovnkube-script-lib\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.380759 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-cni-bin\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.380778 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.380847 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-kubelet\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.380865 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-etc-openvswitch\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.380888 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/415fdacb-85c1-4265-89b7-6771a84ffc89-ovn-node-metrics-cert\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.380920 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-run-systemd\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.380942 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-run-ovn\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.380958 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/415fdacb-85c1-4265-89b7-6771a84ffc89-ovnkube-config\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.380980 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-log-socket\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.380999 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-run-ovn-kubernetes\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.381018 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l696l\" (UniqueName: \"kubernetes.io/projected/415fdacb-85c1-4265-89b7-6771a84ffc89-kube-api-access-l696l\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.381077 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/415fdacb-85c1-4265-89b7-6771a84ffc89-env-overrides\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.381120 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-slash\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.381180 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-cni-netd\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.403132 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.434078 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.481910 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-systemd-units\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.481961 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-var-lib-openvswitch\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.481987 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-run-netns\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482010 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-run-openvswitch\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482031 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-node-log\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482055 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/415fdacb-85c1-4265-89b7-6771a84ffc89-ovnkube-script-lib\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482076 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-cni-bin\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482098 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482144 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-kubelet\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482166 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-etc-openvswitch\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482188 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/415fdacb-85c1-4265-89b7-6771a84ffc89-ovn-node-metrics-cert\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482207 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-run-systemd\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482235 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/415fdacb-85c1-4265-89b7-6771a84ffc89-ovnkube-config\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482258 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-run-ovn\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482277 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-log-socket\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482297 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-run-ovn-kubernetes\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482318 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l696l\" (UniqueName: \"kubernetes.io/projected/415fdacb-85c1-4265-89b7-6771a84ffc89-kube-api-access-l696l\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482339 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/415fdacb-85c1-4265-89b7-6771a84ffc89-env-overrides\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482371 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-cni-netd\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482394 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-slash\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482460 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-slash\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482521 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-systemd-units\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482553 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-var-lib-openvswitch\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482582 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-run-netns\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482620 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-run-openvswitch\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.482662 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-node-log\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.483360 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/415fdacb-85c1-4265-89b7-6771a84ffc89-ovnkube-script-lib\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.483412 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-cni-bin\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.483448 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.483480 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-kubelet\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.483508 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-etc-openvswitch\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.483763 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-log-socket\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.483843 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-run-systemd\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.484435 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-run-ovn-kubernetes\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.484476 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-cni-netd\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.484616 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/415fdacb-85c1-4265-89b7-6771a84ffc89-env-overrides\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.484740 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/415fdacb-85c1-4265-89b7-6771a84ffc89-ovnkube-config\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.485900 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-run-ovn\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.489341 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/415fdacb-85c1-4265-89b7-6771a84ffc89-ovn-node-metrics-cert\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.492947 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.514597 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l696l\" (UniqueName: \"kubernetes.io/projected/415fdacb-85c1-4265-89b7-6771a84ffc89-kube-api-access-l696l\") pod \"ovnkube-node-fxnln\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.537824 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.539498 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.563038 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.575080 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.585467 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.600840 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.611533 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.620623 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.631782 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.826797 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.826928 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:04 crc kubenswrapper[4766]: E1126 00:24:04.826981 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:04 crc kubenswrapper[4766]: E1126 00:24:04.827060 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.827310 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:04 crc kubenswrapper[4766]: E1126 00:24:04.827408 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.892421 4766 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.893910 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.893957 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.893970 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.894085 4766 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.901236 4766 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.901667 4766 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.902907 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.902951 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.902963 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.902980 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.902992 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:04Z","lastTransitionTime":"2025-11-26T00:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:04 crc kubenswrapper[4766]: E1126 00:24:04.922777 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.926733 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.926832 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.926851 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.926869 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.926879 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:04Z","lastTransitionTime":"2025-11-26T00:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.937739 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-8s7f9" event={"ID":"561423b1-063f-4968-a1b3-4187bb470b20","Type":"ContainerStarted","Data":"ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71"} Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.937796 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-8s7f9" event={"ID":"561423b1-063f-4968-a1b3-4187bb470b20","Type":"ContainerStarted","Data":"a706e6009200662d778652a46c28c02f527c6ae963a011c7b936b80a785f4c1c"} Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.941791 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472"} Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.941842 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf"} Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.941853 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"69c849398194770a92ac7e9ab9737bec8adbd68e2658c4a7f15499e58695ae2b"} Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.943017 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-dm6rv" event={"ID":"95807d5c-2f7e-40f9-9116-a7e4766b85c3","Type":"ContainerStarted","Data":"85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a"} Nov 26 00:24:04 crc kubenswrapper[4766]: E1126 00:24:04.944095 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.944858 4766 generic.go:334] "Generic (PLEG): container finished" podID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerID="d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7" exitCode=0 Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.944913 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerDied","Data":"d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7"} Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.944947 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerStarted","Data":"cde41151cb667db048d6f2ca5f4878532b799d90003bd6ebba008b716686f3c2"} Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.946546 4766 generic.go:334] "Generic (PLEG): container finished" podID="beeb3530-3b3b-40d9-8b0e-1b927c64920a" containerID="50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977" exitCode=0 Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.946605 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" event={"ID":"beeb3530-3b3b-40d9-8b0e-1b927c64920a","Type":"ContainerDied","Data":"50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977"} Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.946623 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" event={"ID":"beeb3530-3b3b-40d9-8b0e-1b927c64920a","Type":"ContainerStarted","Data":"c572fcf84560ef05831763abfa2878f1333e241f3030bd68af9b3370f8bc6643"} Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.948031 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-prqrp" event={"ID":"234068f7-4097-4e6b-aac3-4bafe425c1f2","Type":"ContainerStarted","Data":"aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c"} Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.948067 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-prqrp" event={"ID":"234068f7-4097-4e6b-aac3-4bafe425c1f2","Type":"ContainerStarted","Data":"7660318bd71c1d6ab6ea9754e0907a4f916ea4a8568cda23ff518d60fa92bec7"} Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.948561 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.948582 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.948593 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.948606 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.948617 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:04Z","lastTransitionTime":"2025-11-26T00:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.957025 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: E1126 00:24:04.965012 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.969305 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.969349 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.969392 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.969431 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.969448 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:04Z","lastTransitionTime":"2025-11-26T00:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.970901 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.982427 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: E1126 00:24:04.982568 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.985838 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.985863 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.985874 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.985889 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.985901 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:04Z","lastTransitionTime":"2025-11-26T00:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.995596 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: E1126 00:24:04.997313 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:04Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:04 crc kubenswrapper[4766]: E1126 00:24:04.997450 4766 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.998888 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.998929 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.998941 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.998961 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:04 crc kubenswrapper[4766]: I1126 00:24:04.998972 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:04Z","lastTransitionTime":"2025-11-26T00:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.007928 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.019060 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.032856 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.046672 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.070731 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.080760 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.092674 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.101110 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.101152 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.101165 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.101184 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.101197 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:05Z","lastTransitionTime":"2025-11-26T00:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.103604 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.115914 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.132456 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.153732 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.165915 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.176141 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.188335 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.204114 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.204148 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.204159 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.204174 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.204185 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:05Z","lastTransitionTime":"2025-11-26T00:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.204242 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.221591 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.233460 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.243147 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.254427 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.264676 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.277789 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.288417 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.299691 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.306499 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.306541 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.306551 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.306566 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.306577 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:05Z","lastTransitionTime":"2025-11-26T00:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.312010 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.331461 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.342062 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.409153 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.409184 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.409192 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.409206 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.409215 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:05Z","lastTransitionTime":"2025-11-26T00:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.511884 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.511925 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.511936 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.511952 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.511964 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:05Z","lastTransitionTime":"2025-11-26T00:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.613897 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.613939 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.613951 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.613981 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.613993 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:05Z","lastTransitionTime":"2025-11-26T00:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.716799 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.716836 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.716847 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.716863 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.716874 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:05Z","lastTransitionTime":"2025-11-26T00:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.819192 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.819244 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.819270 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.819289 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.819300 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:05Z","lastTransitionTime":"2025-11-26T00:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.922555 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.922611 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.922627 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.922672 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.922691 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:05Z","lastTransitionTime":"2025-11-26T00:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.952951 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" event={"ID":"beeb3530-3b3b-40d9-8b0e-1b927c64920a","Type":"ContainerStarted","Data":"b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e"} Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.957116 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerStarted","Data":"2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820"} Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.957164 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerStarted","Data":"619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498"} Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.957181 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerStarted","Data":"cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99"} Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.957193 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerStarted","Data":"abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b"} Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.969901 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:05 crc kubenswrapper[4766]: I1126 00:24:05.990053 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:05Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.008756 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:06Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.025321 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:06Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.025417 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.025444 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.025456 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.025471 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.025484 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:06Z","lastTransitionTime":"2025-11-26T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.041475 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:06Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.053436 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:06Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.062899 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:06Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.077913 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:06Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.094339 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:06Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.109306 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:06Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.124942 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:06Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.128355 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.128392 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.128403 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.128423 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.128436 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:06Z","lastTransitionTime":"2025-11-26T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.136502 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:06Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.149929 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:06Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.168509 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:06Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.178856 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:06Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.230719 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.230758 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.230769 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.230785 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.230796 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:06Z","lastTransitionTime":"2025-11-26T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.332815 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.332855 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.332866 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.332882 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.332894 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:06Z","lastTransitionTime":"2025-11-26T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.397562 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.397691 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.397726 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:06 crc kubenswrapper[4766]: E1126 00:24:06.397813 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:24:14.397779623 +0000 UTC m=+35.246550063 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:24:06 crc kubenswrapper[4766]: E1126 00:24:06.397860 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 00:24:06 crc kubenswrapper[4766]: E1126 00:24:06.397885 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 00:24:06 crc kubenswrapper[4766]: E1126 00:24:06.397886 4766 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 00:24:06 crc kubenswrapper[4766]: E1126 00:24:06.397898 4766 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.397930 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:06 crc kubenswrapper[4766]: E1126 00:24:06.397947 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 00:24:14.397933607 +0000 UTC m=+35.246704087 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 00:24:06 crc kubenswrapper[4766]: E1126 00:24:06.397970 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 00:24:14.397961567 +0000 UTC m=+35.246732057 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:24:06 crc kubenswrapper[4766]: E1126 00:24:06.398042 4766 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 00:24:06 crc kubenswrapper[4766]: E1126 00:24:06.398075 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 00:24:14.39806706 +0000 UTC m=+35.246837530 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.435125 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.435161 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.435170 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.435184 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.435195 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:06Z","lastTransitionTime":"2025-11-26T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.499113 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:06 crc kubenswrapper[4766]: E1126 00:24:06.499332 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 00:24:06 crc kubenswrapper[4766]: E1126 00:24:06.499366 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 00:24:06 crc kubenswrapper[4766]: E1126 00:24:06.499379 4766 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:24:06 crc kubenswrapper[4766]: E1126 00:24:06.499436 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 00:24:14.49942134 +0000 UTC m=+35.348191770 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.537285 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.537587 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.537596 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.537611 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.537621 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:06Z","lastTransitionTime":"2025-11-26T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.639555 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.639594 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.639604 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.639618 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.639628 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:06Z","lastTransitionTime":"2025-11-26T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.741892 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.741930 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.741941 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.741957 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.741968 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:06Z","lastTransitionTime":"2025-11-26T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.826050 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:06 crc kubenswrapper[4766]: E1126 00:24:06.826192 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.826072 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.826072 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:06 crc kubenswrapper[4766]: E1126 00:24:06.826273 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:06 crc kubenswrapper[4766]: E1126 00:24:06.826458 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.844374 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.844412 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.844421 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.844436 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.844448 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:06Z","lastTransitionTime":"2025-11-26T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.947547 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.947639 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.947680 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.947701 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.947713 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:06Z","lastTransitionTime":"2025-11-26T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.966286 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerStarted","Data":"4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52"} Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.966330 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerStarted","Data":"ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff"} Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.967906 4766 generic.go:334] "Generic (PLEG): container finished" podID="beeb3530-3b3b-40d9-8b0e-1b927c64920a" containerID="b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e" exitCode=0 Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.967941 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" event={"ID":"beeb3530-3b3b-40d9-8b0e-1b927c64920a","Type":"ContainerDied","Data":"b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e"} Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.983212 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:06Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:06 crc kubenswrapper[4766]: I1126 00:24:06.996104 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:06Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.010420 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:07Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.024580 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:07Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.037503 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:07Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.048615 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:07Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.052479 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.053153 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.053166 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.053182 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.053192 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:07Z","lastTransitionTime":"2025-11-26T00:24:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.061732 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:07Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.076974 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:07Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.098215 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:07Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.111044 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:07Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.129526 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:07Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.146697 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:07Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.156068 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.156547 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.156560 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.156581 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.156594 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:07Z","lastTransitionTime":"2025-11-26T00:24:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.163445 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:07Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.176431 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:07Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.192987 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:07Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.259249 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.259296 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.259307 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.259325 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.259339 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:07Z","lastTransitionTime":"2025-11-26T00:24:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.361638 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.361693 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.361703 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.361716 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.361726 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:07Z","lastTransitionTime":"2025-11-26T00:24:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.464115 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.464156 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.464165 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.464182 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.464192 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:07Z","lastTransitionTime":"2025-11-26T00:24:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.567135 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.567193 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.567204 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.567224 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.567239 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:07Z","lastTransitionTime":"2025-11-26T00:24:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.669979 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.670026 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.670042 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.670066 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.670082 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:07Z","lastTransitionTime":"2025-11-26T00:24:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.772727 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.772781 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.772793 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.772811 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.772824 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:07Z","lastTransitionTime":"2025-11-26T00:24:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.876119 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.876193 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.876219 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.876251 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.876281 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:07Z","lastTransitionTime":"2025-11-26T00:24:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.974256 4766 generic.go:334] "Generic (PLEG): container finished" podID="beeb3530-3b3b-40d9-8b0e-1b927c64920a" containerID="dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2" exitCode=0 Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.974301 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" event={"ID":"beeb3530-3b3b-40d9-8b0e-1b927c64920a","Type":"ContainerDied","Data":"dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2"} Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.978790 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.978827 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.978838 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.978855 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.978866 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:07Z","lastTransitionTime":"2025-11-26T00:24:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:07 crc kubenswrapper[4766]: I1126 00:24:07.988554 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:07Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.006118 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:08Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.024939 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:08Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.038343 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:08Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.050577 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:08Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.064374 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:08Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.082858 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.082911 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.082924 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.082947 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.082963 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:08Z","lastTransitionTime":"2025-11-26T00:24:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.084880 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:08Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.095566 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:08Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.111332 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:08Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.123527 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:08Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.133991 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:08Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.148310 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:08Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.163735 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:08Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.175423 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:08Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.184155 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:08Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.184909 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.184940 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.184953 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.184971 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.184982 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:08Z","lastTransitionTime":"2025-11-26T00:24:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.290261 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.290327 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.290355 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.290379 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.290397 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:08Z","lastTransitionTime":"2025-11-26T00:24:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.392320 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.392349 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.392359 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.392374 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.392384 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:08Z","lastTransitionTime":"2025-11-26T00:24:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.494987 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.495014 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.495022 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.495040 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.495059 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:08Z","lastTransitionTime":"2025-11-26T00:24:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.597420 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.597465 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.597478 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.597495 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.597510 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:08Z","lastTransitionTime":"2025-11-26T00:24:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.700574 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.700619 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.700631 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.700659 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.700668 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:08Z","lastTransitionTime":"2025-11-26T00:24:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.803262 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.803307 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.803320 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.803337 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.803349 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:08Z","lastTransitionTime":"2025-11-26T00:24:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.826145 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.826211 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.826261 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:08 crc kubenswrapper[4766]: E1126 00:24:08.826293 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:08 crc kubenswrapper[4766]: E1126 00:24:08.826363 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:08 crc kubenswrapper[4766]: E1126 00:24:08.826406 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.905739 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.905782 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.905791 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.905808 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.905824 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:08Z","lastTransitionTime":"2025-11-26T00:24:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.981639 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerStarted","Data":"4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da"} Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.984073 4766 generic.go:334] "Generic (PLEG): container finished" podID="beeb3530-3b3b-40d9-8b0e-1b927c64920a" containerID="d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af" exitCode=0 Nov 26 00:24:08 crc kubenswrapper[4766]: I1126 00:24:08.984111 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" event={"ID":"beeb3530-3b3b-40d9-8b0e-1b927c64920a","Type":"ContainerDied","Data":"d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af"} Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.000638 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:08Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.009231 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.009267 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.009279 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.009296 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.009309 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:09Z","lastTransitionTime":"2025-11-26T00:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.019920 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.033606 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.049238 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.063103 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.081845 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.092303 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.106697 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.110977 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.111024 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.111034 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.111067 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.111078 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:09Z","lastTransitionTime":"2025-11-26T00:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.118362 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.127919 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.138902 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.152800 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.172812 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.185145 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.195063 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.213376 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.213412 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.213424 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.213439 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.213450 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:09Z","lastTransitionTime":"2025-11-26T00:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.316118 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.316161 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.316173 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.316188 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.316199 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:09Z","lastTransitionTime":"2025-11-26T00:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.418822 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.418860 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.418872 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.418888 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.418900 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:09Z","lastTransitionTime":"2025-11-26T00:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.521224 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.521259 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.521269 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.521285 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.521295 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:09Z","lastTransitionTime":"2025-11-26T00:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.623165 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.623202 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.623215 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.623230 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.623240 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:09Z","lastTransitionTime":"2025-11-26T00:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.726046 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.726096 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.726108 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.726132 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.726144 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:09Z","lastTransitionTime":"2025-11-26T00:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.829643 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.829730 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.829747 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.829791 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.829811 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:09Z","lastTransitionTime":"2025-11-26T00:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.840917 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.852703 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.864681 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.881469 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.931846 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.931891 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.931901 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.931916 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.931926 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:09Z","lastTransitionTime":"2025-11-26T00:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.942265 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.959125 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.967471 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.979335 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.988777 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" event={"ID":"beeb3530-3b3b-40d9-8b0e-1b927c64920a","Type":"ContainerStarted","Data":"3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4"} Nov 26 00:24:09 crc kubenswrapper[4766]: I1126 00:24:09.991594 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.001020 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:09Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.011190 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:10Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.021360 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:10Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.034364 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.034391 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.034401 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.034415 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.034425 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:10Z","lastTransitionTime":"2025-11-26T00:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.037523 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:10Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.048439 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:10Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.059742 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:10Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.069128 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:10Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.081968 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:10Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.094888 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:10Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.109038 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:10Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.120852 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:10Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.132280 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:10Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.136497 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.136535 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.136545 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.136560 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.136577 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:10Z","lastTransitionTime":"2025-11-26T00:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.154590 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:10Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.164927 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:10Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.176124 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:10Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.189000 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:10Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.210646 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:10Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.228565 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:10Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.239169 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.239224 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.239239 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.239259 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.239273 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:10Z","lastTransitionTime":"2025-11-26T00:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.243701 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:10Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.254816 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:10Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.264762 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:10Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.341512 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.341554 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.341565 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.341581 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.341592 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:10Z","lastTransitionTime":"2025-11-26T00:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.443897 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.444193 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.444205 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.444222 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.444233 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:10Z","lastTransitionTime":"2025-11-26T00:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.546237 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.546278 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.546292 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.546309 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.546320 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:10Z","lastTransitionTime":"2025-11-26T00:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.648915 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.648976 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.648988 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.649006 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.649018 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:10Z","lastTransitionTime":"2025-11-26T00:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.752215 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.752267 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.752279 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.752294 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.752306 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:10Z","lastTransitionTime":"2025-11-26T00:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.826269 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.826363 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:10 crc kubenswrapper[4766]: E1126 00:24:10.826417 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.826465 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:10 crc kubenswrapper[4766]: E1126 00:24:10.826507 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:10 crc kubenswrapper[4766]: E1126 00:24:10.826731 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.854545 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.854582 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.854595 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.854629 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.854641 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:10Z","lastTransitionTime":"2025-11-26T00:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.957335 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.957388 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.957404 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.957428 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.957442 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:10Z","lastTransitionTime":"2025-11-26T00:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.995417 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerStarted","Data":"f07ffa8c50c44fa2abfa4579ca7b53f16f8630f358f0375fc57920edfe771505"} Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.995840 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.995901 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:10 crc kubenswrapper[4766]: I1126 00:24:10.999876 4766 generic.go:334] "Generic (PLEG): container finished" podID="beeb3530-3b3b-40d9-8b0e-1b927c64920a" containerID="3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4" exitCode=0 Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:10.999938 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" event={"ID":"beeb3530-3b3b-40d9-8b0e-1b927c64920a","Type":"ContainerDied","Data":"3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4"} Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.010730 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.023471 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.034720 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.036297 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.036358 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.046204 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.057936 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.059345 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.059384 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.059393 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.059407 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.059416 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:11Z","lastTransitionTime":"2025-11-26T00:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.070238 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.094062 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.108051 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.119739 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.132168 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.144102 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.158679 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.161224 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.161259 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.161267 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.161281 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.161290 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:11Z","lastTransitionTime":"2025-11-26T00:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.175890 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f07ffa8c50c44fa2abfa4579ca7b53f16f8630f358f0375fc57920edfe771505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.186574 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.196113 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.207123 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.216206 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.227226 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.239578 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.251771 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.264348 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.265397 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.265464 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.265483 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.266488 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.266539 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:11Z","lastTransitionTime":"2025-11-26T00:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.275817 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.291961 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.315928 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.330210 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.347905 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.366185 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f07ffa8c50c44fa2abfa4579ca7b53f16f8630f358f0375fc57920edfe771505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.368629 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.368667 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.368679 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.368695 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.368706 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:11Z","lastTransitionTime":"2025-11-26T00:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.378107 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.388200 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.399290 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:11Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.471609 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.471661 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.471671 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.471686 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.471697 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:11Z","lastTransitionTime":"2025-11-26T00:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.574261 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.574316 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.574331 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.574348 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.574359 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:11Z","lastTransitionTime":"2025-11-26T00:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.677309 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.677345 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.677353 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.677369 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.677379 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:11Z","lastTransitionTime":"2025-11-26T00:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.779992 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.780034 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.780045 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.780060 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.780071 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:11Z","lastTransitionTime":"2025-11-26T00:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.882067 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.882148 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.882172 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.882201 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.882218 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:11Z","lastTransitionTime":"2025-11-26T00:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.984575 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.984610 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.984619 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.984635 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:11 crc kubenswrapper[4766]: I1126 00:24:11.984644 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:11Z","lastTransitionTime":"2025-11-26T00:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.007039 4766 generic.go:334] "Generic (PLEG): container finished" podID="beeb3530-3b3b-40d9-8b0e-1b927c64920a" containerID="27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b" exitCode=0 Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.007125 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" event={"ID":"beeb3530-3b3b-40d9-8b0e-1b927c64920a","Type":"ContainerDied","Data":"27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b"} Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.007224 4766 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.020160 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:12Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.032842 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:12Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.046504 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:12Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.060571 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:12Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.073058 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:12Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.084103 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:12Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.086870 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.086912 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.086920 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.086936 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.086944 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:12Z","lastTransitionTime":"2025-11-26T00:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.096174 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:12Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.109270 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:12Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.126592 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:12Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.137014 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:12Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.155777 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f07ffa8c50c44fa2abfa4579ca7b53f16f8630f358f0375fc57920edfe771505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:12Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.168978 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:12Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.180516 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:12Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.189120 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.189162 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.189174 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.189192 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.189204 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:12Z","lastTransitionTime":"2025-11-26T00:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.191013 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:12Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.205911 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:12Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.291227 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.291269 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.291282 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.291300 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.291313 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:12Z","lastTransitionTime":"2025-11-26T00:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.395415 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.395457 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.395468 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.395485 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.395496 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:12Z","lastTransitionTime":"2025-11-26T00:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.498668 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.498731 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.498749 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.498770 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.498786 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:12Z","lastTransitionTime":"2025-11-26T00:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.601203 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.601231 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.601238 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.601253 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.601262 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:12Z","lastTransitionTime":"2025-11-26T00:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.703708 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.703749 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.703763 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.703785 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.703798 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:12Z","lastTransitionTime":"2025-11-26T00:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.806202 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.806526 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.806538 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.806555 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.806567 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:12Z","lastTransitionTime":"2025-11-26T00:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.836849 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:12 crc kubenswrapper[4766]: E1126 00:24:12.836964 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.837260 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:12 crc kubenswrapper[4766]: E1126 00:24:12.837305 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.837339 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:12 crc kubenswrapper[4766]: E1126 00:24:12.837372 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.908894 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.908930 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.908939 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.908955 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:12 crc kubenswrapper[4766]: I1126 00:24:12.908964 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:12Z","lastTransitionTime":"2025-11-26T00:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.010575 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.010616 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.010628 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.010644 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.010677 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:13Z","lastTransitionTime":"2025-11-26T00:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.013129 4766 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.013842 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" event={"ID":"beeb3530-3b3b-40d9-8b0e-1b927c64920a","Type":"ContainerStarted","Data":"fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd"} Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.029042 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:13Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.044470 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:13Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.058333 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:13Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.071623 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:13Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.086537 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:13Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.099733 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:13Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.114377 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.114418 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.114429 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.114444 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.114454 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:13Z","lastTransitionTime":"2025-11-26T00:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.119995 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:13Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.131758 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:13Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.143511 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:13Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.153760 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:13Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.163842 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:13Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.180207 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:13Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.198297 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f07ffa8c50c44fa2abfa4579ca7b53f16f8630f358f0375fc57920edfe771505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:13Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.210098 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:13Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.216553 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.216586 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.216595 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.216608 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.216617 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:13Z","lastTransitionTime":"2025-11-26T00:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.222694 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:13Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.318633 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.318683 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.318692 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.318705 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.318715 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:13Z","lastTransitionTime":"2025-11-26T00:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.421848 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.421892 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.421909 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.421931 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.421986 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:13Z","lastTransitionTime":"2025-11-26T00:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.524059 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.524103 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.524115 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.524132 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.524144 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:13Z","lastTransitionTime":"2025-11-26T00:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.626556 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.626616 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.626630 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.626676 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.626691 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:13Z","lastTransitionTime":"2025-11-26T00:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.729172 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.729207 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.729220 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.729235 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.729246 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:13Z","lastTransitionTime":"2025-11-26T00:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.831784 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.831839 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.831855 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.831876 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:13 crc kubenswrapper[4766]: I1126 00:24:13.831894 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:13Z","lastTransitionTime":"2025-11-26T00:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.375904 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.375950 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.375959 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.375975 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.375984 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:14Z","lastTransitionTime":"2025-11-26T00:24:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.378942 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fxnln_415fdacb-85c1-4265-89b7-6771a84ffc89/ovnkube-controller/0.log" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.382166 4766 generic.go:334] "Generic (PLEG): container finished" podID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerID="f07ffa8c50c44fa2abfa4579ca7b53f16f8630f358f0375fc57920edfe771505" exitCode=1 Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.382209 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerDied","Data":"f07ffa8c50c44fa2abfa4579ca7b53f16f8630f358f0375fc57920edfe771505"} Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.382841 4766 scope.go:117] "RemoveContainer" containerID="f07ffa8c50c44fa2abfa4579ca7b53f16f8630f358f0375fc57920edfe771505" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.396631 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:14Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.406972 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:14Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.417913 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:14Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.433521 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:14Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.465363 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f07ffa8c50c44fa2abfa4579ca7b53f16f8630f358f0375fc57920edfe771505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07ffa8c50c44fa2abfa4579ca7b53f16f8630f358f0375fc57920edfe771505\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"message\\\":\\\" handler 8\\\\nI1126 00:24:13.651998 6045 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 00:24:13.652004 6045 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 00:24:13.652053 6045 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652140 6045 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652270 6045 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652350 6045 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1126 00:24:13.652644 6045 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 00:24:13.652675 6045 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 00:24:13.652694 6045 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 00:24:13.652709 6045 factory.go:656] Stopping watch factory\\\\nI1126 00:24:13.652717 6045 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 00:24:13.652732 6045 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 00:24:13.652725 6045 ovnkube.go:599] Stopped ovnkube\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:14Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.478326 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.478368 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.478378 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.478393 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.478416 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:14Z","lastTransitionTime":"2025-11-26T00:24:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.480064 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:14Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.481215 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:24:14 crc kubenswrapper[4766]: E1126 00:24:14.481363 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:24:30.481345416 +0000 UTC m=+51.330115846 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.481412 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.481465 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.481498 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:14 crc kubenswrapper[4766]: E1126 00:24:14.481537 4766 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 00:24:14 crc kubenswrapper[4766]: E1126 00:24:14.481590 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 00:24:14 crc kubenswrapper[4766]: E1126 00:24:14.481603 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 00:24:30.481588132 +0000 UTC m=+51.330358552 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 00:24:14 crc kubenswrapper[4766]: E1126 00:24:14.481606 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 00:24:14 crc kubenswrapper[4766]: E1126 00:24:14.482841 4766 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:24:14 crc kubenswrapper[4766]: E1126 00:24:14.482920 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 00:24:30.482896765 +0000 UTC m=+51.331667195 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:24:14 crc kubenswrapper[4766]: E1126 00:24:14.481634 4766 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 00:24:14 crc kubenswrapper[4766]: E1126 00:24:14.483222 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 00:24:30.483198613 +0000 UTC m=+51.331969053 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.492002 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:14Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.504259 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:14Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.515999 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:14Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.526729 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:14Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.537477 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:14Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.551069 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:14Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.569777 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:14Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.580432 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.580468 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.580480 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.580496 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.580508 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:14Z","lastTransitionTime":"2025-11-26T00:24:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.583316 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:14 crc kubenswrapper[4766]: E1126 00:24:14.583504 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 00:24:14 crc kubenswrapper[4766]: E1126 00:24:14.583535 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 00:24:14 crc kubenswrapper[4766]: E1126 00:24:14.583553 4766 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:24:14 crc kubenswrapper[4766]: E1126 00:24:14.583628 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 00:24:30.583606459 +0000 UTC m=+51.432376929 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.592310 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:14Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.603758 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:14Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.682640 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.682691 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.682701 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.682716 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.682729 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:14Z","lastTransitionTime":"2025-11-26T00:24:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.784745 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.784789 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.784801 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.784819 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.784840 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:14Z","lastTransitionTime":"2025-11-26T00:24:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.825729 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.825739 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:14 crc kubenswrapper[4766]: E1126 00:24:14.825935 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:14 crc kubenswrapper[4766]: E1126 00:24:14.825852 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.825752 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:14 crc kubenswrapper[4766]: E1126 00:24:14.826013 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.887932 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.887966 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.887977 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.887990 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.888000 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:14Z","lastTransitionTime":"2025-11-26T00:24:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.990625 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.990685 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.990694 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.990709 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:14 crc kubenswrapper[4766]: I1126 00:24:14.990719 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:14Z","lastTransitionTime":"2025-11-26T00:24:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.093011 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.093047 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.093057 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.093072 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.093082 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:15Z","lastTransitionTime":"2025-11-26T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.194971 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.195071 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.195131 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.195156 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.195175 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:15Z","lastTransitionTime":"2025-11-26T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.285611 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.285675 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.285686 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.285703 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.285715 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:15Z","lastTransitionTime":"2025-11-26T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:15 crc kubenswrapper[4766]: E1126 00:24:15.299070 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:15Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.303715 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.303754 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.303765 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.303781 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.303792 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:15Z","lastTransitionTime":"2025-11-26T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:15 crc kubenswrapper[4766]: E1126 00:24:15.320901 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:15Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.330833 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.330882 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.330894 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.330914 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.330928 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:15Z","lastTransitionTime":"2025-11-26T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:15 crc kubenswrapper[4766]: E1126 00:24:15.350700 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:15Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.354795 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.354831 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.354840 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.354856 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.354868 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:15Z","lastTransitionTime":"2025-11-26T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:15 crc kubenswrapper[4766]: E1126 00:24:15.371361 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:15Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.374404 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.374448 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.374460 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.374476 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.374486 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:15Z","lastTransitionTime":"2025-11-26T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.386664 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fxnln_415fdacb-85c1-4265-89b7-6771a84ffc89/ovnkube-controller/0.log" Nov 26 00:24:15 crc kubenswrapper[4766]: E1126 00:24:15.387551 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:15Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:15 crc kubenswrapper[4766]: E1126 00:24:15.387699 4766 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.388786 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.388816 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.388827 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.388840 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.388850 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:15Z","lastTransitionTime":"2025-11-26T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.388914 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerStarted","Data":"7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945"} Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.389039 4766 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.404373 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:15Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.415105 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:15Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.427411 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:15Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.439494 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:15Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.448377 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:15Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.458591 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:15Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.471358 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:15Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.488491 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:15Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.492104 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.492131 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.492142 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.492157 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.492166 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:15Z","lastTransitionTime":"2025-11-26T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.499527 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:15Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.511435 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:15Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.522052 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:15Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.531833 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:15Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.541834 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:15Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.554865 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:15Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.572235 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07ffa8c50c44fa2abfa4579ca7b53f16f8630f358f0375fc57920edfe771505\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"message\\\":\\\" handler 8\\\\nI1126 00:24:13.651998 6045 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 00:24:13.652004 6045 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 00:24:13.652053 6045 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652140 6045 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652270 6045 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652350 6045 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1126 00:24:13.652644 6045 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 00:24:13.652675 6045 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 00:24:13.652694 6045 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 00:24:13.652709 6045 factory.go:656] Stopping watch factory\\\\nI1126 00:24:13.652717 6045 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 00:24:13.652732 6045 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 00:24:13.652725 6045 ovnkube.go:599] Stopped ovnkube\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:15Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.594065 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.594116 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.594129 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.594148 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.594163 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:15Z","lastTransitionTime":"2025-11-26T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.696857 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.697164 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.697405 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.697631 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.697864 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:15Z","lastTransitionTime":"2025-11-26T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.800536 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.801567 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.801773 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.801913 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.802046 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:15Z","lastTransitionTime":"2025-11-26T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.904298 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.904346 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.904354 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.904370 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:15 crc kubenswrapper[4766]: I1126 00:24:15.904383 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:15Z","lastTransitionTime":"2025-11-26T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.007420 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.007466 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.007478 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.007494 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.007506 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:16Z","lastTransitionTime":"2025-11-26T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.109496 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.109540 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.109549 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.109562 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.109570 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:16Z","lastTransitionTime":"2025-11-26T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.211944 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.211988 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.211998 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.212013 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.212023 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:16Z","lastTransitionTime":"2025-11-26T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.314349 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.314396 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.314411 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.314431 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.314445 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:16Z","lastTransitionTime":"2025-11-26T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.396631 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fxnln_415fdacb-85c1-4265-89b7-6771a84ffc89/ovnkube-controller/1.log" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.397778 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fxnln_415fdacb-85c1-4265-89b7-6771a84ffc89/ovnkube-controller/0.log" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.401747 4766 generic.go:334] "Generic (PLEG): container finished" podID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerID="7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945" exitCode=1 Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.401812 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerDied","Data":"7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945"} Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.401897 4766 scope.go:117] "RemoveContainer" containerID="f07ffa8c50c44fa2abfa4579ca7b53f16f8630f358f0375fc57920edfe771505" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.402811 4766 scope.go:117] "RemoveContainer" containerID="7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945" Nov 26 00:24:16 crc kubenswrapper[4766]: E1126 00:24:16.403066 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fxnln_openshift-ovn-kubernetes(415fdacb-85c1-4265-89b7-6771a84ffc89)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.416978 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.417015 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.417025 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.417042 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.417052 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:16Z","lastTransitionTime":"2025-11-26T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.421129 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:16Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.454977 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:16Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.468450 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:16Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.483070 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:16Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.503023 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:16Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.518340 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:16Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.519794 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.519929 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.519993 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.520078 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.520139 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:16Z","lastTransitionTime":"2025-11-26T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.537501 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:16Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.562465 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07ffa8c50c44fa2abfa4579ca7b53f16f8630f358f0375fc57920edfe771505\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"message\\\":\\\" handler 8\\\\nI1126 00:24:13.651998 6045 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 00:24:13.652004 6045 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 00:24:13.652053 6045 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652140 6045 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652270 6045 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652350 6045 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1126 00:24:13.652644 6045 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 00:24:13.652675 6045 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 00:24:13.652694 6045 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 00:24:13.652709 6045 factory.go:656] Stopping watch factory\\\\nI1126 00:24:13.652717 6045 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 00:24:13.652732 6045 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 00:24:13.652725 6045 ovnkube.go:599] Stopped ovnkube\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"423049 6211 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 00:24:15.423069 6211 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1126 00:24:15.423443 6211 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1126 00:24:15.423455 6211 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1126 00:24:15.423492 6211 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1126 00:24:15.423506 6211 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 00:24:15.423508 6211 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 00:24:15.423518 6211 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1126 00:24:15.423520 6211 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 00:24:15.423530 6211 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 00:24:15.423557 6211 factory.go:656] Stopping watch factory\\\\nI1126 00:24:15.423569 6211 ovnkube.go:599] Stopped ovnkube\\\\nI1126 00:24:15.423580 6211 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 00:24:15.423604 6211 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1126 00:24:15.423687 6211 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:16Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.580249 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:16Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.595384 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:16Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.612442 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:16Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.623044 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.623095 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.623114 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.623135 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.623148 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:16Z","lastTransitionTime":"2025-11-26T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.626388 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:16Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.640227 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:16Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.653459 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:16Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.664544 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:16Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.726352 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.726418 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.726432 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.726451 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.726466 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:16Z","lastTransitionTime":"2025-11-26T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.825720 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.825849 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.825863 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:16 crc kubenswrapper[4766]: E1126 00:24:16.825984 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:16 crc kubenswrapper[4766]: E1126 00:24:16.826094 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:16 crc kubenswrapper[4766]: E1126 00:24:16.826281 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.829334 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.829367 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.829376 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.829391 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.829401 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:16Z","lastTransitionTime":"2025-11-26T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.932685 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.932731 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.932743 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.932765 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:16 crc kubenswrapper[4766]: I1126 00:24:16.932803 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:16Z","lastTransitionTime":"2025-11-26T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.035171 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.035228 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.035246 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.035267 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.035284 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:17Z","lastTransitionTime":"2025-11-26T00:24:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.137732 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.137806 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.137844 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.137879 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.137903 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:17Z","lastTransitionTime":"2025-11-26T00:24:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.159475 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc"] Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.160019 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.163064 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.163546 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.183298 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.193287 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.211743 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.224448 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.237976 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.239941 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.239970 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.239978 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.239992 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.240002 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:17Z","lastTransitionTime":"2025-11-26T00:24:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.250276 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.263560 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.277789 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.299675 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.311269 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.312685 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/443534cf-37a7-4831-a581-9d04dee05ead-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-49ztc\" (UID: \"443534cf-37a7-4831-a581-9d04dee05ead\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.312738 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/443534cf-37a7-4831-a581-9d04dee05ead-env-overrides\") pod \"ovnkube-control-plane-749d76644c-49ztc\" (UID: \"443534cf-37a7-4831-a581-9d04dee05ead\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.312782 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/443534cf-37a7-4831-a581-9d04dee05ead-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-49ztc\" (UID: \"443534cf-37a7-4831-a581-9d04dee05ead\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.312903 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcwjf\" (UniqueName: \"kubernetes.io/projected/443534cf-37a7-4831-a581-9d04dee05ead-kube-api-access-pcwjf\") pod \"ovnkube-control-plane-749d76644c-49ztc\" (UID: \"443534cf-37a7-4831-a581-9d04dee05ead\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.322468 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"443534cf-37a7-4831-a581-9d04dee05ead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-49ztc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.335695 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.342141 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.342193 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.342204 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.342224 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.342286 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:17Z","lastTransitionTime":"2025-11-26T00:24:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.350273 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.363635 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.380404 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.397693 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07ffa8c50c44fa2abfa4579ca7b53f16f8630f358f0375fc57920edfe771505\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"message\\\":\\\" handler 8\\\\nI1126 00:24:13.651998 6045 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 00:24:13.652004 6045 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 00:24:13.652053 6045 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652140 6045 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652270 6045 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652350 6045 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1126 00:24:13.652644 6045 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 00:24:13.652675 6045 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 00:24:13.652694 6045 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 00:24:13.652709 6045 factory.go:656] Stopping watch factory\\\\nI1126 00:24:13.652717 6045 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 00:24:13.652732 6045 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 00:24:13.652725 6045 ovnkube.go:599] Stopped ovnkube\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"423049 6211 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 00:24:15.423069 6211 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1126 00:24:15.423443 6211 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1126 00:24:15.423455 6211 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1126 00:24:15.423492 6211 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1126 00:24:15.423506 6211 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 00:24:15.423508 6211 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 00:24:15.423518 6211 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1126 00:24:15.423520 6211 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 00:24:15.423530 6211 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 00:24:15.423557 6211 factory.go:656] Stopping watch factory\\\\nI1126 00:24:15.423569 6211 ovnkube.go:599] Stopped ovnkube\\\\nI1126 00:24:15.423580 6211 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 00:24:15.423604 6211 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1126 00:24:15.423687 6211 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.405961 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fxnln_415fdacb-85c1-4265-89b7-6771a84ffc89/ovnkube-controller/1.log" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.413615 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/443534cf-37a7-4831-a581-9d04dee05ead-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-49ztc\" (UID: \"443534cf-37a7-4831-a581-9d04dee05ead\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.413672 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/443534cf-37a7-4831-a581-9d04dee05ead-env-overrides\") pod \"ovnkube-control-plane-749d76644c-49ztc\" (UID: \"443534cf-37a7-4831-a581-9d04dee05ead\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.413698 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/443534cf-37a7-4831-a581-9d04dee05ead-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-49ztc\" (UID: \"443534cf-37a7-4831-a581-9d04dee05ead\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.413727 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcwjf\" (UniqueName: \"kubernetes.io/projected/443534cf-37a7-4831-a581-9d04dee05ead-kube-api-access-pcwjf\") pod \"ovnkube-control-plane-749d76644c-49ztc\" (UID: \"443534cf-37a7-4831-a581-9d04dee05ead\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.414326 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/443534cf-37a7-4831-a581-9d04dee05ead-env-overrides\") pod \"ovnkube-control-plane-749d76644c-49ztc\" (UID: \"443534cf-37a7-4831-a581-9d04dee05ead\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.414446 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/443534cf-37a7-4831-a581-9d04dee05ead-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-49ztc\" (UID: \"443534cf-37a7-4831-a581-9d04dee05ead\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.419926 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/443534cf-37a7-4831-a581-9d04dee05ead-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-49ztc\" (UID: \"443534cf-37a7-4831-a581-9d04dee05ead\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.428575 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcwjf\" (UniqueName: \"kubernetes.io/projected/443534cf-37a7-4831-a581-9d04dee05ead-kube-api-access-pcwjf\") pod \"ovnkube-control-plane-749d76644c-49ztc\" (UID: \"443534cf-37a7-4831-a581-9d04dee05ead\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.444236 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.444272 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.444280 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.444295 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.444306 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:17Z","lastTransitionTime":"2025-11-26T00:24:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.445778 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.457083 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.467117 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.472373 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.476856 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: W1126 00:24:17.488813 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod443534cf_37a7_4831_a581_9d04dee05ead.slice/crio-92da230303eac62e2b0ce6bd2111a648adb6b08c5ef7bd29e1ad9ba35ae69ca1 WatchSource:0}: Error finding container 92da230303eac62e2b0ce6bd2111a648adb6b08c5ef7bd29e1ad9ba35ae69ca1: Status 404 returned error can't find the container with id 92da230303eac62e2b0ce6bd2111a648adb6b08c5ef7bd29e1ad9ba35ae69ca1 Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.491105 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.505674 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.526264 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.536246 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.545149 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"443534cf-37a7-4831-a581-9d04dee05ead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-49ztc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.546393 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.546426 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.546438 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.546458 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.546469 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:17Z","lastTransitionTime":"2025-11-26T00:24:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.557564 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.568617 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.578808 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.589595 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.602771 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.620387 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07ffa8c50c44fa2abfa4579ca7b53f16f8630f358f0375fc57920edfe771505\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"message\\\":\\\" handler 8\\\\nI1126 00:24:13.651998 6045 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 00:24:13.652004 6045 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 00:24:13.652053 6045 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652140 6045 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652270 6045 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652350 6045 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1126 00:24:13.652644 6045 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 00:24:13.652675 6045 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 00:24:13.652694 6045 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 00:24:13.652709 6045 factory.go:656] Stopping watch factory\\\\nI1126 00:24:13.652717 6045 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 00:24:13.652732 6045 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 00:24:13.652725 6045 ovnkube.go:599] Stopped ovnkube\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"423049 6211 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 00:24:15.423069 6211 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1126 00:24:15.423443 6211 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1126 00:24:15.423455 6211 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1126 00:24:15.423492 6211 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1126 00:24:15.423506 6211 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 00:24:15.423508 6211 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 00:24:15.423518 6211 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1126 00:24:15.423520 6211 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 00:24:15.423530 6211 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 00:24:15.423557 6211 factory.go:656] Stopping watch factory\\\\nI1126 00:24:15.423569 6211 ovnkube.go:599] Stopped ovnkube\\\\nI1126 00:24:15.423580 6211 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 00:24:15.423604 6211 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1126 00:24:15.423687 6211 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.632750 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.643218 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:17Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.648974 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.649086 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.649113 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.649130 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.649141 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:17Z","lastTransitionTime":"2025-11-26T00:24:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.751758 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.751807 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.751817 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.751832 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.751842 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:17Z","lastTransitionTime":"2025-11-26T00:24:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.854087 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.854126 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.854137 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.854152 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.854164 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:17Z","lastTransitionTime":"2025-11-26T00:24:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.956618 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.956706 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.956722 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.956743 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:17 crc kubenswrapper[4766]: I1126 00:24:17.956759 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:17Z","lastTransitionTime":"2025-11-26T00:24:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.059410 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.059470 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.059484 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.059503 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.059514 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:18Z","lastTransitionTime":"2025-11-26T00:24:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.161871 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.161930 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.161944 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.161964 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.161976 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:18Z","lastTransitionTime":"2025-11-26T00:24:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.235385 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-5z9ll"] Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.236012 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:18 crc kubenswrapper[4766]: E1126 00:24:18.236087 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.249355 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.264363 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.264407 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.264418 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.264435 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.264448 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:18Z","lastTransitionTime":"2025-11-26T00:24:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.264626 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.280415 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.290697 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.305299 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.319477 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5z9ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6600efa-4858-41aa-a890-44ce262b63dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5z9ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.331711 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.352362 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.363596 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.367151 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.367188 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.367198 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.367212 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.367223 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:18Z","lastTransitionTime":"2025-11-26T00:24:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.375521 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"443534cf-37a7-4831-a581-9d04dee05ead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-49ztc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.386758 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.399054 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.412800 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" event={"ID":"443534cf-37a7-4831-a581-9d04dee05ead","Type":"ContainerStarted","Data":"fb484d4736b942d10794d5e39064bec67b5b30e9d1ac668f5a62dbc15119c3eb"} Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.412852 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" event={"ID":"443534cf-37a7-4831-a581-9d04dee05ead","Type":"ContainerStarted","Data":"0874e594efd8c1df4af245ea08256417c1b6cfe83fd8ce08848ea6710081cf1e"} Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.412862 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" event={"ID":"443534cf-37a7-4831-a581-9d04dee05ead","Type":"ContainerStarted","Data":"92da230303eac62e2b0ce6bd2111a648adb6b08c5ef7bd29e1ad9ba35ae69ca1"} Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.413614 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.425473 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs\") pod \"network-metrics-daemon-5z9ll\" (UID: \"b6600efa-4858-41aa-a890-44ce262b63dd\") " pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.425533 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnqtx\" (UniqueName: \"kubernetes.io/projected/b6600efa-4858-41aa-a890-44ce262b63dd-kube-api-access-dnqtx\") pod \"network-metrics-daemon-5z9ll\" (UID: \"b6600efa-4858-41aa-a890-44ce262b63dd\") " pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.427933 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.447286 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07ffa8c50c44fa2abfa4579ca7b53f16f8630f358f0375fc57920edfe771505\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"message\\\":\\\" handler 8\\\\nI1126 00:24:13.651998 6045 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 00:24:13.652004 6045 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 00:24:13.652053 6045 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652140 6045 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652270 6045 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652350 6045 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1126 00:24:13.652644 6045 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 00:24:13.652675 6045 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 00:24:13.652694 6045 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 00:24:13.652709 6045 factory.go:656] Stopping watch factory\\\\nI1126 00:24:13.652717 6045 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 00:24:13.652732 6045 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 00:24:13.652725 6045 ovnkube.go:599] Stopped ovnkube\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"423049 6211 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 00:24:15.423069 6211 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1126 00:24:15.423443 6211 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1126 00:24:15.423455 6211 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1126 00:24:15.423492 6211 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1126 00:24:15.423506 6211 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 00:24:15.423508 6211 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 00:24:15.423518 6211 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1126 00:24:15.423520 6211 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 00:24:15.423530 6211 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 00:24:15.423557 6211 factory.go:656] Stopping watch factory\\\\nI1126 00:24:15.423569 6211 ovnkube.go:599] Stopped ovnkube\\\\nI1126 00:24:15.423580 6211 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 00:24:15.423604 6211 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1126 00:24:15.423687 6211 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.460499 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.469331 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.469369 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.469378 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.469392 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.469405 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:18Z","lastTransitionTime":"2025-11-26T00:24:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.470672 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.481338 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.491160 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.503192 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.512629 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.525113 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.527074 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnqtx\" (UniqueName: \"kubernetes.io/projected/b6600efa-4858-41aa-a890-44ce262b63dd-kube-api-access-dnqtx\") pod \"network-metrics-daemon-5z9ll\" (UID: \"b6600efa-4858-41aa-a890-44ce262b63dd\") " pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.527787 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs\") pod \"network-metrics-daemon-5z9ll\" (UID: \"b6600efa-4858-41aa-a890-44ce262b63dd\") " pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:18 crc kubenswrapper[4766]: E1126 00:24:18.527971 4766 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 00:24:18 crc kubenswrapper[4766]: E1126 00:24:18.528038 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs podName:b6600efa-4858-41aa-a890-44ce262b63dd nodeName:}" failed. No retries permitted until 2025-11-26 00:24:19.028018168 +0000 UTC m=+39.876788598 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs") pod "network-metrics-daemon-5z9ll" (UID: "b6600efa-4858-41aa-a890-44ce262b63dd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.536508 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5z9ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6600efa-4858-41aa-a890-44ce262b63dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5z9ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.544086 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnqtx\" (UniqueName: \"kubernetes.io/projected/b6600efa-4858-41aa-a890-44ce262b63dd-kube-api-access-dnqtx\") pod \"network-metrics-daemon-5z9ll\" (UID: \"b6600efa-4858-41aa-a890-44ce262b63dd\") " pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.551853 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.567848 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.573722 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.573779 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.573792 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.573811 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.573826 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:18Z","lastTransitionTime":"2025-11-26T00:24:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.587060 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.596714 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"443534cf-37a7-4831-a581-9d04dee05ead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0874e594efd8c1df4af245ea08256417c1b6cfe83fd8ce08848ea6710081cf1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb484d4736b942d10794d5e39064bec67b5b30e9d1ac668f5a62dbc15119c3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-49ztc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.629214 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.640514 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.651750 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.666153 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.675824 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.675867 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.675878 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.675895 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.675907 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:18Z","lastTransitionTime":"2025-11-26T00:24:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.685598 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07ffa8c50c44fa2abfa4579ca7b53f16f8630f358f0375fc57920edfe771505\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"message\\\":\\\" handler 8\\\\nI1126 00:24:13.651998 6045 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 00:24:13.652004 6045 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 00:24:13.652053 6045 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652140 6045 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652270 6045 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652350 6045 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1126 00:24:13.652644 6045 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 00:24:13.652675 6045 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 00:24:13.652694 6045 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 00:24:13.652709 6045 factory.go:656] Stopping watch factory\\\\nI1126 00:24:13.652717 6045 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 00:24:13.652732 6045 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 00:24:13.652725 6045 ovnkube.go:599] Stopped ovnkube\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"423049 6211 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 00:24:15.423069 6211 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1126 00:24:15.423443 6211 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1126 00:24:15.423455 6211 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1126 00:24:15.423492 6211 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1126 00:24:15.423506 6211 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 00:24:15.423508 6211 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 00:24:15.423518 6211 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1126 00:24:15.423520 6211 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 00:24:15.423530 6211 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 00:24:15.423557 6211 factory.go:656] Stopping watch factory\\\\nI1126 00:24:15.423569 6211 ovnkube.go:599] Stopped ovnkube\\\\nI1126 00:24:15.423580 6211 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 00:24:15.423604 6211 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1126 00:24:15.423687 6211 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.697726 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.708466 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:18Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.779535 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.779621 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.779640 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.779704 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.779730 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:18Z","lastTransitionTime":"2025-11-26T00:24:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.826160 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.826187 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:18 crc kubenswrapper[4766]: E1126 00:24:18.826383 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:18 crc kubenswrapper[4766]: E1126 00:24:18.826475 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.826743 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:18 crc kubenswrapper[4766]: E1126 00:24:18.826996 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.882805 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.882881 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.882942 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.883101 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.883130 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:18Z","lastTransitionTime":"2025-11-26T00:24:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.985838 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.985894 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.985911 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.985934 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:18 crc kubenswrapper[4766]: I1126 00:24:18.985950 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:18Z","lastTransitionTime":"2025-11-26T00:24:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.033082 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs\") pod \"network-metrics-daemon-5z9ll\" (UID: \"b6600efa-4858-41aa-a890-44ce262b63dd\") " pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:19 crc kubenswrapper[4766]: E1126 00:24:19.033250 4766 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 00:24:19 crc kubenswrapper[4766]: E1126 00:24:19.033330 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs podName:b6600efa-4858-41aa-a890-44ce262b63dd nodeName:}" failed. No retries permitted until 2025-11-26 00:24:20.033310725 +0000 UTC m=+40.882081175 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs") pod "network-metrics-daemon-5z9ll" (UID: "b6600efa-4858-41aa-a890-44ce262b63dd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.089437 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.089494 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.089529 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.089547 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.089563 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:19Z","lastTransitionTime":"2025-11-26T00:24:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.192967 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.193009 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.193017 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.193031 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.193040 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:19Z","lastTransitionTime":"2025-11-26T00:24:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.296112 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.296181 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.296203 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.296234 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.296258 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:19Z","lastTransitionTime":"2025-11-26T00:24:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.398972 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.399019 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.399027 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.399044 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.399054 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:19Z","lastTransitionTime":"2025-11-26T00:24:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.501680 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.501731 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.501740 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.501757 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.501771 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:19Z","lastTransitionTime":"2025-11-26T00:24:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.604494 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.604537 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.604548 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.604563 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.604574 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:19Z","lastTransitionTime":"2025-11-26T00:24:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.706982 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.707020 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.707029 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.707044 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.707053 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:19Z","lastTransitionTime":"2025-11-26T00:24:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.809446 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.809547 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.809564 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.809582 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.809596 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:19Z","lastTransitionTime":"2025-11-26T00:24:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.826170 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:19 crc kubenswrapper[4766]: E1126 00:24:19.826319 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.840834 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:19Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.853707 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:19Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.867734 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:19Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.882845 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:19Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.907084 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07ffa8c50c44fa2abfa4579ca7b53f16f8630f358f0375fc57920edfe771505\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"message\\\":\\\" handler 8\\\\nI1126 00:24:13.651998 6045 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 00:24:13.652004 6045 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 00:24:13.652053 6045 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652140 6045 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652270 6045 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652350 6045 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1126 00:24:13.652644 6045 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 00:24:13.652675 6045 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 00:24:13.652694 6045 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 00:24:13.652709 6045 factory.go:656] Stopping watch factory\\\\nI1126 00:24:13.652717 6045 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 00:24:13.652732 6045 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 00:24:13.652725 6045 ovnkube.go:599] Stopped ovnkube\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"423049 6211 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 00:24:15.423069 6211 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1126 00:24:15.423443 6211 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1126 00:24:15.423455 6211 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1126 00:24:15.423492 6211 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1126 00:24:15.423506 6211 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 00:24:15.423508 6211 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 00:24:15.423518 6211 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1126 00:24:15.423520 6211 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 00:24:15.423530 6211 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 00:24:15.423557 6211 factory.go:656] Stopping watch factory\\\\nI1126 00:24:15.423569 6211 ovnkube.go:599] Stopped ovnkube\\\\nI1126 00:24:15.423580 6211 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 00:24:15.423604 6211 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1126 00:24:15.423687 6211 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:19Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.912212 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.912551 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.912632 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.912729 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.912795 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:19Z","lastTransitionTime":"2025-11-26T00:24:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.921925 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:19Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.933437 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:19Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.945134 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:19Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.960147 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:19Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.971857 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:19Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.985841 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:19Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:19 crc kubenswrapper[4766]: I1126 00:24:19.996246 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5z9ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6600efa-4858-41aa-a890-44ce262b63dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5z9ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:19Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.008153 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:20Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.015300 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.015385 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.015398 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.015418 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.015429 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:20Z","lastTransitionTime":"2025-11-26T00:24:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.028128 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:20Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.038514 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:20Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.042025 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs\") pod \"network-metrics-daemon-5z9ll\" (UID: \"b6600efa-4858-41aa-a890-44ce262b63dd\") " pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:20 crc kubenswrapper[4766]: E1126 00:24:20.042162 4766 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 00:24:20 crc kubenswrapper[4766]: E1126 00:24:20.042223 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs podName:b6600efa-4858-41aa-a890-44ce262b63dd nodeName:}" failed. No retries permitted until 2025-11-26 00:24:22.042206527 +0000 UTC m=+42.890976957 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs") pod "network-metrics-daemon-5z9ll" (UID: "b6600efa-4858-41aa-a890-44ce262b63dd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.048544 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"443534cf-37a7-4831-a581-9d04dee05ead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0874e594efd8c1df4af245ea08256417c1b6cfe83fd8ce08848ea6710081cf1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb484d4736b942d10794d5e39064bec67b5b30e9d1ac668f5a62dbc15119c3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-49ztc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:20Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.062851 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:20Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.117341 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.117388 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.117402 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.117420 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.117433 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:20Z","lastTransitionTime":"2025-11-26T00:24:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.220063 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.220117 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.220131 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.220162 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.220175 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:20Z","lastTransitionTime":"2025-11-26T00:24:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.323054 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.323103 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.323113 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.323129 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.323140 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:20Z","lastTransitionTime":"2025-11-26T00:24:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.425378 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.425424 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.425435 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.425453 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.425466 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:20Z","lastTransitionTime":"2025-11-26T00:24:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.528731 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.528774 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.528783 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.528800 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.528808 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:20Z","lastTransitionTime":"2025-11-26T00:24:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.631170 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.631232 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.631242 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.631256 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.631265 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:20Z","lastTransitionTime":"2025-11-26T00:24:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.733315 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.733351 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.733359 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.733374 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.733383 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:20Z","lastTransitionTime":"2025-11-26T00:24:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.825898 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.825964 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.825906 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:20 crc kubenswrapper[4766]: E1126 00:24:20.826063 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:20 crc kubenswrapper[4766]: E1126 00:24:20.826121 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:20 crc kubenswrapper[4766]: E1126 00:24:20.826336 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.835112 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.835164 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.835175 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.835202 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.835214 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:20Z","lastTransitionTime":"2025-11-26T00:24:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.937675 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.937742 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.937765 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.937792 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:20 crc kubenswrapper[4766]: I1126 00:24:20.937812 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:20Z","lastTransitionTime":"2025-11-26T00:24:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.040719 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.040791 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.040828 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.040858 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.040879 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:21Z","lastTransitionTime":"2025-11-26T00:24:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.142732 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.142777 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.142792 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.142809 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.142819 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:21Z","lastTransitionTime":"2025-11-26T00:24:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.245763 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.245816 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.245827 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.245845 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.245856 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:21Z","lastTransitionTime":"2025-11-26T00:24:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.348559 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.348600 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.348612 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.348632 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.348668 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:21Z","lastTransitionTime":"2025-11-26T00:24:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.451339 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.451412 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.451436 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.451467 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.451493 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:21Z","lastTransitionTime":"2025-11-26T00:24:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.553577 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.553658 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.553672 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.553690 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.553702 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:21Z","lastTransitionTime":"2025-11-26T00:24:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.656168 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.656203 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.656212 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.656226 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.656238 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:21Z","lastTransitionTime":"2025-11-26T00:24:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.758448 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.758495 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.758506 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.758522 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.758532 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:21Z","lastTransitionTime":"2025-11-26T00:24:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.826209 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:21 crc kubenswrapper[4766]: E1126 00:24:21.826696 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.861377 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.861433 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.861445 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.861462 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.861473 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:21Z","lastTransitionTime":"2025-11-26T00:24:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.963875 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.963925 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.963948 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.963971 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:21 crc kubenswrapper[4766]: I1126 00:24:21.963987 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:21Z","lastTransitionTime":"2025-11-26T00:24:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.059475 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs\") pod \"network-metrics-daemon-5z9ll\" (UID: \"b6600efa-4858-41aa-a890-44ce262b63dd\") " pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:22 crc kubenswrapper[4766]: E1126 00:24:22.059618 4766 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 00:24:22 crc kubenswrapper[4766]: E1126 00:24:22.059729 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs podName:b6600efa-4858-41aa-a890-44ce262b63dd nodeName:}" failed. No retries permitted until 2025-11-26 00:24:26.059704704 +0000 UTC m=+46.908475144 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs") pod "network-metrics-daemon-5z9ll" (UID: "b6600efa-4858-41aa-a890-44ce262b63dd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.065859 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.065889 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.065897 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.065910 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.065919 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:22Z","lastTransitionTime":"2025-11-26T00:24:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.167969 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.168017 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.168032 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.168057 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.168075 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:22Z","lastTransitionTime":"2025-11-26T00:24:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.271606 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.271666 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.271683 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.271699 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.271711 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:22Z","lastTransitionTime":"2025-11-26T00:24:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.374114 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.374148 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.374160 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.374172 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.374181 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:22Z","lastTransitionTime":"2025-11-26T00:24:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.476055 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.476093 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.476104 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.476120 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.476133 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:22Z","lastTransitionTime":"2025-11-26T00:24:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.578335 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.578369 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.578380 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.578397 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.578407 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:22Z","lastTransitionTime":"2025-11-26T00:24:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.681422 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.681484 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.681500 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.681520 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.681535 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:22Z","lastTransitionTime":"2025-11-26T00:24:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.784797 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.784864 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.784882 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.784906 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.784922 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:22Z","lastTransitionTime":"2025-11-26T00:24:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.826483 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.826573 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:22 crc kubenswrapper[4766]: E1126 00:24:22.826965 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.826598 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:22 crc kubenswrapper[4766]: E1126 00:24:22.826836 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:22 crc kubenswrapper[4766]: E1126 00:24:22.827087 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.887728 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.888032 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.888157 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.888434 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.888540 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:22Z","lastTransitionTime":"2025-11-26T00:24:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.991155 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.991789 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.991811 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.991828 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:22 crc kubenswrapper[4766]: I1126 00:24:22.991840 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:22Z","lastTransitionTime":"2025-11-26T00:24:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.094465 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.094502 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.094510 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.094523 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.094532 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:23Z","lastTransitionTime":"2025-11-26T00:24:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.196394 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.196457 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.196482 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.196497 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.196508 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:23Z","lastTransitionTime":"2025-11-26T00:24:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.298354 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.298397 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.298407 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.298423 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.298434 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:23Z","lastTransitionTime":"2025-11-26T00:24:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.400508 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.400554 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.400566 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.400584 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.400596 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:23Z","lastTransitionTime":"2025-11-26T00:24:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.502746 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.502794 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.502812 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.502830 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.502839 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:23Z","lastTransitionTime":"2025-11-26T00:24:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.606096 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.606166 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.606186 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.606209 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.606225 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:23Z","lastTransitionTime":"2025-11-26T00:24:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.708336 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.708377 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.708386 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.708400 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.708409 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:23Z","lastTransitionTime":"2025-11-26T00:24:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.810945 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.811002 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.811020 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.811041 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.811054 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:23Z","lastTransitionTime":"2025-11-26T00:24:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.825937 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:23 crc kubenswrapper[4766]: E1126 00:24:23.826203 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.914109 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.914159 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.914171 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.914193 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:23 crc kubenswrapper[4766]: I1126 00:24:23.914208 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:23Z","lastTransitionTime":"2025-11-26T00:24:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.016796 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.016841 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.016850 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.016865 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.016876 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:24Z","lastTransitionTime":"2025-11-26T00:24:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.119468 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.119566 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.119584 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.119609 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.119626 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:24Z","lastTransitionTime":"2025-11-26T00:24:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.221415 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.221450 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.221458 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.221470 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.221479 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:24Z","lastTransitionTime":"2025-11-26T00:24:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.323596 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.323643 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.323677 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.323695 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.323707 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:24Z","lastTransitionTime":"2025-11-26T00:24:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.426047 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.426091 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.426102 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.426120 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.426132 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:24Z","lastTransitionTime":"2025-11-26T00:24:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.529732 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.529779 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.529793 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.529810 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.529821 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:24Z","lastTransitionTime":"2025-11-26T00:24:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.632500 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.632553 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.632563 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.632579 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.632591 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:24Z","lastTransitionTime":"2025-11-26T00:24:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.735248 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.735310 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.735324 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.735346 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.735361 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:24Z","lastTransitionTime":"2025-11-26T00:24:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.826317 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.826372 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.826409 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:24 crc kubenswrapper[4766]: E1126 00:24:24.826466 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:24 crc kubenswrapper[4766]: E1126 00:24:24.826556 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:24 crc kubenswrapper[4766]: E1126 00:24:24.826689 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.838006 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.838051 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.838063 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.838080 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.838091 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:24Z","lastTransitionTime":"2025-11-26T00:24:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.941108 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.941153 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.941162 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.941176 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:24 crc kubenswrapper[4766]: I1126 00:24:24.941188 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:24Z","lastTransitionTime":"2025-11-26T00:24:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.044271 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.044304 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.044312 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.044326 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.044335 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:25Z","lastTransitionTime":"2025-11-26T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.147041 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.147081 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.147091 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.147106 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.147118 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:25Z","lastTransitionTime":"2025-11-26T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.250680 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.250750 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.250774 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.250805 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.250828 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:25Z","lastTransitionTime":"2025-11-26T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.353103 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.353171 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.353186 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.353203 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.353216 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:25Z","lastTransitionTime":"2025-11-26T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.456141 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.456229 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.456258 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.456304 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.456360 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:25Z","lastTransitionTime":"2025-11-26T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.558404 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.558445 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.558482 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.558510 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.558521 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:25Z","lastTransitionTime":"2025-11-26T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.654682 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.654717 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.654725 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.654738 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.654748 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:25Z","lastTransitionTime":"2025-11-26T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:25 crc kubenswrapper[4766]: E1126 00:24:25.669744 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:25Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.673404 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.673438 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.673447 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.673460 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.673470 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:25Z","lastTransitionTime":"2025-11-26T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:25 crc kubenswrapper[4766]: E1126 00:24:25.685000 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:25Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.688393 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.688430 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.688440 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.688456 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.688469 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:25Z","lastTransitionTime":"2025-11-26T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:25 crc kubenswrapper[4766]: E1126 00:24:25.701429 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:25Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.705630 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.705704 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.705720 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.705740 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.705755 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:25Z","lastTransitionTime":"2025-11-26T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:25 crc kubenswrapper[4766]: E1126 00:24:25.721319 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:25Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.725740 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.725784 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.725798 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.725814 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.725827 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:25Z","lastTransitionTime":"2025-11-26T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:25 crc kubenswrapper[4766]: E1126 00:24:25.745823 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:25Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:25 crc kubenswrapper[4766]: E1126 00:24:25.745943 4766 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.747407 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.747456 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.747467 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.747482 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.747493 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:25Z","lastTransitionTime":"2025-11-26T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.826208 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:25 crc kubenswrapper[4766]: E1126 00:24:25.826481 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.849612 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.849705 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.849718 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.849738 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.849752 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:25Z","lastTransitionTime":"2025-11-26T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.952360 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.952420 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.952429 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.952441 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:25 crc kubenswrapper[4766]: I1126 00:24:25.952450 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:25Z","lastTransitionTime":"2025-11-26T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.054836 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.054921 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.054955 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.054985 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.055005 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:26Z","lastTransitionTime":"2025-11-26T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.103740 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs\") pod \"network-metrics-daemon-5z9ll\" (UID: \"b6600efa-4858-41aa-a890-44ce262b63dd\") " pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:26 crc kubenswrapper[4766]: E1126 00:24:26.103936 4766 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 00:24:26 crc kubenswrapper[4766]: E1126 00:24:26.104019 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs podName:b6600efa-4858-41aa-a890-44ce262b63dd nodeName:}" failed. No retries permitted until 2025-11-26 00:24:34.103995711 +0000 UTC m=+54.952766161 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs") pod "network-metrics-daemon-5z9ll" (UID: "b6600efa-4858-41aa-a890-44ce262b63dd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.157041 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.157091 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.157104 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.157122 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.157140 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:26Z","lastTransitionTime":"2025-11-26T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.259796 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.259840 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.259852 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.259871 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.259883 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:26Z","lastTransitionTime":"2025-11-26T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.362311 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.362378 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.362396 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.362422 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.362444 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:26Z","lastTransitionTime":"2025-11-26T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.464916 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.464962 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.464976 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.464992 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.465003 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:26Z","lastTransitionTime":"2025-11-26T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.567482 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.567521 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.567529 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.567542 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.567551 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:26Z","lastTransitionTime":"2025-11-26T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.670763 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.670806 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.670818 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.670834 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.670846 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:26Z","lastTransitionTime":"2025-11-26T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.773909 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.773964 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.773974 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.773988 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.773999 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:26Z","lastTransitionTime":"2025-11-26T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.826032 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.826052 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:26 crc kubenswrapper[4766]: E1126 00:24:26.826283 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.826073 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:26 crc kubenswrapper[4766]: E1126 00:24:26.826412 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:26 crc kubenswrapper[4766]: E1126 00:24:26.826439 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.877231 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.877627 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.877714 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.877796 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.877877 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:26Z","lastTransitionTime":"2025-11-26T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.980470 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.980546 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.980567 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.980594 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:26 crc kubenswrapper[4766]: I1126 00:24:26.980616 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:26Z","lastTransitionTime":"2025-11-26T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.084243 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.084303 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.084314 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.084333 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.084343 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:27Z","lastTransitionTime":"2025-11-26T00:24:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.186223 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.186258 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.186270 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.186286 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.186296 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:27Z","lastTransitionTime":"2025-11-26T00:24:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.288948 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.288985 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.288993 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.289007 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.289016 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:27Z","lastTransitionTime":"2025-11-26T00:24:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.391520 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.391614 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.391632 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.391688 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.391708 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:27Z","lastTransitionTime":"2025-11-26T00:24:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.493803 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.493859 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.493875 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.493896 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.493912 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:27Z","lastTransitionTime":"2025-11-26T00:24:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.596890 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.596948 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.596965 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.596988 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.597013 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:27Z","lastTransitionTime":"2025-11-26T00:24:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.699981 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.700054 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.700077 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.700106 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.700129 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:27Z","lastTransitionTime":"2025-11-26T00:24:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.802952 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.802982 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.802991 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.803006 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.803015 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:27Z","lastTransitionTime":"2025-11-26T00:24:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.825981 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:27 crc kubenswrapper[4766]: E1126 00:24:27.826173 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.905322 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.905382 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.905393 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.905410 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:27 crc kubenswrapper[4766]: I1126 00:24:27.905422 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:27Z","lastTransitionTime":"2025-11-26T00:24:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.008945 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.008985 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.008996 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.009011 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.009022 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:28Z","lastTransitionTime":"2025-11-26T00:24:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.111224 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.111277 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.111292 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.111312 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.111327 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:28Z","lastTransitionTime":"2025-11-26T00:24:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.214179 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.214247 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.214271 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.214301 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.214349 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:28Z","lastTransitionTime":"2025-11-26T00:24:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.316637 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.316703 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.316714 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.316730 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.316741 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:28Z","lastTransitionTime":"2025-11-26T00:24:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.420128 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.420230 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.420251 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.420277 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.420295 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:28Z","lastTransitionTime":"2025-11-26T00:24:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.522178 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.522209 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.522218 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.522231 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.522239 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:28Z","lastTransitionTime":"2025-11-26T00:24:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.625151 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.625218 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.625236 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.625259 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.625276 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:28Z","lastTransitionTime":"2025-11-26T00:24:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.728725 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.728774 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.728785 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.728803 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.728817 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:28Z","lastTransitionTime":"2025-11-26T00:24:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.826020 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.826032 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:28 crc kubenswrapper[4766]: E1126 00:24:28.826229 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.826063 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:28 crc kubenswrapper[4766]: E1126 00:24:28.826376 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:28 crc kubenswrapper[4766]: E1126 00:24:28.826532 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.831842 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.831906 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.831930 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.831959 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.831980 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:28Z","lastTransitionTime":"2025-11-26T00:24:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.935172 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.935239 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.935257 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.935285 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:28 crc kubenswrapper[4766]: I1126 00:24:28.935303 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:28Z","lastTransitionTime":"2025-11-26T00:24:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.038119 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.038185 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.038202 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.038226 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.038244 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:29Z","lastTransitionTime":"2025-11-26T00:24:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.140974 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.141040 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.141059 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.141081 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.141098 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:29Z","lastTransitionTime":"2025-11-26T00:24:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.244349 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.244413 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.244437 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.244466 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.244489 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:29Z","lastTransitionTime":"2025-11-26T00:24:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.347809 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.347884 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.347910 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.347941 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.347963 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:29Z","lastTransitionTime":"2025-11-26T00:24:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.450589 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.450698 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.450735 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.450764 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.450806 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:29Z","lastTransitionTime":"2025-11-26T00:24:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.553762 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.553836 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.553861 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.553890 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.553914 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:29Z","lastTransitionTime":"2025-11-26T00:24:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.656785 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.656846 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.656869 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.656898 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.656964 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:29Z","lastTransitionTime":"2025-11-26T00:24:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.759017 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.759090 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.759107 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.759130 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.759144 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:29Z","lastTransitionTime":"2025-11-26T00:24:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.825889 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:29 crc kubenswrapper[4766]: E1126 00:24:29.826383 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.826557 4766 scope.go:117] "RemoveContainer" containerID="7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.844596 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:29Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.862901 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.862948 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.862959 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.862978 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.862991 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:29Z","lastTransitionTime":"2025-11-26T00:24:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.863110 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:29Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.873829 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.888049 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:29Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.920339 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f07ffa8c50c44fa2abfa4579ca7b53f16f8630f358f0375fc57920edfe771505\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"message\\\":\\\" handler 8\\\\nI1126 00:24:13.651998 6045 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 00:24:13.652004 6045 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 00:24:13.652053 6045 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652140 6045 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652270 6045 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:13.652350 6045 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1126 00:24:13.652644 6045 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 00:24:13.652675 6045 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 00:24:13.652694 6045 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 00:24:13.652709 6045 factory.go:656] Stopping watch factory\\\\nI1126 00:24:13.652717 6045 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 00:24:13.652732 6045 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 00:24:13.652725 6045 ovnkube.go:599] Stopped ovnkube\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"423049 6211 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 00:24:15.423069 6211 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1126 00:24:15.423443 6211 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1126 00:24:15.423455 6211 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1126 00:24:15.423492 6211 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1126 00:24:15.423506 6211 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 00:24:15.423508 6211 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 00:24:15.423518 6211 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1126 00:24:15.423520 6211 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 00:24:15.423530 6211 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 00:24:15.423557 6211 factory.go:656] Stopping watch factory\\\\nI1126 00:24:15.423569 6211 ovnkube.go:599] Stopped ovnkube\\\\nI1126 00:24:15.423580 6211 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 00:24:15.423604 6211 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1126 00:24:15.423687 6211 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:29Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.942557 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:29Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.957062 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:29Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.966609 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.966700 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.966713 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.966731 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.966746 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:29Z","lastTransitionTime":"2025-11-26T00:24:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.972095 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:29Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.983400 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:29Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:29 crc kubenswrapper[4766]: I1126 00:24:29.992121 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:29Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.003079 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.012622 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5z9ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6600efa-4858-41aa-a890-44ce262b63dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5z9ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.025890 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.037151 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.048483 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.060134 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"443534cf-37a7-4831-a581-9d04dee05ead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0874e594efd8c1df4af245ea08256417c1b6cfe83fd8ce08848ea6710081cf1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb484d4736b942d10794d5e39064bec67b5b30e9d1ac668f5a62dbc15119c3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-49ztc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.069978 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.070002 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.070011 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.070026 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.070036 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:30Z","lastTransitionTime":"2025-11-26T00:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.072861 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.095893 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.109345 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.119191 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.128111 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.140031 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.158277 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"423049 6211 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 00:24:15.423069 6211 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1126 00:24:15.423443 6211 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1126 00:24:15.423455 6211 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1126 00:24:15.423492 6211 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1126 00:24:15.423506 6211 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 00:24:15.423508 6211 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 00:24:15.423518 6211 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1126 00:24:15.423520 6211 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 00:24:15.423530 6211 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 00:24:15.423557 6211 factory.go:656] Stopping watch factory\\\\nI1126 00:24:15.423569 6211 ovnkube.go:599] Stopped ovnkube\\\\nI1126 00:24:15.423580 6211 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 00:24:15.423604 6211 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1126 00:24:15.423687 6211 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fxnln_openshift-ovn-kubernetes(415fdacb-85c1-4265-89b7-6771a84ffc89)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.171592 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.172297 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.172335 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.172345 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.172360 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.172370 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:30Z","lastTransitionTime":"2025-11-26T00:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.183857 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.195724 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.212271 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.221777 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.236895 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.250387 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5z9ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6600efa-4858-41aa-a890-44ce262b63dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5z9ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.263493 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.275185 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.275225 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.275238 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.275258 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.275270 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:30Z","lastTransitionTime":"2025-11-26T00:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.289680 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.301057 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.311556 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"443534cf-37a7-4831-a581-9d04dee05ead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0874e594efd8c1df4af245ea08256417c1b6cfe83fd8ce08848ea6710081cf1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb484d4736b942d10794d5e39064bec67b5b30e9d1ac668f5a62dbc15119c3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-49ztc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.324798 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.377638 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.377699 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.377710 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.377726 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.377737 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:30Z","lastTransitionTime":"2025-11-26T00:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.453779 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fxnln_415fdacb-85c1-4265-89b7-6771a84ffc89/ovnkube-controller/1.log" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.455981 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerStarted","Data":"c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a"} Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.456389 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.467187 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5z9ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6600efa-4858-41aa-a890-44ce262b63dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5z9ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.478508 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.479257 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.479288 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.479300 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.479315 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.479325 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:30Z","lastTransitionTime":"2025-11-26T00:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.488940 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.499299 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.508620 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.519400 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.541287 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.608501 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:24:30 crc kubenswrapper[4766]: E1126 00:24:30.608742 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:25:02.608683536 +0000 UTC m=+83.457454016 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.608956 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.608982 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.609014 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.609050 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:30 crc kubenswrapper[4766]: E1126 00:24:30.610183 4766 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 00:24:30 crc kubenswrapper[4766]: E1126 00:24:30.610253 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 00:25:02.610234442 +0000 UTC m=+83.459004872 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 00:24:30 crc kubenswrapper[4766]: E1126 00:24:30.610386 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 00:24:30 crc kubenswrapper[4766]: E1126 00:24:30.610595 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 00:24:30 crc kubenswrapper[4766]: E1126 00:24:30.610614 4766 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:24:30 crc kubenswrapper[4766]: E1126 00:24:30.610814 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 00:25:02.610665493 +0000 UTC m=+83.459435923 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:24:30 crc kubenswrapper[4766]: E1126 00:24:30.611062 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 00:24:30 crc kubenswrapper[4766]: E1126 00:24:30.611086 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 00:24:30 crc kubenswrapper[4766]: E1126 00:24:30.611107 4766 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:24:30 crc kubenswrapper[4766]: E1126 00:24:30.611298 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 00:25:02.611281127 +0000 UTC m=+83.460051557 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:24:30 crc kubenswrapper[4766]: E1126 00:24:30.611516 4766 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 00:24:30 crc kubenswrapper[4766]: E1126 00:24:30.611667 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 00:25:02.611556584 +0000 UTC m=+83.460327004 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.613354 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.613478 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.613555 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.613635 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.613747 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:30Z","lastTransitionTime":"2025-11-26T00:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.623889 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.633218 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.643068 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"443534cf-37a7-4831-a581-9d04dee05ead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0874e594efd8c1df4af245ea08256417c1b6cfe83fd8ce08848ea6710081cf1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb484d4736b942d10794d5e39064bec67b5b30e9d1ac668f5a62dbc15119c3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-49ztc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.670334 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"423049 6211 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 00:24:15.423069 6211 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1126 00:24:15.423443 6211 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1126 00:24:15.423455 6211 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1126 00:24:15.423492 6211 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1126 00:24:15.423506 6211 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 00:24:15.423508 6211 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 00:24:15.423518 6211 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1126 00:24:15.423520 6211 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 00:24:15.423530 6211 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 00:24:15.423557 6211 factory.go:656] Stopping watch factory\\\\nI1126 00:24:15.423569 6211 ovnkube.go:599] Stopped ovnkube\\\\nI1126 00:24:15.423580 6211 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 00:24:15.423604 6211 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1126 00:24:15.423687 6211 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.683376 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.694479 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.705838 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.716134 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.716177 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.716186 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.716204 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.716215 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:30Z","lastTransitionTime":"2025-11-26T00:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.721126 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.732608 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.742202 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:30Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.818634 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.818679 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.818689 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.818703 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.818713 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:30Z","lastTransitionTime":"2025-11-26T00:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.825863 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.825889 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.825879 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:30 crc kubenswrapper[4766]: E1126 00:24:30.825985 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:30 crc kubenswrapper[4766]: E1126 00:24:30.826068 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:30 crc kubenswrapper[4766]: E1126 00:24:30.826173 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.920635 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.920690 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.920702 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.920720 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:30 crc kubenswrapper[4766]: I1126 00:24:30.920733 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:30Z","lastTransitionTime":"2025-11-26T00:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.023568 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.023616 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.023632 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.023678 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.023696 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:31Z","lastTransitionTime":"2025-11-26T00:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.125824 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.125865 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.125877 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.125894 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.125906 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:31Z","lastTransitionTime":"2025-11-26T00:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.228480 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.228525 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.228535 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.228549 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.228560 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:31Z","lastTransitionTime":"2025-11-26T00:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.331229 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.331279 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.331298 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.331322 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.331341 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:31Z","lastTransitionTime":"2025-11-26T00:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.433837 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.433874 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.433885 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.433902 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.433913 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:31Z","lastTransitionTime":"2025-11-26T00:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.460791 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fxnln_415fdacb-85c1-4265-89b7-6771a84ffc89/ovnkube-controller/2.log" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.461289 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fxnln_415fdacb-85c1-4265-89b7-6771a84ffc89/ovnkube-controller/1.log" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.464247 4766 generic.go:334] "Generic (PLEG): container finished" podID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerID="c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a" exitCode=1 Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.464306 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerDied","Data":"c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a"} Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.464388 4766 scope.go:117] "RemoveContainer" containerID="7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.464991 4766 scope.go:117] "RemoveContainer" containerID="c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a" Nov 26 00:24:31 crc kubenswrapper[4766]: E1126 00:24:31.465160 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fxnln_openshift-ovn-kubernetes(415fdacb-85c1-4265-89b7-6771a84ffc89)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.480537 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:31Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.492154 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:31Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.504450 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:31Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.514037 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:31Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.524245 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:31Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.534537 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5z9ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6600efa-4858-41aa-a890-44ce262b63dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5z9ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:31Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.535910 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.536079 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.536096 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.536110 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.536121 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:31Z","lastTransitionTime":"2025-11-26T00:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.547245 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:31Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.569145 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:31Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.579529 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:31Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.592021 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"443534cf-37a7-4831-a581-9d04dee05ead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0874e594efd8c1df4af245ea08256417c1b6cfe83fd8ce08848ea6710081cf1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb484d4736b942d10794d5e39064bec67b5b30e9d1ac668f5a62dbc15119c3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-49ztc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:31Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.606313 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:31Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.618557 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:31Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.632206 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:31Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.638444 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.638468 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.638476 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.638489 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.638498 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:31Z","lastTransitionTime":"2025-11-26T00:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.644295 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:31Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.659683 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e25fe0f04ec382e086f1d89659a8f109aa4d375c0b8aab09db19793fb79a945\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:15Z\\\",\\\"message\\\":\\\"423049 6211 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 00:24:15.423069 6211 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1126 00:24:15.423443 6211 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1126 00:24:15.423455 6211 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1126 00:24:15.423492 6211 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1126 00:24:15.423506 6211 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 00:24:15.423508 6211 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 00:24:15.423518 6211 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1126 00:24:15.423520 6211 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 00:24:15.423530 6211 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 00:24:15.423557 6211 factory.go:656] Stopping watch factory\\\\nI1126 00:24:15.423569 6211 ovnkube.go:599] Stopped ovnkube\\\\nI1126 00:24:15.423580 6211 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 00:24:15.423604 6211 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1126 00:24:15.423687 6211 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:30Z\\\",\\\"message\\\":\\\"iserver-crc after 0 failed attempt(s)\\\\nI1126 00:24:30.895910 6449 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1126 00:24:30.895884 6449 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1126 00:24:30.895916 6449 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI1126 00:24:30.895920 6449 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1126 00:24:30.895911 6449 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1126 00:24:30.895679 6449 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1126 00:24:30.895942 6449 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF1126 00:24:30.895947 6449 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:31Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.671424 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:31Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.681086 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:31Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.741203 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.741242 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.741254 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.741270 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.741282 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:31Z","lastTransitionTime":"2025-11-26T00:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.826820 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:31 crc kubenswrapper[4766]: E1126 00:24:31.827008 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.844264 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.844313 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.844327 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.844346 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.844360 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:31Z","lastTransitionTime":"2025-11-26T00:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.946998 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.947037 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.947048 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.947063 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:31 crc kubenswrapper[4766]: I1126 00:24:31.947077 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:31Z","lastTransitionTime":"2025-11-26T00:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.049263 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.049296 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.049312 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.049333 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.049345 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:32Z","lastTransitionTime":"2025-11-26T00:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.152033 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.152081 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.152090 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.152105 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.152115 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:32Z","lastTransitionTime":"2025-11-26T00:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.254305 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.254347 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.254356 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.254369 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.254377 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:32Z","lastTransitionTime":"2025-11-26T00:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.357180 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.357237 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.357249 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.357268 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.357281 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:32Z","lastTransitionTime":"2025-11-26T00:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.459639 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.459686 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.459697 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.459711 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.459724 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:32Z","lastTransitionTime":"2025-11-26T00:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.468195 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fxnln_415fdacb-85c1-4265-89b7-6771a84ffc89/ovnkube-controller/2.log" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.472095 4766 scope.go:117] "RemoveContainer" containerID="c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a" Nov 26 00:24:32 crc kubenswrapper[4766]: E1126 00:24:32.472233 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fxnln_openshift-ovn-kubernetes(415fdacb-85c1-4265-89b7-6771a84ffc89)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.484378 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:32Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.497006 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:32Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.505458 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:32Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.516704 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:32Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.524788 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5z9ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6600efa-4858-41aa-a890-44ce262b63dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5z9ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:32Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.534252 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:32Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.549735 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:32Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.558620 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:32Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.561671 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.561702 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.561713 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.561726 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.561737 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:32Z","lastTransitionTime":"2025-11-26T00:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.568900 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"443534cf-37a7-4831-a581-9d04dee05ead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0874e594efd8c1df4af245ea08256417c1b6cfe83fd8ce08848ea6710081cf1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb484d4736b942d10794d5e39064bec67b5b30e9d1ac668f5a62dbc15119c3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-49ztc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:32Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.583045 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:32Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.594706 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:32Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.605560 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:32Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.616575 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:32Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.628328 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:32Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.648859 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:30Z\\\",\\\"message\\\":\\\"iserver-crc after 0 failed attempt(s)\\\\nI1126 00:24:30.895910 6449 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1126 00:24:30.895884 6449 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1126 00:24:30.895916 6449 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI1126 00:24:30.895920 6449 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1126 00:24:30.895911 6449 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1126 00:24:30.895679 6449 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1126 00:24:30.895942 6449 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF1126 00:24:30.895947 6449 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fxnln_openshift-ovn-kubernetes(415fdacb-85c1-4265-89b7-6771a84ffc89)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:32Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.659824 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:32Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.664057 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.664108 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.664126 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.664149 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.664166 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:32Z","lastTransitionTime":"2025-11-26T00:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.671016 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:32Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.767444 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.767499 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.767516 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.767538 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.767555 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:32Z","lastTransitionTime":"2025-11-26T00:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.826579 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.826610 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.826579 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:32 crc kubenswrapper[4766]: E1126 00:24:32.826824 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:32 crc kubenswrapper[4766]: E1126 00:24:32.826927 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:32 crc kubenswrapper[4766]: E1126 00:24:32.827018 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.870872 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.870945 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.870969 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.870998 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.871020 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:32Z","lastTransitionTime":"2025-11-26T00:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.973968 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.974025 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.974042 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.974064 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:32 crc kubenswrapper[4766]: I1126 00:24:32.974081 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:32Z","lastTransitionTime":"2025-11-26T00:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.032793 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.045729 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.050121 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:33Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.077352 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:33Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.077912 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.077947 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.077961 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.077982 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.077997 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:33Z","lastTransitionTime":"2025-11-26T00:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.088787 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:33Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.099155 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"443534cf-37a7-4831-a581-9d04dee05ead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0874e594efd8c1df4af245ea08256417c1b6cfe83fd8ce08848ea6710081cf1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb484d4736b942d10794d5e39064bec67b5b30e9d1ac668f5a62dbc15119c3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-49ztc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:33Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.111714 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:33Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.123839 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:33Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.136514 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:33Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.150939 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:33Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.173080 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:30Z\\\",\\\"message\\\":\\\"iserver-crc after 0 failed attempt(s)\\\\nI1126 00:24:30.895910 6449 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1126 00:24:30.895884 6449 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1126 00:24:30.895916 6449 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI1126 00:24:30.895920 6449 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1126 00:24:30.895911 6449 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1126 00:24:30.895679 6449 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1126 00:24:30.895942 6449 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF1126 00:24:30.895947 6449 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fxnln_openshift-ovn-kubernetes(415fdacb-85c1-4265-89b7-6771a84ffc89)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:33Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.180616 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.180686 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.180697 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.180712 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.180725 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:33Z","lastTransitionTime":"2025-11-26T00:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.191530 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:33Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.201264 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:33Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.212518 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:33Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.224309 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:33Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.236026 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:33Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.246873 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:33Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.258053 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:33Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.267810 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5z9ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6600efa-4858-41aa-a890-44ce262b63dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5z9ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:33Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.283800 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.283831 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.283839 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.283853 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.283863 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:33Z","lastTransitionTime":"2025-11-26T00:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.386474 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.386524 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.386547 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.386567 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.386581 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:33Z","lastTransitionTime":"2025-11-26T00:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.490079 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.490154 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.490179 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.490208 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.490227 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:33Z","lastTransitionTime":"2025-11-26T00:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.593395 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.593454 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.593477 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.593497 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.593519 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:33Z","lastTransitionTime":"2025-11-26T00:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.696930 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.697018 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.697054 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.697085 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.697110 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:33Z","lastTransitionTime":"2025-11-26T00:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.800571 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.800642 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.800699 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.800729 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.800751 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:33Z","lastTransitionTime":"2025-11-26T00:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.825928 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:33 crc kubenswrapper[4766]: E1126 00:24:33.826070 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.905236 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.905275 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.905285 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.905300 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:33 crc kubenswrapper[4766]: I1126 00:24:33.905311 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:33Z","lastTransitionTime":"2025-11-26T00:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.007588 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.007635 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.007647 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.007674 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.007683 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:34Z","lastTransitionTime":"2025-11-26T00:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.111082 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.111150 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.111168 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.111205 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.111222 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:34Z","lastTransitionTime":"2025-11-26T00:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.144639 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs\") pod \"network-metrics-daemon-5z9ll\" (UID: \"b6600efa-4858-41aa-a890-44ce262b63dd\") " pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:34 crc kubenswrapper[4766]: E1126 00:24:34.144841 4766 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 00:24:34 crc kubenswrapper[4766]: E1126 00:24:34.144911 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs podName:b6600efa-4858-41aa-a890-44ce262b63dd nodeName:}" failed. No retries permitted until 2025-11-26 00:24:50.14489249 +0000 UTC m=+70.993662920 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs") pod "network-metrics-daemon-5z9ll" (UID: "b6600efa-4858-41aa-a890-44ce262b63dd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.214812 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.214843 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.214861 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.214879 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.214890 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:34Z","lastTransitionTime":"2025-11-26T00:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.317804 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.317892 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.317904 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.317924 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.317936 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:34Z","lastTransitionTime":"2025-11-26T00:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.420425 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.420483 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.420494 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.420511 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.420523 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:34Z","lastTransitionTime":"2025-11-26T00:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.523731 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.523774 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.523789 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.523809 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.523825 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:34Z","lastTransitionTime":"2025-11-26T00:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.627271 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.627349 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.627373 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.627403 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.627427 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:34Z","lastTransitionTime":"2025-11-26T00:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.730754 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.730807 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.730818 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.730835 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.730846 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:34Z","lastTransitionTime":"2025-11-26T00:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.826163 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.826258 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.826167 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:34 crc kubenswrapper[4766]: E1126 00:24:34.826364 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:34 crc kubenswrapper[4766]: E1126 00:24:34.826470 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:34 crc kubenswrapper[4766]: E1126 00:24:34.826562 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.833468 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.833510 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.833522 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.833540 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.833553 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:34Z","lastTransitionTime":"2025-11-26T00:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.936146 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.936202 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.936221 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.936247 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:34 crc kubenswrapper[4766]: I1126 00:24:34.936266 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:34Z","lastTransitionTime":"2025-11-26T00:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.040171 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.040232 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.040250 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.040273 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.040291 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:35Z","lastTransitionTime":"2025-11-26T00:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.143113 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.143184 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.143207 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.143238 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.143259 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:35Z","lastTransitionTime":"2025-11-26T00:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.247395 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.247468 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.247483 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.247502 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.247517 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:35Z","lastTransitionTime":"2025-11-26T00:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.350064 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.350098 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.350108 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.350121 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.350134 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:35Z","lastTransitionTime":"2025-11-26T00:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.452507 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.452546 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.452557 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.452574 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.452586 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:35Z","lastTransitionTime":"2025-11-26T00:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.554936 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.554972 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.554981 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.554995 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.555003 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:35Z","lastTransitionTime":"2025-11-26T00:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.657799 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.657840 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.657851 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.657866 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.657878 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:35Z","lastTransitionTime":"2025-11-26T00:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.760398 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.760441 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.760453 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.760470 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.760481 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:35Z","lastTransitionTime":"2025-11-26T00:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.826600 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:35 crc kubenswrapper[4766]: E1126 00:24:35.826823 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.862532 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.862572 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.862583 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.862599 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.862609 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:35Z","lastTransitionTime":"2025-11-26T00:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.965317 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.965355 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.965363 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.965381 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:35 crc kubenswrapper[4766]: I1126 00:24:35.965390 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:35Z","lastTransitionTime":"2025-11-26T00:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.067586 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.067702 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.067722 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.067770 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.067789 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:36Z","lastTransitionTime":"2025-11-26T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.075955 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.076012 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.076030 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.076051 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.076067 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:36Z","lastTransitionTime":"2025-11-26T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:36 crc kubenswrapper[4766]: E1126 00:24:36.100730 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:36Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.104390 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.104434 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.104444 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.104486 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.104498 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:36Z","lastTransitionTime":"2025-11-26T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:36 crc kubenswrapper[4766]: E1126 00:24:36.118787 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:36Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.122392 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.122433 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.122448 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.122468 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.122483 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:36Z","lastTransitionTime":"2025-11-26T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:36 crc kubenswrapper[4766]: E1126 00:24:36.137828 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:36Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.142644 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.142707 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.142718 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.142737 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.143123 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:36Z","lastTransitionTime":"2025-11-26T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:36 crc kubenswrapper[4766]: E1126 00:24:36.155255 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:36Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.159447 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.159486 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.159500 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.159518 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.159532 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:36Z","lastTransitionTime":"2025-11-26T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:36 crc kubenswrapper[4766]: E1126 00:24:36.176358 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:36Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:36 crc kubenswrapper[4766]: E1126 00:24:36.176463 4766 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.178214 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.178268 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.178282 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.178302 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.178316 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:36Z","lastTransitionTime":"2025-11-26T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.281467 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.281533 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.281545 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.281562 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.281576 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:36Z","lastTransitionTime":"2025-11-26T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.384620 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.384723 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.384749 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.384785 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.384810 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:36Z","lastTransitionTime":"2025-11-26T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.487159 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.487216 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.487233 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.487256 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.487277 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:36Z","lastTransitionTime":"2025-11-26T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.589925 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.589986 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.590003 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.590024 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.590041 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:36Z","lastTransitionTime":"2025-11-26T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.692780 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.692842 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.692854 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.692869 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.692880 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:36Z","lastTransitionTime":"2025-11-26T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.795428 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.795509 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.795531 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.795562 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.795585 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:36Z","lastTransitionTime":"2025-11-26T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.826299 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:36 crc kubenswrapper[4766]: E1126 00:24:36.826521 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.826939 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.826943 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:36 crc kubenswrapper[4766]: E1126 00:24:36.827126 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:36 crc kubenswrapper[4766]: E1126 00:24:36.827268 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.898987 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.899047 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.899067 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.899093 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:36 crc kubenswrapper[4766]: I1126 00:24:36.899111 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:36Z","lastTransitionTime":"2025-11-26T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.001849 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.001881 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.001893 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.001909 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.001921 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:37Z","lastTransitionTime":"2025-11-26T00:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.106450 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.106504 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.106516 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.106533 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.106546 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:37Z","lastTransitionTime":"2025-11-26T00:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.209096 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.209171 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.209197 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.209228 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.209250 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:37Z","lastTransitionTime":"2025-11-26T00:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.311717 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.311768 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.311784 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.311805 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.311820 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:37Z","lastTransitionTime":"2025-11-26T00:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.414368 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.414409 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.414418 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.414434 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.414444 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:37Z","lastTransitionTime":"2025-11-26T00:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.516997 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.517029 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.517039 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.517055 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.517067 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:37Z","lastTransitionTime":"2025-11-26T00:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.619107 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.619160 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.619177 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.619198 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.619215 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:37Z","lastTransitionTime":"2025-11-26T00:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.722117 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.722222 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.722229 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.722243 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.722252 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:37Z","lastTransitionTime":"2025-11-26T00:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.825566 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.825640 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.825696 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.825727 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.825746 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:37Z","lastTransitionTime":"2025-11-26T00:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.825773 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:37 crc kubenswrapper[4766]: E1126 00:24:37.825991 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.928350 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.928406 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.928417 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.928435 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:37 crc kubenswrapper[4766]: I1126 00:24:37.928447 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:37Z","lastTransitionTime":"2025-11-26T00:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.031790 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.031836 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.031849 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.031868 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.031881 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:38Z","lastTransitionTime":"2025-11-26T00:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.135044 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.135357 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.135476 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.135580 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.135729 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:38Z","lastTransitionTime":"2025-11-26T00:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.237683 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.237745 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.237763 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.237786 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.237805 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:38Z","lastTransitionTime":"2025-11-26T00:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.340990 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.341057 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.341076 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.341104 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.341122 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:38Z","lastTransitionTime":"2025-11-26T00:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.444369 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.445066 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.445126 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.445146 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.445158 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:38Z","lastTransitionTime":"2025-11-26T00:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.547855 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.547895 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.547905 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.547923 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.547934 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:38Z","lastTransitionTime":"2025-11-26T00:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.650129 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.650197 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.650220 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.650248 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.650270 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:38Z","lastTransitionTime":"2025-11-26T00:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.753327 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.753353 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.753361 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.753373 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.753382 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:38Z","lastTransitionTime":"2025-11-26T00:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.826316 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.826379 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:38 crc kubenswrapper[4766]: E1126 00:24:38.826469 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.826492 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:38 crc kubenswrapper[4766]: E1126 00:24:38.826680 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:38 crc kubenswrapper[4766]: E1126 00:24:38.826807 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.856673 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.856915 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.857183 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.857265 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.857342 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:38Z","lastTransitionTime":"2025-11-26T00:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.960276 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.960325 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.960340 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.960359 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:38 crc kubenswrapper[4766]: I1126 00:24:38.960373 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:38Z","lastTransitionTime":"2025-11-26T00:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.062810 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.062843 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.062852 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.062864 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.062873 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:39Z","lastTransitionTime":"2025-11-26T00:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.165416 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.165446 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.165454 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.165466 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.165476 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:39Z","lastTransitionTime":"2025-11-26T00:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.267554 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.267600 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.267609 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.267624 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.267675 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:39Z","lastTransitionTime":"2025-11-26T00:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.370560 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.370613 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.370624 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.370641 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.370672 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:39Z","lastTransitionTime":"2025-11-26T00:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.474124 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.474169 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.474181 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.474200 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.474212 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:39Z","lastTransitionTime":"2025-11-26T00:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.576605 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.576681 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.576700 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.576721 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.576736 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:39Z","lastTransitionTime":"2025-11-26T00:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.680804 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.680846 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.680854 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.680869 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.680877 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:39Z","lastTransitionTime":"2025-11-26T00:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.782445 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.782481 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.782492 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.782507 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.782516 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:39Z","lastTransitionTime":"2025-11-26T00:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.825906 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:39 crc kubenswrapper[4766]: E1126 00:24:39.826033 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.841302 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:39Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.864731 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:39Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.876699 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:39Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.884727 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.884756 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.884764 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.884777 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.884786 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:39Z","lastTransitionTime":"2025-11-26T00:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.890746 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:39Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.902996 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5z9ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6600efa-4858-41aa-a890-44ce262b63dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5z9ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:39Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.915897 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:39Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.935775 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:39Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.948970 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:39Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.961296 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"443534cf-37a7-4831-a581-9d04dee05ead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0874e594efd8c1df4af245ea08256417c1b6cfe83fd8ce08848ea6710081cf1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb484d4736b942d10794d5e39064bec67b5b30e9d1ac668f5a62dbc15119c3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-49ztc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:39Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.973891 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:39Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.986417 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:39Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.987741 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.987828 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.987856 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.987888 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:39 crc kubenswrapper[4766]: I1126 00:24:39.987923 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:39Z","lastTransitionTime":"2025-11-26T00:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.000074 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:39Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.014092 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:40Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.033699 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:40Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.054923 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:30Z\\\",\\\"message\\\":\\\"iserver-crc after 0 failed attempt(s)\\\\nI1126 00:24:30.895910 6449 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1126 00:24:30.895884 6449 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1126 00:24:30.895916 6449 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI1126 00:24:30.895920 6449 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1126 00:24:30.895911 6449 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1126 00:24:30.895679 6449 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1126 00:24:30.895942 6449 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF1126 00:24:30.895947 6449 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fxnln_openshift-ovn-kubernetes(415fdacb-85c1-4265-89b7-6771a84ffc89)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:40Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.067125 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:40Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.077737 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:40Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.087589 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20c9b10f-a395-40ae-a0aa-b3f0b5f63682\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01950717656ec2766fa80a35e2f7a9136409c2331979c67bffa25b965864be0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2293d0b6924d276777e52994713be62a0f98ff846ecee3ed7f597835e18a8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2dd6961899214f94684d3cb24a3c75f9870bf0fc667ef4e2b5c25fdc4b0e5da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2b174d3a0b70785380dfbac2eb1062d84fafedff25f8322cc25aea9c362b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e2b174d3a0b70785380dfbac2eb1062d84fafedff25f8322cc25aea9c362b6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:40Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.090152 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.090195 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.090205 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.090217 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.090225 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:40Z","lastTransitionTime":"2025-11-26T00:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.192703 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.193004 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.193136 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.193323 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.193443 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:40Z","lastTransitionTime":"2025-11-26T00:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.296634 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.296703 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.296713 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.296726 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.296737 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:40Z","lastTransitionTime":"2025-11-26T00:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.399747 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.399789 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.399800 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.399823 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.399836 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:40Z","lastTransitionTime":"2025-11-26T00:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.502128 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.502744 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.502781 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.502797 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.502808 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:40Z","lastTransitionTime":"2025-11-26T00:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.606075 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.606151 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.606174 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.606203 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.606226 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:40Z","lastTransitionTime":"2025-11-26T00:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.709038 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.709312 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.709328 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.709347 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.709361 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:40Z","lastTransitionTime":"2025-11-26T00:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.812279 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.812346 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.812369 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.812397 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.812419 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:40Z","lastTransitionTime":"2025-11-26T00:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.825723 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.825743 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:40 crc kubenswrapper[4766]: E1126 00:24:40.825820 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.825728 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:40 crc kubenswrapper[4766]: E1126 00:24:40.826049 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:40 crc kubenswrapper[4766]: E1126 00:24:40.826125 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.914871 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.914921 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.914936 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.914965 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:40 crc kubenswrapper[4766]: I1126 00:24:40.914976 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:40Z","lastTransitionTime":"2025-11-26T00:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.017123 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.017164 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.017172 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.017186 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.017195 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:41Z","lastTransitionTime":"2025-11-26T00:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.120141 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.120178 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.120188 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.120202 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.120213 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:41Z","lastTransitionTime":"2025-11-26T00:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.223002 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.223265 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.223353 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.223471 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.223561 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:41Z","lastTransitionTime":"2025-11-26T00:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.326812 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.327264 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.327406 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.327490 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.327558 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:41Z","lastTransitionTime":"2025-11-26T00:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.430026 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.430060 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.430072 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.430089 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.430103 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:41Z","lastTransitionTime":"2025-11-26T00:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.533059 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.533104 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.533119 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.533139 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.533154 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:41Z","lastTransitionTime":"2025-11-26T00:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.635863 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.635897 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.635908 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.635922 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.635932 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:41Z","lastTransitionTime":"2025-11-26T00:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.738606 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.738719 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.738746 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.738776 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.738798 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:41Z","lastTransitionTime":"2025-11-26T00:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.826923 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:41 crc kubenswrapper[4766]: E1126 00:24:41.827559 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.845519 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.845568 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.845585 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.845609 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.845626 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:41Z","lastTransitionTime":"2025-11-26T00:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.947958 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.948206 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.948280 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.948348 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:41 crc kubenswrapper[4766]: I1126 00:24:41.948420 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:41Z","lastTransitionTime":"2025-11-26T00:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.051133 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.051420 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.051528 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.051637 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.051766 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:42Z","lastTransitionTime":"2025-11-26T00:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.154804 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.154854 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.154866 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.154885 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.154899 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:42Z","lastTransitionTime":"2025-11-26T00:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.257709 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.257759 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.257768 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.257782 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.257791 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:42Z","lastTransitionTime":"2025-11-26T00:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.360863 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.360901 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.360913 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.360929 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.360939 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:42Z","lastTransitionTime":"2025-11-26T00:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.463744 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.463782 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.463791 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.463806 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.463823 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:42Z","lastTransitionTime":"2025-11-26T00:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.566705 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.566755 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.566766 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.566780 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.566792 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:42Z","lastTransitionTime":"2025-11-26T00:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.668601 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.668663 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.668673 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.668691 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.668701 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:42Z","lastTransitionTime":"2025-11-26T00:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.771339 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.771369 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.771379 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.771393 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.771409 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:42Z","lastTransitionTime":"2025-11-26T00:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.826625 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.826635 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:42 crc kubenswrapper[4766]: E1126 00:24:42.826773 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:42 crc kubenswrapper[4766]: E1126 00:24:42.826829 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.826671 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:42 crc kubenswrapper[4766]: E1126 00:24:42.826899 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.873960 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.874366 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.874509 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.874713 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.874902 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:42Z","lastTransitionTime":"2025-11-26T00:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.977017 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.977045 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.977053 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.977066 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:42 crc kubenswrapper[4766]: I1126 00:24:42.977075 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:42Z","lastTransitionTime":"2025-11-26T00:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.080099 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.080175 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.080197 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.080226 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.080250 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:43Z","lastTransitionTime":"2025-11-26T00:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.183061 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.183114 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.183131 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.183157 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.183174 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:43Z","lastTransitionTime":"2025-11-26T00:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.286357 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.286638 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.286757 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.286864 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.286959 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:43Z","lastTransitionTime":"2025-11-26T00:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.390463 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.390755 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.390870 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.390975 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.391075 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:43Z","lastTransitionTime":"2025-11-26T00:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.493579 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.493921 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.494011 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.494105 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.494178 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:43Z","lastTransitionTime":"2025-11-26T00:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.596293 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.596326 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.596334 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.596348 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.596356 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:43Z","lastTransitionTime":"2025-11-26T00:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.698434 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.698464 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.698472 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.698485 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.698494 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:43Z","lastTransitionTime":"2025-11-26T00:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.801773 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.801836 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.801850 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.801870 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.801885 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:43Z","lastTransitionTime":"2025-11-26T00:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.826575 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:43 crc kubenswrapper[4766]: E1126 00:24:43.827182 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.905355 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.905397 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.905408 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.905425 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:43 crc kubenswrapper[4766]: I1126 00:24:43.905436 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:43Z","lastTransitionTime":"2025-11-26T00:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.008477 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.008533 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.008544 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.008559 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.008568 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:44Z","lastTransitionTime":"2025-11-26T00:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.112023 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.112103 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.112124 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.112151 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.112175 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:44Z","lastTransitionTime":"2025-11-26T00:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.214493 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.214545 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.214555 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.214569 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.214578 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:44Z","lastTransitionTime":"2025-11-26T00:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.317109 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.317146 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.317158 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.317174 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.317185 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:44Z","lastTransitionTime":"2025-11-26T00:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.423699 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.423745 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.423759 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.423778 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.423793 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:44Z","lastTransitionTime":"2025-11-26T00:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.526735 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.526777 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.526790 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.526807 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.526820 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:44Z","lastTransitionTime":"2025-11-26T00:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.629482 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.629542 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.629560 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.629584 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.629601 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:44Z","lastTransitionTime":"2025-11-26T00:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.732708 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.732793 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.732815 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.732840 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.732889 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:44Z","lastTransitionTime":"2025-11-26T00:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.826267 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.826307 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:44 crc kubenswrapper[4766]: E1126 00:24:44.826390 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:44 crc kubenswrapper[4766]: E1126 00:24:44.826476 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.826556 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:44 crc kubenswrapper[4766]: E1126 00:24:44.826622 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.835866 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.835908 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.835917 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.835930 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.835938 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:44Z","lastTransitionTime":"2025-11-26T00:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.939178 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.939726 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.939823 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.939915 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:44 crc kubenswrapper[4766]: I1126 00:24:44.940003 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:44Z","lastTransitionTime":"2025-11-26T00:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.043109 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.043160 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.043169 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.043186 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.043197 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:45Z","lastTransitionTime":"2025-11-26T00:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.145980 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.146037 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.146053 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.146086 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.146107 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:45Z","lastTransitionTime":"2025-11-26T00:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.249009 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.249263 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.249326 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.249389 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.249445 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:45Z","lastTransitionTime":"2025-11-26T00:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.352071 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.352331 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.352427 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.352524 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.352620 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:45Z","lastTransitionTime":"2025-11-26T00:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.455166 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.455233 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.455247 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.455265 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.455276 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:45Z","lastTransitionTime":"2025-11-26T00:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.557616 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.557701 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.557728 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.557758 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.557779 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:45Z","lastTransitionTime":"2025-11-26T00:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.660294 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.660328 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.660339 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.660355 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.660366 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:45Z","lastTransitionTime":"2025-11-26T00:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.763058 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.763103 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.763116 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.763134 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.763149 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:45Z","lastTransitionTime":"2025-11-26T00:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.826200 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:45 crc kubenswrapper[4766]: E1126 00:24:45.826362 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.865298 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.865363 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.865380 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.865401 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.865418 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:45Z","lastTransitionTime":"2025-11-26T00:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.968132 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.968205 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.968217 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.968246 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:45 crc kubenswrapper[4766]: I1126 00:24:45.968259 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:45Z","lastTransitionTime":"2025-11-26T00:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.070508 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.070536 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.070545 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.070557 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.070568 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:46Z","lastTransitionTime":"2025-11-26T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.173363 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.173434 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.173446 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.173467 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.173479 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:46Z","lastTransitionTime":"2025-11-26T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.275983 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.276020 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.276031 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.276045 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.276054 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:46Z","lastTransitionTime":"2025-11-26T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.378734 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.378784 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.378798 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.378815 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.378825 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:46Z","lastTransitionTime":"2025-11-26T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.481057 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.481126 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.481141 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.481160 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.481171 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:46Z","lastTransitionTime":"2025-11-26T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.519488 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.519530 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.519542 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.519560 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.519573 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:46Z","lastTransitionTime":"2025-11-26T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:46 crc kubenswrapper[4766]: E1126 00:24:46.530951 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:46Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.534393 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.534445 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.534457 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.534472 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.534483 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:46Z","lastTransitionTime":"2025-11-26T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:46 crc kubenswrapper[4766]: E1126 00:24:46.549921 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:46Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.553088 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.553114 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.553124 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.553137 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.553147 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:46Z","lastTransitionTime":"2025-11-26T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:46 crc kubenswrapper[4766]: E1126 00:24:46.569119 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:46Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.573253 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.573275 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.573283 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.573296 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.573307 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:46Z","lastTransitionTime":"2025-11-26T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:46 crc kubenswrapper[4766]: E1126 00:24:46.584151 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:46Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.586864 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.586900 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.586912 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.586931 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.586944 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:46Z","lastTransitionTime":"2025-11-26T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:46 crc kubenswrapper[4766]: E1126 00:24:46.600670 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:46Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:46 crc kubenswrapper[4766]: E1126 00:24:46.600783 4766 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.602352 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.602384 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.602396 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.602412 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.602423 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:46Z","lastTransitionTime":"2025-11-26T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.704997 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.705024 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.705033 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.705045 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.705054 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:46Z","lastTransitionTime":"2025-11-26T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.807338 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.807373 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.807384 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.807399 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.807420 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:46Z","lastTransitionTime":"2025-11-26T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.826504 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.826580 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.826883 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:46 crc kubenswrapper[4766]: E1126 00:24:46.827100 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:46 crc kubenswrapper[4766]: E1126 00:24:46.827293 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:46 crc kubenswrapper[4766]: E1126 00:24:46.827701 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.828280 4766 scope.go:117] "RemoveContainer" containerID="c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a" Nov 26 00:24:46 crc kubenswrapper[4766]: E1126 00:24:46.828574 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fxnln_openshift-ovn-kubernetes(415fdacb-85c1-4265-89b7-6771a84ffc89)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.910293 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.910560 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.910670 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.910773 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:46 crc kubenswrapper[4766]: I1126 00:24:46.910855 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:46Z","lastTransitionTime":"2025-11-26T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.013194 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.013239 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.013250 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.013269 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.013279 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:47Z","lastTransitionTime":"2025-11-26T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.115693 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.115983 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.116067 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.116137 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.116198 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:47Z","lastTransitionTime":"2025-11-26T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.218230 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.218257 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.218265 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.218278 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.218287 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:47Z","lastTransitionTime":"2025-11-26T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.321392 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.321423 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.321432 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.321445 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.321455 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:47Z","lastTransitionTime":"2025-11-26T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.424592 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.424637 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.424674 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.424695 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.424707 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:47Z","lastTransitionTime":"2025-11-26T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.526735 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.526766 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.526775 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.526788 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.526796 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:47Z","lastTransitionTime":"2025-11-26T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.629806 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.629843 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.629860 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.629883 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.629897 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:47Z","lastTransitionTime":"2025-11-26T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.732781 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.732810 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.732819 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.732831 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.732840 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:47Z","lastTransitionTime":"2025-11-26T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.826191 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:47 crc kubenswrapper[4766]: E1126 00:24:47.826555 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.834793 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.834824 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.834832 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.834843 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.834852 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:47Z","lastTransitionTime":"2025-11-26T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.936614 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.936896 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.936986 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.937076 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:47 crc kubenswrapper[4766]: I1126 00:24:47.937167 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:47Z","lastTransitionTime":"2025-11-26T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.039949 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.039992 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.040004 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.040021 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.040033 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:48Z","lastTransitionTime":"2025-11-26T00:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.142519 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.142554 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.142564 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.142579 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.142590 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:48Z","lastTransitionTime":"2025-11-26T00:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.244317 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.244354 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.244364 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.244381 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.244394 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:48Z","lastTransitionTime":"2025-11-26T00:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.347684 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.347756 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.347780 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.347810 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.347881 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:48Z","lastTransitionTime":"2025-11-26T00:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.451222 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.451258 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.451267 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.451283 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.451293 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:48Z","lastTransitionTime":"2025-11-26T00:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.553328 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.553382 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.553400 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.553422 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.553438 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:48Z","lastTransitionTime":"2025-11-26T00:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.656429 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.656484 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.656497 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.656514 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.656526 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:48Z","lastTransitionTime":"2025-11-26T00:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.758708 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.758755 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.758768 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.758789 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.758800 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:48Z","lastTransitionTime":"2025-11-26T00:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.825937 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.826014 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:48 crc kubenswrapper[4766]: E1126 00:24:48.826116 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.825953 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:48 crc kubenswrapper[4766]: E1126 00:24:48.826270 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:48 crc kubenswrapper[4766]: E1126 00:24:48.826349 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.861831 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.861872 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.861885 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.861900 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.861911 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:48Z","lastTransitionTime":"2025-11-26T00:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.963800 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.963838 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.963849 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.963863 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:48 crc kubenswrapper[4766]: I1126 00:24:48.963875 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:48Z","lastTransitionTime":"2025-11-26T00:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.067023 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.067313 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.067446 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.067544 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.067624 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:49Z","lastTransitionTime":"2025-11-26T00:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.170296 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.170335 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.170347 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.170363 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.170374 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:49Z","lastTransitionTime":"2025-11-26T00:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.272621 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.272727 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.272784 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.272812 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.272828 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:49Z","lastTransitionTime":"2025-11-26T00:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.375222 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.375268 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.375277 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.375293 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.375302 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:49Z","lastTransitionTime":"2025-11-26T00:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.477037 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.477070 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.477079 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.477092 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.477106 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:49Z","lastTransitionTime":"2025-11-26T00:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.579051 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.579103 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.579116 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.579135 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.579147 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:49Z","lastTransitionTime":"2025-11-26T00:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.681586 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.681635 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.681648 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.681688 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.681701 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:49Z","lastTransitionTime":"2025-11-26T00:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.784108 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.784155 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.784170 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.784187 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.784198 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:49Z","lastTransitionTime":"2025-11-26T00:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.826287 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:49 crc kubenswrapper[4766]: E1126 00:24:49.826598 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.838390 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.845619 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:49Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.857632 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:49Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.870032 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:49Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.881127 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:49Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.886306 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.886346 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.886357 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.886375 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.886388 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:49Z","lastTransitionTime":"2025-11-26T00:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.893607 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:49Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.904906 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5z9ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6600efa-4858-41aa-a890-44ce262b63dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5z9ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:49Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.917715 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:49Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.936064 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:49Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.946077 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:49Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.956350 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"443534cf-37a7-4831-a581-9d04dee05ead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0874e594efd8c1df4af245ea08256417c1b6cfe83fd8ce08848ea6710081cf1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb484d4736b942d10794d5e39064bec67b5b30e9d1ac668f5a62dbc15119c3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-49ztc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:49Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.967894 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:49Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.978915 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:49Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.992574 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.992616 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.992625 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.992667 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.992678 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:49Z","lastTransitionTime":"2025-11-26T00:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:49 crc kubenswrapper[4766]: I1126 00:24:49.995752 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:49Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.012475 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:50Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.029592 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:30Z\\\",\\\"message\\\":\\\"iserver-crc after 0 failed attempt(s)\\\\nI1126 00:24:30.895910 6449 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1126 00:24:30.895884 6449 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1126 00:24:30.895916 6449 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI1126 00:24:30.895920 6449 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1126 00:24:30.895911 6449 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1126 00:24:30.895679 6449 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1126 00:24:30.895942 6449 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF1126 00:24:30.895947 6449 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fxnln_openshift-ovn-kubernetes(415fdacb-85c1-4265-89b7-6771a84ffc89)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:50Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.040409 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20c9b10f-a395-40ae-a0aa-b3f0b5f63682\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01950717656ec2766fa80a35e2f7a9136409c2331979c67bffa25b965864be0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2293d0b6924d276777e52994713be62a0f98ff846ecee3ed7f597835e18a8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2dd6961899214f94684d3cb24a3c75f9870bf0fc667ef4e2b5c25fdc4b0e5da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2b174d3a0b70785380dfbac2eb1062d84fafedff25f8322cc25aea9c362b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e2b174d3a0b70785380dfbac2eb1062d84fafedff25f8322cc25aea9c362b6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:50Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.051565 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:50Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.060737 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:50Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.095971 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.096008 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.096019 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.096035 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.096045 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:50Z","lastTransitionTime":"2025-11-26T00:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.198495 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.198550 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.198563 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.198577 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.198589 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:50Z","lastTransitionTime":"2025-11-26T00:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.234226 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs\") pod \"network-metrics-daemon-5z9ll\" (UID: \"b6600efa-4858-41aa-a890-44ce262b63dd\") " pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:50 crc kubenswrapper[4766]: E1126 00:24:50.234386 4766 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 00:24:50 crc kubenswrapper[4766]: E1126 00:24:50.234433 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs podName:b6600efa-4858-41aa-a890-44ce262b63dd nodeName:}" failed. No retries permitted until 2025-11-26 00:25:22.234416431 +0000 UTC m=+103.083186861 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs") pod "network-metrics-daemon-5z9ll" (UID: "b6600efa-4858-41aa-a890-44ce262b63dd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.301325 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.301373 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.301388 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.301406 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.301420 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:50Z","lastTransitionTime":"2025-11-26T00:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.404953 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.405016 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.405028 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.405044 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.405057 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:50Z","lastTransitionTime":"2025-11-26T00:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.507419 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.507461 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.507471 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.507489 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.507500 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:50Z","lastTransitionTime":"2025-11-26T00:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.610029 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.610069 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.610081 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.610097 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.610107 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:50Z","lastTransitionTime":"2025-11-26T00:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.712260 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.712299 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.712310 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.712328 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.712341 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:50Z","lastTransitionTime":"2025-11-26T00:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.815566 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.815601 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.815610 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.815625 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.815635 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:50Z","lastTransitionTime":"2025-11-26T00:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.826329 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.826364 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.826421 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:50 crc kubenswrapper[4766]: E1126 00:24:50.826477 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:50 crc kubenswrapper[4766]: E1126 00:24:50.826603 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:50 crc kubenswrapper[4766]: E1126 00:24:50.826767 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.917548 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.917587 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.917597 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.917610 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:50 crc kubenswrapper[4766]: I1126 00:24:50.917619 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:50Z","lastTransitionTime":"2025-11-26T00:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.019515 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.019589 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.019600 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.019612 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.019621 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:51Z","lastTransitionTime":"2025-11-26T00:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.122276 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.122344 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.122362 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.122377 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.122390 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:51Z","lastTransitionTime":"2025-11-26T00:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.224766 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.224810 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.224822 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.224838 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.224850 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:51Z","lastTransitionTime":"2025-11-26T00:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.326774 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.326824 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.326834 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.326848 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.326859 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:51Z","lastTransitionTime":"2025-11-26T00:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.429891 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.429934 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.429946 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.429964 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.429976 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:51Z","lastTransitionTime":"2025-11-26T00:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.531925 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.531973 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.531984 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.532014 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.532025 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:51Z","lastTransitionTime":"2025-11-26T00:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.633807 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.633852 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.633863 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.633882 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.633897 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:51Z","lastTransitionTime":"2025-11-26T00:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.736627 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.736842 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.736903 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.736936 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.736958 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:51Z","lastTransitionTime":"2025-11-26T00:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.826311 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:51 crc kubenswrapper[4766]: E1126 00:24:51.827420 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.839098 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.839135 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.839146 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.839160 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.839170 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:51Z","lastTransitionTime":"2025-11-26T00:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.942688 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.942715 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.942724 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.942737 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:51 crc kubenswrapper[4766]: I1126 00:24:51.942747 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:51Z","lastTransitionTime":"2025-11-26T00:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.045063 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.045163 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.045190 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.045221 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.045246 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:52Z","lastTransitionTime":"2025-11-26T00:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.147015 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.147049 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.147058 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.147076 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.147086 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:52Z","lastTransitionTime":"2025-11-26T00:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.249596 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.249683 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.249701 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.249756 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.249775 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:52Z","lastTransitionTime":"2025-11-26T00:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.351726 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.351765 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.351775 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.351787 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.351798 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:52Z","lastTransitionTime":"2025-11-26T00:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.454633 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.454698 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.454710 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.454729 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.454742 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:52Z","lastTransitionTime":"2025-11-26T00:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.534311 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-prqrp_234068f7-4097-4e6b-aac3-4bafe425c1f2/kube-multus/0.log" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.534363 4766 generic.go:334] "Generic (PLEG): container finished" podID="234068f7-4097-4e6b-aac3-4bafe425c1f2" containerID="aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c" exitCode=1 Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.534393 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-prqrp" event={"ID":"234068f7-4097-4e6b-aac3-4bafe425c1f2","Type":"ContainerDied","Data":"aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c"} Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.534794 4766 scope.go:117] "RemoveContainer" containerID="aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.549308 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:52Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.557298 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.557332 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.557342 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.557357 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.557368 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:52Z","lastTransitionTime":"2025-11-26T00:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.570178 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:52Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.581777 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:52Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.593782 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"443534cf-37a7-4831-a581-9d04dee05ead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0874e594efd8c1df4af245ea08256417c1b6cfe83fd8ce08848ea6710081cf1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb484d4736b942d10794d5e39064bec67b5b30e9d1ac668f5a62dbc15119c3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-49ztc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:52Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.606821 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c2c717-366a-4934-a0f5-3bb5d2d3f7ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a577316cc6eba55a72b6af823345750b50ca7cf138ec4393039ead51d70089b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0624ff399d744217874fd106831a1e945b55b700ad96cf3fc6cc6edf4f42ba0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0624ff399d744217874fd106831a1e945b55b700ad96cf3fc6cc6edf4f42ba0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:52Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.621566 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:52Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.636354 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:52Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.649929 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:52Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.659433 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.659470 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.659478 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.659492 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.659502 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:52Z","lastTransitionTime":"2025-11-26T00:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.665279 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:52Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.691228 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:30Z\\\",\\\"message\\\":\\\"iserver-crc after 0 failed attempt(s)\\\\nI1126 00:24:30.895910 6449 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1126 00:24:30.895884 6449 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1126 00:24:30.895916 6449 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI1126 00:24:30.895920 6449 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1126 00:24:30.895911 6449 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1126 00:24:30.895679 6449 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1126 00:24:30.895942 6449 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF1126 00:24:30.895947 6449 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fxnln_openshift-ovn-kubernetes(415fdacb-85c1-4265-89b7-6771a84ffc89)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:52Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.703890 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20c9b10f-a395-40ae-a0aa-b3f0b5f63682\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01950717656ec2766fa80a35e2f7a9136409c2331979c67bffa25b965864be0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2293d0b6924d276777e52994713be62a0f98ff846ecee3ed7f597835e18a8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2dd6961899214f94684d3cb24a3c75f9870bf0fc667ef4e2b5c25fdc4b0e5da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2b174d3a0b70785380dfbac2eb1062d84fafedff25f8322cc25aea9c362b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e2b174d3a0b70785380dfbac2eb1062d84fafedff25f8322cc25aea9c362b6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:52Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.717540 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:52Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.726767 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:52Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.740029 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:52Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.753016 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:52Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.764812 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.764870 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.764880 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.764902 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.764914 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:52Z","lastTransitionTime":"2025-11-26T00:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.768825 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:52Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.783528 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:52Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.797466 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:52Z\\\",\\\"message\\\":\\\"2025-11-26T00:24:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a427e75d-eccf-4d39-9926-137bf66d6947\\\\n2025-11-26T00:24:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a427e75d-eccf-4d39-9926-137bf66d6947 to /host/opt/cni/bin/\\\\n2025-11-26T00:24:07Z [verbose] multus-daemon started\\\\n2025-11-26T00:24:07Z [verbose] Readiness Indicator file check\\\\n2025-11-26T00:24:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:52Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.811666 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5z9ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6600efa-4858-41aa-a890-44ce262b63dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5z9ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:52Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.825929 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.826002 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.826070 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:52 crc kubenswrapper[4766]: E1126 00:24:52.826065 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:52 crc kubenswrapper[4766]: E1126 00:24:52.826131 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:52 crc kubenswrapper[4766]: E1126 00:24:52.826178 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.866834 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.866890 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.866902 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.866920 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.866932 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:52Z","lastTransitionTime":"2025-11-26T00:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.969495 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.969542 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.969550 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.969566 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:52 crc kubenswrapper[4766]: I1126 00:24:52.969578 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:52Z","lastTransitionTime":"2025-11-26T00:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.072739 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.072797 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.072810 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.072831 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.072844 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:53Z","lastTransitionTime":"2025-11-26T00:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.176348 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.176385 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.176397 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.176411 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.176419 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:53Z","lastTransitionTime":"2025-11-26T00:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.279133 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.279173 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.279181 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.279196 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.279206 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:53Z","lastTransitionTime":"2025-11-26T00:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.381759 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.381809 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.381821 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.381838 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.381847 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:53Z","lastTransitionTime":"2025-11-26T00:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.484334 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.484374 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.484384 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.484398 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.484409 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:53Z","lastTransitionTime":"2025-11-26T00:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.540335 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-prqrp_234068f7-4097-4e6b-aac3-4bafe425c1f2/kube-multus/0.log" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.540408 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-prqrp" event={"ID":"234068f7-4097-4e6b-aac3-4bafe425c1f2","Type":"ContainerStarted","Data":"58d961f26d62b958e729b2bd5129790335de3019e968595091099e5e416353d1"} Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.554209 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20c9b10f-a395-40ae-a0aa-b3f0b5f63682\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01950717656ec2766fa80a35e2f7a9136409c2331979c67bffa25b965864be0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2293d0b6924d276777e52994713be62a0f98ff846ecee3ed7f597835e18a8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2dd6961899214f94684d3cb24a3c75f9870bf0fc667ef4e2b5c25fdc4b0e5da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2b174d3a0b70785380dfbac2eb1062d84fafedff25f8322cc25aea9c362b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e2b174d3a0b70785380dfbac2eb1062d84fafedff25f8322cc25aea9c362b6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:53Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.567148 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:53Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.579116 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:53Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.587169 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.587218 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.587233 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.587251 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.587266 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:53Z","lastTransitionTime":"2025-11-26T00:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.593187 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:53Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.605757 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:53Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.624257 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:53Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.637604 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:53Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.655138 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58d961f26d62b958e729b2bd5129790335de3019e968595091099e5e416353d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:52Z\\\",\\\"message\\\":\\\"2025-11-26T00:24:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a427e75d-eccf-4d39-9926-137bf66d6947\\\\n2025-11-26T00:24:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a427e75d-eccf-4d39-9926-137bf66d6947 to /host/opt/cni/bin/\\\\n2025-11-26T00:24:07Z [verbose] multus-daemon started\\\\n2025-11-26T00:24:07Z [verbose] Readiness Indicator file check\\\\n2025-11-26T00:24:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:53Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.670524 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5z9ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6600efa-4858-41aa-a890-44ce262b63dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5z9ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:53Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.687569 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:53Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.689603 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.689674 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.689688 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.689709 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.689720 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:53Z","lastTransitionTime":"2025-11-26T00:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.709754 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:53Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.753626 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:53Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.766512 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"443534cf-37a7-4831-a581-9d04dee05ead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0874e594efd8c1df4af245ea08256417c1b6cfe83fd8ce08848ea6710081cf1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb484d4736b942d10794d5e39064bec67b5b30e9d1ac668f5a62dbc15119c3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-49ztc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:53Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.776059 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c2c717-366a-4934-a0f5-3bb5d2d3f7ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a577316cc6eba55a72b6af823345750b50ca7cf138ec4393039ead51d70089b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0624ff399d744217874fd106831a1e945b55b700ad96cf3fc6cc6edf4f42ba0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0624ff399d744217874fd106831a1e945b55b700ad96cf3fc6cc6edf4f42ba0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:53Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.787615 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:53Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.792710 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.792780 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.792798 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.792821 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.792837 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:53Z","lastTransitionTime":"2025-11-26T00:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.802352 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:53Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.816581 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:53Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.826032 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:53 crc kubenswrapper[4766]: E1126 00:24:53.826208 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.834222 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:53Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.854547 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:30Z\\\",\\\"message\\\":\\\"iserver-crc after 0 failed attempt(s)\\\\nI1126 00:24:30.895910 6449 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1126 00:24:30.895884 6449 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1126 00:24:30.895916 6449 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI1126 00:24:30.895920 6449 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1126 00:24:30.895911 6449 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1126 00:24:30.895679 6449 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1126 00:24:30.895942 6449 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF1126 00:24:30.895947 6449 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fxnln_openshift-ovn-kubernetes(415fdacb-85c1-4265-89b7-6771a84ffc89)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:53Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.895748 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.895799 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.895814 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.895838 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.895850 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:53Z","lastTransitionTime":"2025-11-26T00:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.998726 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.998778 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.998788 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.998805 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:53 crc kubenswrapper[4766]: I1126 00:24:53.998817 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:53Z","lastTransitionTime":"2025-11-26T00:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.102029 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.102082 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.102091 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.102113 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.102124 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:54Z","lastTransitionTime":"2025-11-26T00:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.205124 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.205178 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.205189 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.205209 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.205224 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:54Z","lastTransitionTime":"2025-11-26T00:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.307872 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.307905 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.307912 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.307926 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.307936 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:54Z","lastTransitionTime":"2025-11-26T00:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.409721 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.409755 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.409765 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.409777 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.409787 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:54Z","lastTransitionTime":"2025-11-26T00:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.512312 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.512356 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.512368 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.512385 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.512398 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:54Z","lastTransitionTime":"2025-11-26T00:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.614494 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.614553 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.614568 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.614589 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.614604 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:54Z","lastTransitionTime":"2025-11-26T00:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.717872 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.717904 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.717913 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.717929 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.717938 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:54Z","lastTransitionTime":"2025-11-26T00:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.821503 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.821552 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.821562 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.821578 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.821588 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:54Z","lastTransitionTime":"2025-11-26T00:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.825949 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.826015 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.826089 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:54 crc kubenswrapper[4766]: E1126 00:24:54.826100 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:54 crc kubenswrapper[4766]: E1126 00:24:54.826207 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:54 crc kubenswrapper[4766]: E1126 00:24:54.826301 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.924984 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.925026 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.925040 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.925056 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:54 crc kubenswrapper[4766]: I1126 00:24:54.925069 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:54Z","lastTransitionTime":"2025-11-26T00:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.028223 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.028279 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.028291 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.028312 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.028327 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:55Z","lastTransitionTime":"2025-11-26T00:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.130396 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.130438 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.130449 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.130464 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.130475 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:55Z","lastTransitionTime":"2025-11-26T00:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.234181 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.234238 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.234260 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.234318 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.234338 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:55Z","lastTransitionTime":"2025-11-26T00:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.337772 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.337840 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.337869 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.337915 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.337940 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:55Z","lastTransitionTime":"2025-11-26T00:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.440195 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.440257 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.440275 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.440297 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.440315 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:55Z","lastTransitionTime":"2025-11-26T00:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.543171 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.543204 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.543214 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.543229 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.543240 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:55Z","lastTransitionTime":"2025-11-26T00:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.646348 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.646419 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.646433 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.646447 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.646458 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:55Z","lastTransitionTime":"2025-11-26T00:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.749739 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.749806 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.749819 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.749837 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.749851 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:55Z","lastTransitionTime":"2025-11-26T00:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.826351 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:55 crc kubenswrapper[4766]: E1126 00:24:55.826846 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.852707 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.852756 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.852766 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.852782 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.852796 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:55Z","lastTransitionTime":"2025-11-26T00:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.955818 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.955893 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.955931 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.955951 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:55 crc kubenswrapper[4766]: I1126 00:24:55.955964 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:55Z","lastTransitionTime":"2025-11-26T00:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.059132 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.059209 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.059240 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.059277 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.059302 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:56Z","lastTransitionTime":"2025-11-26T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.162160 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.162216 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.162228 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.162243 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.162253 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:56Z","lastTransitionTime":"2025-11-26T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.265368 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.265455 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.265505 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.265540 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.265561 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:56Z","lastTransitionTime":"2025-11-26T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.367627 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.367694 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.367707 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.367726 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.367737 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:56Z","lastTransitionTime":"2025-11-26T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.471024 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.471078 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.471177 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.471215 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.471232 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:56Z","lastTransitionTime":"2025-11-26T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.573761 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.573830 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.573851 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.573877 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.573894 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:56Z","lastTransitionTime":"2025-11-26T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.608898 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.608953 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.608969 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.608991 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.609008 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:56Z","lastTransitionTime":"2025-11-26T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:56 crc kubenswrapper[4766]: E1126 00:24:56.621710 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:56Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.624988 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.625024 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.625036 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.625053 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.625066 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:56Z","lastTransitionTime":"2025-11-26T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:56 crc kubenswrapper[4766]: E1126 00:24:56.636880 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:56Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.640565 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.640603 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.640614 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.640629 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.640639 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:56Z","lastTransitionTime":"2025-11-26T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:56 crc kubenswrapper[4766]: E1126 00:24:56.654176 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:56Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.657901 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.657948 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.657958 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.657977 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.657988 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:56Z","lastTransitionTime":"2025-11-26T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:56 crc kubenswrapper[4766]: E1126 00:24:56.668996 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:56Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.672859 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.672987 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.673106 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.673216 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.673298 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:56Z","lastTransitionTime":"2025-11-26T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:56 crc kubenswrapper[4766]: E1126 00:24:56.686072 4766 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e5fef464-cee5-4575-b7c5-a728a420eb4a\\\",\\\"systemUUID\\\":\\\"4cba354f-139e-43ce-b547-af06ad27febb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:56Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:56 crc kubenswrapper[4766]: E1126 00:24:56.686537 4766 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.688285 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.688326 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.688337 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.688352 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.688362 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:56Z","lastTransitionTime":"2025-11-26T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.791176 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.791226 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.791243 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.791266 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.791283 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:56Z","lastTransitionTime":"2025-11-26T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.825729 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.825743 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:56 crc kubenswrapper[4766]: E1126 00:24:56.825833 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.825980 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:56 crc kubenswrapper[4766]: E1126 00:24:56.826164 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:56 crc kubenswrapper[4766]: E1126 00:24:56.826017 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.895050 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.895100 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.895118 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.895139 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.895155 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:56Z","lastTransitionTime":"2025-11-26T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.997294 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.997327 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.997337 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.997352 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:56 crc kubenswrapper[4766]: I1126 00:24:56.997363 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:56Z","lastTransitionTime":"2025-11-26T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.100287 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.100564 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.100804 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.101005 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.101366 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:57Z","lastTransitionTime":"2025-11-26T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.205498 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.205573 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.205599 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.205631 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.205716 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:57Z","lastTransitionTime":"2025-11-26T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.308388 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.308432 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.308455 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.308480 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.308504 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:57Z","lastTransitionTime":"2025-11-26T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.412563 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.412619 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.412636 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.412700 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.412725 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:57Z","lastTransitionTime":"2025-11-26T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.516568 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.516619 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.516636 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.516689 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.516707 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:57Z","lastTransitionTime":"2025-11-26T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.619461 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.619519 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.619538 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.619562 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.619582 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:57Z","lastTransitionTime":"2025-11-26T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.722746 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.722822 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.722843 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.722871 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.722897 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:57Z","lastTransitionTime":"2025-11-26T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.825610 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.825721 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.825747 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.825773 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.825802 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.825795 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:57Z","lastTransitionTime":"2025-11-26T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:57 crc kubenswrapper[4766]: E1126 00:24:57.825996 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.827132 4766 scope.go:117] "RemoveContainer" containerID="c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.928518 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.928580 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.928594 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.928614 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:57 crc kubenswrapper[4766]: I1126 00:24:57.928627 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:57Z","lastTransitionTime":"2025-11-26T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.032814 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.032859 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.032868 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.032886 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.032897 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:58Z","lastTransitionTime":"2025-11-26T00:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.135826 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.135870 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.135882 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.135898 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.135908 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:58Z","lastTransitionTime":"2025-11-26T00:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.239524 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.239604 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.239620 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.239640 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.239691 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:58Z","lastTransitionTime":"2025-11-26T00:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.343103 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.343180 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.343206 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.343237 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.343261 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:58Z","lastTransitionTime":"2025-11-26T00:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.446295 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.446348 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.446368 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.446393 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.446413 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:58Z","lastTransitionTime":"2025-11-26T00:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.551016 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.551089 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.551110 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.551136 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.551160 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:58Z","lastTransitionTime":"2025-11-26T00:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.654982 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.655050 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.655068 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.655090 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.655107 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:58Z","lastTransitionTime":"2025-11-26T00:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.757131 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.757159 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.757167 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.757179 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.757187 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:58Z","lastTransitionTime":"2025-11-26T00:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.826567 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.826678 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:24:58 crc kubenswrapper[4766]: E1126 00:24:58.826698 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.826815 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:24:58 crc kubenswrapper[4766]: E1126 00:24:58.826908 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:24:58 crc kubenswrapper[4766]: E1126 00:24:58.827019 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.859257 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.859319 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.859334 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.859353 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.859371 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:58Z","lastTransitionTime":"2025-11-26T00:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.962490 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.962526 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.962535 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.962548 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:58 crc kubenswrapper[4766]: I1126 00:24:58.962558 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:58Z","lastTransitionTime":"2025-11-26T00:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.065427 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.065468 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.065479 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.065496 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.065508 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:59Z","lastTransitionTime":"2025-11-26T00:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.168302 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.168382 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.168410 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.168444 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.168468 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:59Z","lastTransitionTime":"2025-11-26T00:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.271393 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.271436 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.271448 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.271466 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.271478 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:59Z","lastTransitionTime":"2025-11-26T00:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.373980 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.374024 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.374033 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.374049 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.374058 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:59Z","lastTransitionTime":"2025-11-26T00:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.476325 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.476388 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.476404 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.476427 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.476444 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:59Z","lastTransitionTime":"2025-11-26T00:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.563050 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fxnln_415fdacb-85c1-4265-89b7-6771a84ffc89/ovnkube-controller/2.log" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.566606 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerStarted","Data":"3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d"} Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.568707 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.579387 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.579429 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.579437 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.579452 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.579461 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:59Z","lastTransitionTime":"2025-11-26T00:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.600543 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.617449 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.634902 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.645543 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.657765 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58d961f26d62b958e729b2bd5129790335de3019e968595091099e5e416353d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:52Z\\\",\\\"message\\\":\\\"2025-11-26T00:24:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a427e75d-eccf-4d39-9926-137bf66d6947\\\\n2025-11-26T00:24:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a427e75d-eccf-4d39-9926-137bf66d6947 to /host/opt/cni/bin/\\\\n2025-11-26T00:24:07Z [verbose] multus-daemon started\\\\n2025-11-26T00:24:07Z [verbose] Readiness Indicator file check\\\\n2025-11-26T00:24:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.668121 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5z9ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6600efa-4858-41aa-a890-44ce262b63dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5z9ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.679624 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.681178 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.681204 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.681214 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.681226 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.681236 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:59Z","lastTransitionTime":"2025-11-26T00:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.698299 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.708900 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.719783 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"443534cf-37a7-4831-a581-9d04dee05ead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0874e594efd8c1df4af245ea08256417c1b6cfe83fd8ce08848ea6710081cf1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb484d4736b942d10794d5e39064bec67b5b30e9d1ac668f5a62dbc15119c3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-49ztc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.729888 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c2c717-366a-4934-a0f5-3bb5d2d3f7ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a577316cc6eba55a72b6af823345750b50ca7cf138ec4393039ead51d70089b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0624ff399d744217874fd106831a1e945b55b700ad96cf3fc6cc6edf4f42ba0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0624ff399d744217874fd106831a1e945b55b700ad96cf3fc6cc6edf4f42ba0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.744483 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.756153 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.768156 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.781605 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.783268 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.783322 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.783342 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.783358 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.783369 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:59Z","lastTransitionTime":"2025-11-26T00:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.805045 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:30Z\\\",\\\"message\\\":\\\"iserver-crc after 0 failed attempt(s)\\\\nI1126 00:24:30.895910 6449 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1126 00:24:30.895884 6449 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1126 00:24:30.895916 6449 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI1126 00:24:30.895920 6449 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1126 00:24:30.895911 6449 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1126 00:24:30.895679 6449 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1126 00:24:30.895942 6449 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF1126 00:24:30.895947 6449 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.817825 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20c9b10f-a395-40ae-a0aa-b3f0b5f63682\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01950717656ec2766fa80a35e2f7a9136409c2331979c67bffa25b965864be0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2293d0b6924d276777e52994713be62a0f98ff846ecee3ed7f597835e18a8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2dd6961899214f94684d3cb24a3c75f9870bf0fc667ef4e2b5c25fdc4b0e5da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2b174d3a0b70785380dfbac2eb1062d84fafedff25f8322cc25aea9c362b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e2b174d3a0b70785380dfbac2eb1062d84fafedff25f8322cc25aea9c362b6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.826820 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:24:59 crc kubenswrapper[4766]: E1126 00:24:59.827037 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.830089 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.839831 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.849176 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20c9b10f-a395-40ae-a0aa-b3f0b5f63682\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01950717656ec2766fa80a35e2f7a9136409c2331979c67bffa25b965864be0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2293d0b6924d276777e52994713be62a0f98ff846ecee3ed7f597835e18a8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2dd6961899214f94684d3cb24a3c75f9870bf0fc667ef4e2b5c25fdc4b0e5da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2b174d3a0b70785380dfbac2eb1062d84fafedff25f8322cc25aea9c362b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e2b174d3a0b70785380dfbac2eb1062d84fafedff25f8322cc25aea9c362b6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.860802 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.870540 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.881057 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.885637 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.885699 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.885711 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.885727 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.885738 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:59Z","lastTransitionTime":"2025-11-26T00:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.891738 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.906961 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.916419 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.926735 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58d961f26d62b958e729b2bd5129790335de3019e968595091099e5e416353d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:52Z\\\",\\\"message\\\":\\\"2025-11-26T00:24:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a427e75d-eccf-4d39-9926-137bf66d6947\\\\n2025-11-26T00:24:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a427e75d-eccf-4d39-9926-137bf66d6947 to /host/opt/cni/bin/\\\\n2025-11-26T00:24:07Z [verbose] multus-daemon started\\\\n2025-11-26T00:24:07Z [verbose] Readiness Indicator file check\\\\n2025-11-26T00:24:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.936501 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5z9ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6600efa-4858-41aa-a890-44ce262b63dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5z9ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.951240 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.967353 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.977066 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.986943 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"443534cf-37a7-4831-a581-9d04dee05ead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0874e594efd8c1df4af245ea08256417c1b6cfe83fd8ce08848ea6710081cf1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb484d4736b942d10794d5e39064bec67b5b30e9d1ac668f5a62dbc15119c3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-49ztc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.988484 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.988513 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.988524 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.988538 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.988549 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:24:59Z","lastTransitionTime":"2025-11-26T00:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:24:59 crc kubenswrapper[4766]: I1126 00:24:59.996500 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c2c717-366a-4934-a0f5-3bb5d2d3f7ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a577316cc6eba55a72b6af823345750b50ca7cf138ec4393039ead51d70089b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0624ff399d744217874fd106831a1e945b55b700ad96cf3fc6cc6edf4f42ba0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0624ff399d744217874fd106831a1e945b55b700ad96cf3fc6cc6edf4f42ba0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:24:59Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.008131 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.018517 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.028665 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.043650 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.064237 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:30Z\\\",\\\"message\\\":\\\"iserver-crc after 0 failed attempt(s)\\\\nI1126 00:24:30.895910 6449 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1126 00:24:30.895884 6449 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1126 00:24:30.895916 6449 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI1126 00:24:30.895920 6449 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1126 00:24:30.895911 6449 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1126 00:24:30.895679 6449 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1126 00:24:30.895942 6449 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF1126 00:24:30.895947 6449 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.091293 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.091350 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.091373 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.091432 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.091458 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:00Z","lastTransitionTime":"2025-11-26T00:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.194052 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.194185 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.194216 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.194246 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.194269 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:00Z","lastTransitionTime":"2025-11-26T00:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.296599 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.296632 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.296640 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.296688 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.296703 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:00Z","lastTransitionTime":"2025-11-26T00:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.406166 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.406267 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.406287 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.406325 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.406348 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:00Z","lastTransitionTime":"2025-11-26T00:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.509802 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.509871 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.509889 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.509913 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.509930 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:00Z","lastTransitionTime":"2025-11-26T00:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.572467 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fxnln_415fdacb-85c1-4265-89b7-6771a84ffc89/ovnkube-controller/3.log" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.573167 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fxnln_415fdacb-85c1-4265-89b7-6771a84ffc89/ovnkube-controller/2.log" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.575943 4766 generic.go:334] "Generic (PLEG): container finished" podID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerID="3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d" exitCode=1 Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.575995 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerDied","Data":"3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d"} Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.576043 4766 scope.go:117] "RemoveContainer" containerID="c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.577295 4766 scope.go:117] "RemoveContainer" containerID="3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d" Nov 26 00:25:00 crc kubenswrapper[4766]: E1126 00:25:00.577713 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fxnln_openshift-ovn-kubernetes(415fdacb-85c1-4265-89b7-6771a84ffc89)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.593254 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c2c717-366a-4934-a0f5-3bb5d2d3f7ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a577316cc6eba55a72b6af823345750b50ca7cf138ec4393039ead51d70089b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0624ff399d744217874fd106831a1e945b55b700ad96cf3fc6cc6edf4f42ba0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0624ff399d744217874fd106831a1e945b55b700ad96cf3fc6cc6edf4f42ba0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.612779 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.613012 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.613060 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.613078 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.613108 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.613129 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:00Z","lastTransitionTime":"2025-11-26T00:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.628537 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.647292 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.664564 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.692697 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c166b3906c7b42372189a1586af250b95133b4b94d13373943859dbe202a911a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:30Z\\\",\\\"message\\\":\\\"iserver-crc after 0 failed attempt(s)\\\\nI1126 00:24:30.895910 6449 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1126 00:24:30.895884 6449 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1126 00:24:30.895916 6449 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI1126 00:24:30.895920 6449 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1126 00:24:30.895911 6449 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1126 00:24:30.895679 6449 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1126 00:24:30.895942 6449 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF1126 00:24:30.895947 6449 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:25:00Z\\\",\\\"message\\\":\\\" 6816 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:59.627207 6816 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:59.627821 6816 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1126 00:24:59.627840 6816 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1126 00:24:59.627858 6816 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 00:24:59.627910 6816 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 00:24:59.627951 6816 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1126 00:24:59.627953 6816 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1126 00:24:59.627959 6816 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1126 00:24:59.627972 6816 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1126 00:24:59.627976 6816 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 00:24:59.627993 6816 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 00:24:59.628008 6816 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 00:24:59.628028 6816 factory.go:656] Stopping watch factory\\\\nI1126 00:24:59.628052 6816 ovnkube.go:599] Stopped ovnkube\\\\nI1126 00:24:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.705931 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20c9b10f-a395-40ae-a0aa-b3f0b5f63682\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01950717656ec2766fa80a35e2f7a9136409c2331979c67bffa25b965864be0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2293d0b6924d276777e52994713be62a0f98ff846ecee3ed7f597835e18a8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2dd6961899214f94684d3cb24a3c75f9870bf0fc667ef4e2b5c25fdc4b0e5da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2b174d3a0b70785380dfbac2eb1062d84fafedff25f8322cc25aea9c362b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e2b174d3a0b70785380dfbac2eb1062d84fafedff25f8322cc25aea9c362b6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.715515 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.715548 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.715556 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.715571 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.715580 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:00Z","lastTransitionTime":"2025-11-26T00:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.719244 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.733339 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.750451 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.767528 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.781758 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.794080 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.813624 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58d961f26d62b958e729b2bd5129790335de3019e968595091099e5e416353d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:52Z\\\",\\\"message\\\":\\\"2025-11-26T00:24:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a427e75d-eccf-4d39-9926-137bf66d6947\\\\n2025-11-26T00:24:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a427e75d-eccf-4d39-9926-137bf66d6947 to /host/opt/cni/bin/\\\\n2025-11-26T00:24:07Z [verbose] multus-daemon started\\\\n2025-11-26T00:24:07Z [verbose] Readiness Indicator file check\\\\n2025-11-26T00:24:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.818831 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.818914 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.818939 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.818970 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.818995 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:00Z","lastTransitionTime":"2025-11-26T00:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.826230 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:00 crc kubenswrapper[4766]: E1126 00:25:00.826337 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.826334 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.826374 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:00 crc kubenswrapper[4766]: E1126 00:25:00.826402 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:00 crc kubenswrapper[4766]: E1126 00:25:00.826485 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.830315 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5z9ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6600efa-4858-41aa-a890-44ce262b63dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5z9ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.842049 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.860509 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.873717 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.884457 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"443534cf-37a7-4831-a581-9d04dee05ead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0874e594efd8c1df4af245ea08256417c1b6cfe83fd8ce08848ea6710081cf1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb484d4736b942d10794d5e39064bec67b5b30e9d1ac668f5a62dbc15119c3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-49ztc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:00Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.921893 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.921962 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.921974 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.921996 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:00 crc kubenswrapper[4766]: I1126 00:25:00.922014 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:00Z","lastTransitionTime":"2025-11-26T00:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.024581 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.024677 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.024696 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.024716 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.024731 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:01Z","lastTransitionTime":"2025-11-26T00:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.126988 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.127049 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.127068 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.127092 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.127109 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:01Z","lastTransitionTime":"2025-11-26T00:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.229549 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.229611 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.229628 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.229676 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.229693 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:01Z","lastTransitionTime":"2025-11-26T00:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.332745 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.332821 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.332837 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.332860 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.332876 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:01Z","lastTransitionTime":"2025-11-26T00:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.435573 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.435700 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.435726 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.435754 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.435772 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:01Z","lastTransitionTime":"2025-11-26T00:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.538971 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.539053 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.539078 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.539109 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.539130 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:01Z","lastTransitionTime":"2025-11-26T00:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.585116 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fxnln_415fdacb-85c1-4265-89b7-6771a84ffc89/ovnkube-controller/3.log" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.591767 4766 scope.go:117] "RemoveContainer" containerID="3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d" Nov 26 00:25:01 crc kubenswrapper[4766]: E1126 00:25:01.592039 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fxnln_openshift-ovn-kubernetes(415fdacb-85c1-4265-89b7-6771a84ffc89)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.610922 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7f9d435-86f6-40ca-9e87-a1b4e44c18f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 00:23:53.217591 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 00:23:53.220150 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-227991850/tls.crt::/tmp/serving-cert-227991850/tls.key\\\\\\\"\\\\nI1126 00:23:58.529960 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 00:23:58.534151 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 00:23:58.534172 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 00:23:58.534189 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 00:23:58.534194 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 00:23:58.541287 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1126 00:23:58.541306 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1126 00:23:58.541312 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541317 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 00:23:58.541322 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 00:23:58.541325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 00:23:58.541328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 00:23:58.541332 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1126 00:23:58.543421 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.642239 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.642294 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.642312 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.642336 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.642353 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:01Z","lastTransitionTime":"2025-11-26T00:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.642241 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6564f5c2-6e79-4554-85f4-3911d362a41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58db8f537c721c688ea85d6b0c3f2f822fdc7c4a517c399760ee77fb3bbece41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://382970b4692ce7f4a5b11f5cb55c607850c4fed5e717c1aefb04da41a794ac72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://506dfd9c827da83dea868b2e2a776c828ccc6c16bd2e6973f2ed89585ebf3c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f437c27eafd944c464d395b171fb349b9f30c83385d3deee7a3e086009c7ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d22edd16d9ff62975e39402a824bcf8133bd431c4fb76e19ab81d85636c2cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6dce3c84a75f2021663b911801408d41fc645fa5a4beccf5d135b2170c1735a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c585047295da08a38e9b29a340591cf18f068990945d1edf4ff56022c43c6699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://951f231de245118b354438983fd8542538d9cf1524e5a1bf34135ac426cf9379\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.657439 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63471884-61f7-41d5-8967-e3f81eba90d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8941880e099dcff757577091518c2490b67b166de784341a0ca8896fb5788472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sb9l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wf9c2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.668700 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"443534cf-37a7-4831-a581-9d04dee05ead\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0874e594efd8c1df4af245ea08256417c1b6cfe83fd8ce08848ea6710081cf1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb484d4736b942d10794d5e39064bec67b5b30e9d1ac668f5a62dbc15119c3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pcwjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-49ztc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.679877 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c2c717-366a-4934-a0f5-3bb5d2d3f7ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a577316cc6eba55a72b6af823345750b50ca7cf138ec4393039ead51d70089b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0624ff399d744217874fd106831a1e945b55b700ad96cf3fc6cc6edf4f42ba0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0624ff399d744217874fd106831a1e945b55b700ad96cf3fc6cc6edf4f42ba0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.695513 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c18015cfee59f871e239ad3c53c8f1a73d218a560747f4d9c6f1719a96f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.707751 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4f12b1f2f3e6d43bc5e632a772f6ca38ce5898a345a053b6f28434a0daf83c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.718510 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.730828 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"beeb3530-3b3b-40d9-8b0e-1b927c64920a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa0e3c1bbfacd42e7b1b006404e6aaddb956b44c5682774ac667f98e467e67cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50ad1c979435821a3343b76126a80f29d2b949da0401c685413b4ed9a41f1977\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b02eaa726a51e5cbfb666a465d8fcc17c247e3a3ffebf8e539c3f431024d4d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfa6d242c58373dd78d41529126aeda1fc7ea54636debfc83d4ca8092e77bdd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e15ef753802dae7bcaa113c27451c02768cf112737589adbe360d8a8e2e4af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b4588d52144d1e82472a8dc76aeac769f81ee78200fd8a6ff025fe911d7d6f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27e91b450bebb047bcca05cda468a47e42ebb58ce6f7466738a3afd8832eed5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-55p6x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b8bk6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.744946 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.744995 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.745007 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.745023 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.745035 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:01Z","lastTransitionTime":"2025-11-26T00:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.748040 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"415fdacb-85c1-4265-89b7-6771a84ffc89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:25:00Z\\\",\\\"message\\\":\\\" 6816 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:59.627207 6816 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 00:24:59.627821 6816 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1126 00:24:59.627840 6816 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1126 00:24:59.627858 6816 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 00:24:59.627910 6816 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 00:24:59.627951 6816 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1126 00:24:59.627953 6816 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1126 00:24:59.627959 6816 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1126 00:24:59.627972 6816 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1126 00:24:59.627976 6816 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 00:24:59.627993 6816 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 00:24:59.628008 6816 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 00:24:59.628028 6816 factory.go:656] Stopping watch factory\\\\nI1126 00:24:59.628052 6816 ovnkube.go:599] Stopped ovnkube\\\\nI1126 00:24:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fxnln_openshift-ovn-kubernetes(415fdacb-85c1-4265-89b7-6771a84ffc89)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l696l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fxnln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.761404 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20c9b10f-a395-40ae-a0aa-b3f0b5f63682\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01950717656ec2766fa80a35e2f7a9136409c2331979c67bffa25b965864be0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2293d0b6924d276777e52994713be62a0f98ff846ecee3ed7f597835e18a8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2dd6961899214f94684d3cb24a3c75f9870bf0fc667ef4e2b5c25fdc4b0e5da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e2b174d3a0b70785380dfbac2eb1062d84fafedff25f8322cc25aea9c362b6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e2b174d3a0b70785380dfbac2eb1062d84fafedff25f8322cc25aea9c362b6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.775765 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.786680 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8s7f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561423b1-063f-4968-a1b3-4187bb470b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba36bbf2b52d2e447842e050f56f8fbfe172d3d1b9d9be6903f376b19026da71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8s7f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.798211 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c03689c3-db5c-4ac1-9c83-ca3de76eace5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac7078b152097a389f6fb981041691859828af0a09b304caf912447175be3d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcba97f96e462bb3385955d2e98b42681514d4d182283f8e96ee26f61be9275e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6be26f2312bf15a6b9925811cb3c3ebfa94c3dd7c8b35602a36b0a31d8359b0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:23:39Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.811800 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.825361 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T00:23:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28f7cb1993b6fbb3dd0b3e685f4abe3ad5efa5896ad160db36678fbd7e2f4a62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a2bfe1a51aba2c5c891e78f105bc574fd959c36fa3f494e9aa6acc71954fa5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:23:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.825693 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:01 crc kubenswrapper[4766]: E1126 00:25:01.825817 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.835615 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dm6rv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95807d5c-2f7e-40f9-9116-a7e4766b85c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85deb49249eda41c77830b5b1fc2c93b3261f9472ad25cd49f3d92526a582c6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dm6rv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.847224 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.847273 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.847285 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.847302 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.847316 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:01Z","lastTransitionTime":"2025-11-26T00:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.848351 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-prqrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"234068f7-4097-4e6b-aac3-4bafe425c1f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58d961f26d62b958e729b2bd5129790335de3019e968595091099e5e416353d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T00:24:52Z\\\",\\\"message\\\":\\\"2025-11-26T00:24:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a427e75d-eccf-4d39-9926-137bf66d6947\\\\n2025-11-26T00:24:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a427e75d-eccf-4d39-9926-137bf66d6947 to /host/opt/cni/bin/\\\\n2025-11-26T00:24:07Z [verbose] multus-daemon started\\\\n2025-11-26T00:24:07Z [verbose] Readiness Indicator file check\\\\n2025-11-26T00:24:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T00:24:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T00:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kchh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-prqrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.858395 4766 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-5z9ll" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6600efa-4858-41aa-a890-44ce262b63dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T00:24:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnqtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T00:24:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-5z9ll\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T00:25:01Z is after 2025-08-24T17:21:41Z" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.949796 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.949844 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.949854 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.949868 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:01 crc kubenswrapper[4766]: I1126 00:25:01.949880 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:01Z","lastTransitionTime":"2025-11-26T00:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.052693 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.052869 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.052895 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.052917 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.052932 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:02Z","lastTransitionTime":"2025-11-26T00:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.156426 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.156512 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.156536 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.156563 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.156582 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:02Z","lastTransitionTime":"2025-11-26T00:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.259935 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.259985 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.260003 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.260023 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.260037 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:02Z","lastTransitionTime":"2025-11-26T00:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.363341 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.363398 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.363423 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.363450 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.363470 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:02Z","lastTransitionTime":"2025-11-26T00:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.466561 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.467186 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.467212 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.467240 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.467301 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:02Z","lastTransitionTime":"2025-11-26T00:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.569996 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.570207 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.570288 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.570413 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.570515 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:02Z","lastTransitionTime":"2025-11-26T00:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.666719 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.666906 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:02 crc kubenswrapper[4766]: E1126 00:25:02.666979 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:06.666933714 +0000 UTC m=+147.515704174 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:25:02 crc kubenswrapper[4766]: E1126 00:25:02.667053 4766 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.667103 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:02 crc kubenswrapper[4766]: E1126 00:25:02.667127 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 00:26:06.667105358 +0000 UTC m=+147.515875998 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.667171 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.667239 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:02 crc kubenswrapper[4766]: E1126 00:25:02.667381 4766 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 00:25:02 crc kubenswrapper[4766]: E1126 00:25:02.667432 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 00:26:06.667420706 +0000 UTC m=+147.516191366 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 00:25:02 crc kubenswrapper[4766]: E1126 00:25:02.667458 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 00:25:02 crc kubenswrapper[4766]: E1126 00:25:02.667485 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 00:25:02 crc kubenswrapper[4766]: E1126 00:25:02.667504 4766 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:25:02 crc kubenswrapper[4766]: E1126 00:25:02.667579 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 00:26:06.66756328 +0000 UTC m=+147.516333740 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:25:02 crc kubenswrapper[4766]: E1126 00:25:02.667761 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 00:25:02 crc kubenswrapper[4766]: E1126 00:25:02.667850 4766 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 00:25:02 crc kubenswrapper[4766]: E1126 00:25:02.667919 4766 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:25:02 crc kubenswrapper[4766]: E1126 00:25:02.668028 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 00:26:06.668014901 +0000 UTC m=+147.516785541 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.673112 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.673162 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.673174 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.673191 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.673205 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:02Z","lastTransitionTime":"2025-11-26T00:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.776246 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.776320 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.776337 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.776357 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.776372 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:02Z","lastTransitionTime":"2025-11-26T00:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.826113 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.826140 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.826789 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:02 crc kubenswrapper[4766]: E1126 00:25:02.826383 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:02 crc kubenswrapper[4766]: E1126 00:25:02.827460 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:02 crc kubenswrapper[4766]: E1126 00:25:02.827464 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.879180 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.879222 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.879231 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.879248 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.879260 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:02Z","lastTransitionTime":"2025-11-26T00:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.982059 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.982113 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.982128 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.982148 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:02 crc kubenswrapper[4766]: I1126 00:25:02.982160 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:02Z","lastTransitionTime":"2025-11-26T00:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.083849 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.083884 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.083895 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.083911 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.083923 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:03Z","lastTransitionTime":"2025-11-26T00:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.186200 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.186263 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.186272 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.186286 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.186295 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:03Z","lastTransitionTime":"2025-11-26T00:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.288991 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.289051 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.289067 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.289089 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.289105 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:03Z","lastTransitionTime":"2025-11-26T00:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.391804 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.391862 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.391875 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.391898 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.391912 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:03Z","lastTransitionTime":"2025-11-26T00:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.494516 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.494887 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.494992 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.495089 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.495182 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:03Z","lastTransitionTime":"2025-11-26T00:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.596985 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.597045 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.597057 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.597070 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.597081 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:03Z","lastTransitionTime":"2025-11-26T00:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.699728 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.699769 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.699779 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.699794 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.699804 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:03Z","lastTransitionTime":"2025-11-26T00:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.802512 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.802556 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.802570 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.802589 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.802600 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:03Z","lastTransitionTime":"2025-11-26T00:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.826475 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:03 crc kubenswrapper[4766]: E1126 00:25:03.826741 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.904432 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.904462 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.904471 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.904485 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:03 crc kubenswrapper[4766]: I1126 00:25:03.904496 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:03Z","lastTransitionTime":"2025-11-26T00:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.006752 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.006994 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.007057 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.007122 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.007190 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:04Z","lastTransitionTime":"2025-11-26T00:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.110220 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.110295 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.110331 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.110348 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.110360 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:04Z","lastTransitionTime":"2025-11-26T00:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.212460 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.212494 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.212507 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.212527 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.212544 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:04Z","lastTransitionTime":"2025-11-26T00:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.315837 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.316045 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.316117 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.316236 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.316302 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:04Z","lastTransitionTime":"2025-11-26T00:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.418568 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.418620 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.418635 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.418677 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.418691 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:04Z","lastTransitionTime":"2025-11-26T00:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.520884 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.520922 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.520933 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.520951 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.520964 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:04Z","lastTransitionTime":"2025-11-26T00:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.623681 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.623710 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.623719 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.623732 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.623740 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:04Z","lastTransitionTime":"2025-11-26T00:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.726575 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.726625 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.726644 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.726719 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.726749 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:04Z","lastTransitionTime":"2025-11-26T00:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.826370 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.826415 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.826432 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:04 crc kubenswrapper[4766]: E1126 00:25:04.826607 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:04 crc kubenswrapper[4766]: E1126 00:25:04.826718 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:04 crc kubenswrapper[4766]: E1126 00:25:04.826843 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.828785 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.828810 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.828819 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.828832 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.828842 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:04Z","lastTransitionTime":"2025-11-26T00:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.932298 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.932729 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.932944 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.933096 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:04 crc kubenswrapper[4766]: I1126 00:25:04.933243 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:04Z","lastTransitionTime":"2025-11-26T00:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.038750 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.038819 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.038840 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.038868 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.038886 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:05Z","lastTransitionTime":"2025-11-26T00:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.142693 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.142763 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.142785 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.142815 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.142837 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:05Z","lastTransitionTime":"2025-11-26T00:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.245517 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.246038 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.246106 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.246212 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.246286 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:05Z","lastTransitionTime":"2025-11-26T00:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.348526 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.349020 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.349127 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.349246 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.349337 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:05Z","lastTransitionTime":"2025-11-26T00:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.452356 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.452423 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.452440 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.452465 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.452482 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:05Z","lastTransitionTime":"2025-11-26T00:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.556310 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.556364 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.556376 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.556394 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.556409 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:05Z","lastTransitionTime":"2025-11-26T00:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.659021 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.659079 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.659097 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.659121 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.659138 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:05Z","lastTransitionTime":"2025-11-26T00:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.763218 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.763270 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.763282 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.763298 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.763311 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:05Z","lastTransitionTime":"2025-11-26T00:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.827025 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:05 crc kubenswrapper[4766]: E1126 00:25:05.827238 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.866929 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.867061 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.867083 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.867111 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.867130 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:05Z","lastTransitionTime":"2025-11-26T00:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.969744 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.969786 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.969794 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.969809 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:05 crc kubenswrapper[4766]: I1126 00:25:05.969819 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:05Z","lastTransitionTime":"2025-11-26T00:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.072123 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.072172 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.072183 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.072197 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.072206 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:06Z","lastTransitionTime":"2025-11-26T00:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.175733 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.175796 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.175843 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.175869 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.175887 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:06Z","lastTransitionTime":"2025-11-26T00:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.278821 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.278906 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.278925 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.278951 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.278970 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:06Z","lastTransitionTime":"2025-11-26T00:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.381580 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.381632 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.381644 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.381684 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.381694 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:06Z","lastTransitionTime":"2025-11-26T00:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.484463 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.484512 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.484526 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.484547 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.484559 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:06Z","lastTransitionTime":"2025-11-26T00:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.587417 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.587469 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.587488 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.587511 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.587526 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:06Z","lastTransitionTime":"2025-11-26T00:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.689985 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.690023 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.690035 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.690051 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.690062 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:06Z","lastTransitionTime":"2025-11-26T00:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.792264 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.792300 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.792312 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.792327 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.792337 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:06Z","lastTransitionTime":"2025-11-26T00:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.826155 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.826197 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.826197 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:06 crc kubenswrapper[4766]: E1126 00:25:06.826310 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:06 crc kubenswrapper[4766]: E1126 00:25:06.826381 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:06 crc kubenswrapper[4766]: E1126 00:25:06.826445 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.861704 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.861742 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.861751 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.861768 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.861777 4766 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T00:25:06Z","lastTransitionTime":"2025-11-26T00:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.913725 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-txnff"] Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.914129 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-txnff" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.915752 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.916290 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.916373 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.916525 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.944293 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=33.944275169 podStartE2EDuration="33.944275169s" podCreationTimestamp="2025-11-26 00:24:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:25:06.941461981 +0000 UTC m=+87.790232431" watchObservedRunningTime="2025-11-26 00:25:06.944275169 +0000 UTC m=+87.793045599" Nov 26 00:25:06 crc kubenswrapper[4766]: I1126 00:25:06.967211 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-8s7f9" podStartSLOduration=63.967196146 podStartE2EDuration="1m3.967196146s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:25:06.966293514 +0000 UTC m=+87.815063944" watchObservedRunningTime="2025-11-26 00:25:06.967196146 +0000 UTC m=+87.815966576" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.005808 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-dm6rv" podStartSLOduration=64.005787243 podStartE2EDuration="1m4.005787243s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:25:06.984802224 +0000 UTC m=+87.833572654" watchObservedRunningTime="2025-11-26 00:25:07.005787243 +0000 UTC m=+87.854557673" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.016959 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-txnff\" (UID: \"2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-txnff" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.017014 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-txnff\" (UID: \"2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-txnff" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.017050 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-txnff\" (UID: \"2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-txnff" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.017076 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-txnff\" (UID: \"2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-txnff" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.017098 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-txnff\" (UID: \"2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-txnff" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.018405 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-prqrp" podStartSLOduration=64.01839201 podStartE2EDuration="1m4.01839201s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:25:07.006089991 +0000 UTC m=+87.854860421" watchObservedRunningTime="2025-11-26 00:25:07.01839201 +0000 UTC m=+87.867162440" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.056944 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=68.056922446 podStartE2EDuration="1m8.056922446s" podCreationTimestamp="2025-11-26 00:23:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:25:07.039106423 +0000 UTC m=+87.887876883" watchObservedRunningTime="2025-11-26 00:25:07.056922446 +0000 UTC m=+87.905692886" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.108440 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-49ztc" podStartSLOduration=63.108421217 podStartE2EDuration="1m3.108421217s" podCreationTimestamp="2025-11-26 00:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:25:07.091380553 +0000 UTC m=+87.940150983" watchObservedRunningTime="2025-11-26 00:25:07.108421217 +0000 UTC m=+87.957191647" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.108733 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=69.108728714 podStartE2EDuration="1m9.108728714s" podCreationTimestamp="2025-11-26 00:23:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:25:07.108233382 +0000 UTC m=+87.957003822" watchObservedRunningTime="2025-11-26 00:25:07.108728714 +0000 UTC m=+87.957499144" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.118283 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-txnff\" (UID: \"2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-txnff" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.118319 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-txnff\" (UID: \"2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-txnff" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.118348 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-txnff\" (UID: \"2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-txnff" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.118365 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-txnff\" (UID: \"2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-txnff" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.118381 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-txnff\" (UID: \"2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-txnff" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.118459 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-txnff\" (UID: \"2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-txnff" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.118459 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-txnff\" (UID: \"2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-txnff" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.119194 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-txnff\" (UID: \"2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-txnff" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.132145 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=69.132125183 podStartE2EDuration="1m9.132125183s" podCreationTimestamp="2025-11-26 00:23:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:25:07.131762394 +0000 UTC m=+87.980532824" watchObservedRunningTime="2025-11-26 00:25:07.132125183 +0000 UTC m=+87.980895613" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.133667 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-txnff\" (UID: \"2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-txnff" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.143615 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podStartSLOduration=64.143595751 podStartE2EDuration="1m4.143595751s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:25:07.143338215 +0000 UTC m=+87.992108645" watchObservedRunningTime="2025-11-26 00:25:07.143595751 +0000 UTC m=+87.992366181" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.146311 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-txnff\" (UID: \"2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-txnff" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.191347 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-b8bk6" podStartSLOduration=64.191325741 podStartE2EDuration="1m4.191325741s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:25:07.169856219 +0000 UTC m=+88.018626669" watchObservedRunningTime="2025-11-26 00:25:07.191325741 +0000 UTC m=+88.040096171" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.203139 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=18.203122558 podStartE2EDuration="18.203122558s" podCreationTimestamp="2025-11-26 00:24:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:25:07.202487812 +0000 UTC m=+88.051258252" watchObservedRunningTime="2025-11-26 00:25:07.203122558 +0000 UTC m=+88.051892988" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.228349 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-txnff" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.608908 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-txnff" event={"ID":"2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3","Type":"ContainerStarted","Data":"2056d6dc79c091a20d3cfaf1e86cfb7ab46a1d070dc46dd08a9551ac80174a22"} Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.609255 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-txnff" event={"ID":"2d904dfd-deb7-4c4c-b17e-b8f152bd7ed3","Type":"ContainerStarted","Data":"539987515a20b5e2cff76c76f9ab125dfbf72cc3aad9e4ad050ddbfe68c12f9f"} Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.623095 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-txnff" podStartSLOduration=64.62307639 podStartE2EDuration="1m4.62307639s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:25:07.622178569 +0000 UTC m=+88.470948999" watchObservedRunningTime="2025-11-26 00:25:07.62307639 +0000 UTC m=+88.471846840" Nov 26 00:25:07 crc kubenswrapper[4766]: I1126 00:25:07.826003 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:07 crc kubenswrapper[4766]: E1126 00:25:07.826109 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:08 crc kubenswrapper[4766]: I1126 00:25:08.826358 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:08 crc kubenswrapper[4766]: I1126 00:25:08.826361 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:08 crc kubenswrapper[4766]: E1126 00:25:08.826495 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:08 crc kubenswrapper[4766]: I1126 00:25:08.826530 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:08 crc kubenswrapper[4766]: E1126 00:25:08.826641 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:08 crc kubenswrapper[4766]: E1126 00:25:08.826788 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:09 crc kubenswrapper[4766]: I1126 00:25:09.826311 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:09 crc kubenswrapper[4766]: E1126 00:25:09.829073 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:10 crc kubenswrapper[4766]: I1126 00:25:10.826693 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:10 crc kubenswrapper[4766]: I1126 00:25:10.826732 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:10 crc kubenswrapper[4766]: I1126 00:25:10.826820 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:10 crc kubenswrapper[4766]: E1126 00:25:10.826882 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:10 crc kubenswrapper[4766]: E1126 00:25:10.827007 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:10 crc kubenswrapper[4766]: E1126 00:25:10.827124 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:11 crc kubenswrapper[4766]: I1126 00:25:11.826687 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:11 crc kubenswrapper[4766]: E1126 00:25:11.826792 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:12 crc kubenswrapper[4766]: I1126 00:25:12.826438 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:12 crc kubenswrapper[4766]: I1126 00:25:12.826521 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:12 crc kubenswrapper[4766]: E1126 00:25:12.826563 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:12 crc kubenswrapper[4766]: I1126 00:25:12.826438 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:12 crc kubenswrapper[4766]: E1126 00:25:12.826683 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:12 crc kubenswrapper[4766]: E1126 00:25:12.826769 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:13 crc kubenswrapper[4766]: I1126 00:25:13.826417 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:13 crc kubenswrapper[4766]: E1126 00:25:13.826571 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:14 crc kubenswrapper[4766]: I1126 00:25:14.825838 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:14 crc kubenswrapper[4766]: I1126 00:25:14.825870 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:14 crc kubenswrapper[4766]: E1126 00:25:14.825974 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:14 crc kubenswrapper[4766]: I1126 00:25:14.826085 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:14 crc kubenswrapper[4766]: E1126 00:25:14.826187 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:14 crc kubenswrapper[4766]: E1126 00:25:14.826264 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:14 crc kubenswrapper[4766]: I1126 00:25:14.827514 4766 scope.go:117] "RemoveContainer" containerID="3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d" Nov 26 00:25:14 crc kubenswrapper[4766]: E1126 00:25:14.828008 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fxnln_openshift-ovn-kubernetes(415fdacb-85c1-4265-89b7-6771a84ffc89)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" Nov 26 00:25:15 crc kubenswrapper[4766]: I1126 00:25:15.826251 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:15 crc kubenswrapper[4766]: E1126 00:25:15.826393 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:16 crc kubenswrapper[4766]: I1126 00:25:16.826213 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:16 crc kubenswrapper[4766]: I1126 00:25:16.826289 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:16 crc kubenswrapper[4766]: E1126 00:25:16.826855 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:16 crc kubenswrapper[4766]: I1126 00:25:16.826289 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:16 crc kubenswrapper[4766]: E1126 00:25:16.826983 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:16 crc kubenswrapper[4766]: E1126 00:25:16.827085 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:17 crc kubenswrapper[4766]: I1126 00:25:17.826725 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:17 crc kubenswrapper[4766]: E1126 00:25:17.827458 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:18 crc kubenswrapper[4766]: I1126 00:25:18.825827 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:18 crc kubenswrapper[4766]: I1126 00:25:18.826045 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:18 crc kubenswrapper[4766]: E1126 00:25:18.826537 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:18 crc kubenswrapper[4766]: E1126 00:25:18.826991 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:18 crc kubenswrapper[4766]: I1126 00:25:18.827031 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:18 crc kubenswrapper[4766]: E1126 00:25:18.827677 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:19 crc kubenswrapper[4766]: I1126 00:25:19.825935 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:19 crc kubenswrapper[4766]: E1126 00:25:19.829185 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:20 crc kubenswrapper[4766]: I1126 00:25:20.826201 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:20 crc kubenswrapper[4766]: I1126 00:25:20.826267 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:20 crc kubenswrapper[4766]: E1126 00:25:20.826601 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:20 crc kubenswrapper[4766]: E1126 00:25:20.826805 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:20 crc kubenswrapper[4766]: I1126 00:25:20.827013 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:20 crc kubenswrapper[4766]: E1126 00:25:20.827109 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:21 crc kubenswrapper[4766]: I1126 00:25:21.826290 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:21 crc kubenswrapper[4766]: E1126 00:25:21.826943 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:22 crc kubenswrapper[4766]: I1126 00:25:22.286526 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs\") pod \"network-metrics-daemon-5z9ll\" (UID: \"b6600efa-4858-41aa-a890-44ce262b63dd\") " pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:22 crc kubenswrapper[4766]: E1126 00:25:22.286697 4766 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 00:25:22 crc kubenswrapper[4766]: E1126 00:25:22.286754 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs podName:b6600efa-4858-41aa-a890-44ce262b63dd nodeName:}" failed. No retries permitted until 2025-11-26 00:26:26.286736238 +0000 UTC m=+167.135506668 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs") pod "network-metrics-daemon-5z9ll" (UID: "b6600efa-4858-41aa-a890-44ce262b63dd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 00:25:22 crc kubenswrapper[4766]: I1126 00:25:22.826261 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:22 crc kubenswrapper[4766]: E1126 00:25:22.826740 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:22 crc kubenswrapper[4766]: I1126 00:25:22.826835 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:22 crc kubenswrapper[4766]: E1126 00:25:22.826969 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:22 crc kubenswrapper[4766]: I1126 00:25:22.826779 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:22 crc kubenswrapper[4766]: E1126 00:25:22.827056 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:23 crc kubenswrapper[4766]: I1126 00:25:23.826678 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:23 crc kubenswrapper[4766]: E1126 00:25:23.826826 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:24 crc kubenswrapper[4766]: I1126 00:25:24.826337 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:24 crc kubenswrapper[4766]: I1126 00:25:24.826369 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:24 crc kubenswrapper[4766]: E1126 00:25:24.826456 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:24 crc kubenswrapper[4766]: I1126 00:25:24.827043 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:24 crc kubenswrapper[4766]: E1126 00:25:24.827195 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:24 crc kubenswrapper[4766]: E1126 00:25:24.827386 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:25 crc kubenswrapper[4766]: I1126 00:25:25.826567 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:25 crc kubenswrapper[4766]: E1126 00:25:25.826825 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:26 crc kubenswrapper[4766]: I1126 00:25:26.826678 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:26 crc kubenswrapper[4766]: E1126 00:25:26.826825 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:26 crc kubenswrapper[4766]: I1126 00:25:26.826696 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:26 crc kubenswrapper[4766]: E1126 00:25:26.826911 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:26 crc kubenswrapper[4766]: I1126 00:25:26.826683 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:26 crc kubenswrapper[4766]: E1126 00:25:26.826998 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:27 crc kubenswrapper[4766]: I1126 00:25:27.825848 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:27 crc kubenswrapper[4766]: E1126 00:25:27.826046 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:28 crc kubenswrapper[4766]: I1126 00:25:28.825711 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:28 crc kubenswrapper[4766]: I1126 00:25:28.825858 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:28 crc kubenswrapper[4766]: I1126 00:25:28.825873 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:28 crc kubenswrapper[4766]: E1126 00:25:28.826848 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:28 crc kubenswrapper[4766]: E1126 00:25:28.826950 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:28 crc kubenswrapper[4766]: I1126 00:25:28.827026 4766 scope.go:117] "RemoveContainer" containerID="3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d" Nov 26 00:25:28 crc kubenswrapper[4766]: E1126 00:25:28.827076 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:28 crc kubenswrapper[4766]: E1126 00:25:28.827186 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fxnln_openshift-ovn-kubernetes(415fdacb-85c1-4265-89b7-6771a84ffc89)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" Nov 26 00:25:29 crc kubenswrapper[4766]: I1126 00:25:29.826111 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:29 crc kubenswrapper[4766]: E1126 00:25:29.828112 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:30 crc kubenswrapper[4766]: I1126 00:25:30.826123 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:30 crc kubenswrapper[4766]: I1126 00:25:30.826195 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:30 crc kubenswrapper[4766]: I1126 00:25:30.826254 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:30 crc kubenswrapper[4766]: E1126 00:25:30.826428 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:30 crc kubenswrapper[4766]: E1126 00:25:30.826855 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:30 crc kubenswrapper[4766]: E1126 00:25:30.827075 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:31 crc kubenswrapper[4766]: I1126 00:25:31.826115 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:31 crc kubenswrapper[4766]: E1126 00:25:31.826317 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:32 crc kubenswrapper[4766]: I1126 00:25:32.825863 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:32 crc kubenswrapper[4766]: I1126 00:25:32.825987 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:32 crc kubenswrapper[4766]: I1126 00:25:32.825878 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:32 crc kubenswrapper[4766]: E1126 00:25:32.826090 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:32 crc kubenswrapper[4766]: E1126 00:25:32.826165 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:32 crc kubenswrapper[4766]: E1126 00:25:32.826261 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:33 crc kubenswrapper[4766]: I1126 00:25:33.826797 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:33 crc kubenswrapper[4766]: E1126 00:25:33.826988 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:34 crc kubenswrapper[4766]: I1126 00:25:34.825956 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:34 crc kubenswrapper[4766]: I1126 00:25:34.825973 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:34 crc kubenswrapper[4766]: I1126 00:25:34.826044 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:34 crc kubenswrapper[4766]: E1126 00:25:34.826075 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:34 crc kubenswrapper[4766]: E1126 00:25:34.826198 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:34 crc kubenswrapper[4766]: E1126 00:25:34.826274 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:35 crc kubenswrapper[4766]: I1126 00:25:35.826858 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:35 crc kubenswrapper[4766]: E1126 00:25:35.827055 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:36 crc kubenswrapper[4766]: I1126 00:25:36.826031 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:36 crc kubenswrapper[4766]: I1126 00:25:36.826113 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:36 crc kubenswrapper[4766]: I1126 00:25:36.826298 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:36 crc kubenswrapper[4766]: E1126 00:25:36.826447 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:36 crc kubenswrapper[4766]: E1126 00:25:36.826638 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:36 crc kubenswrapper[4766]: E1126 00:25:36.826757 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:37 crc kubenswrapper[4766]: I1126 00:25:37.826213 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:37 crc kubenswrapper[4766]: E1126 00:25:37.826383 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:38 crc kubenswrapper[4766]: I1126 00:25:38.709632 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-prqrp_234068f7-4097-4e6b-aac3-4bafe425c1f2/kube-multus/1.log" Nov 26 00:25:38 crc kubenswrapper[4766]: I1126 00:25:38.710326 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-prqrp_234068f7-4097-4e6b-aac3-4bafe425c1f2/kube-multus/0.log" Nov 26 00:25:38 crc kubenswrapper[4766]: I1126 00:25:38.710385 4766 generic.go:334] "Generic (PLEG): container finished" podID="234068f7-4097-4e6b-aac3-4bafe425c1f2" containerID="58d961f26d62b958e729b2bd5129790335de3019e968595091099e5e416353d1" exitCode=1 Nov 26 00:25:38 crc kubenswrapper[4766]: I1126 00:25:38.710434 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-prqrp" event={"ID":"234068f7-4097-4e6b-aac3-4bafe425c1f2","Type":"ContainerDied","Data":"58d961f26d62b958e729b2bd5129790335de3019e968595091099e5e416353d1"} Nov 26 00:25:38 crc kubenswrapper[4766]: I1126 00:25:38.710479 4766 scope.go:117] "RemoveContainer" containerID="aa08cb69552885c5253ed0399cee46f0e6a16acc0b6e7425bf111dcc056d5a0c" Nov 26 00:25:38 crc kubenswrapper[4766]: I1126 00:25:38.710938 4766 scope.go:117] "RemoveContainer" containerID="58d961f26d62b958e729b2bd5129790335de3019e968595091099e5e416353d1" Nov 26 00:25:38 crc kubenswrapper[4766]: E1126 00:25:38.711143 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-prqrp_openshift-multus(234068f7-4097-4e6b-aac3-4bafe425c1f2)\"" pod="openshift-multus/multus-prqrp" podUID="234068f7-4097-4e6b-aac3-4bafe425c1f2" Nov 26 00:25:38 crc kubenswrapper[4766]: I1126 00:25:38.826249 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:38 crc kubenswrapper[4766]: I1126 00:25:38.826333 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:38 crc kubenswrapper[4766]: E1126 00:25:38.826375 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:38 crc kubenswrapper[4766]: E1126 00:25:38.826461 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:38 crc kubenswrapper[4766]: I1126 00:25:38.826537 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:38 crc kubenswrapper[4766]: E1126 00:25:38.826610 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:39 crc kubenswrapper[4766]: I1126 00:25:39.717568 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-prqrp_234068f7-4097-4e6b-aac3-4bafe425c1f2/kube-multus/1.log" Nov 26 00:25:39 crc kubenswrapper[4766]: I1126 00:25:39.826623 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:39 crc kubenswrapper[4766]: E1126 00:25:39.828032 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:39 crc kubenswrapper[4766]: E1126 00:25:39.838926 4766 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 26 00:25:39 crc kubenswrapper[4766]: E1126 00:25:39.911725 4766 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 00:25:40 crc kubenswrapper[4766]: I1126 00:25:40.827045 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:40 crc kubenswrapper[4766]: I1126 00:25:40.827122 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:40 crc kubenswrapper[4766]: I1126 00:25:40.827152 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:40 crc kubenswrapper[4766]: E1126 00:25:40.827772 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:40 crc kubenswrapper[4766]: E1126 00:25:40.827932 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:40 crc kubenswrapper[4766]: E1126 00:25:40.828141 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:40 crc kubenswrapper[4766]: I1126 00:25:40.828773 4766 scope.go:117] "RemoveContainer" containerID="3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d" Nov 26 00:25:41 crc kubenswrapper[4766]: I1126 00:25:41.641340 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-5z9ll"] Nov 26 00:25:41 crc kubenswrapper[4766]: I1126 00:25:41.641517 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:41 crc kubenswrapper[4766]: E1126 00:25:41.641680 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:41 crc kubenswrapper[4766]: I1126 00:25:41.726199 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fxnln_415fdacb-85c1-4265-89b7-6771a84ffc89/ovnkube-controller/3.log" Nov 26 00:25:41 crc kubenswrapper[4766]: I1126 00:25:41.729604 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerStarted","Data":"2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1"} Nov 26 00:25:41 crc kubenswrapper[4766]: I1126 00:25:41.730017 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:25:41 crc kubenswrapper[4766]: I1126 00:25:41.766809 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" podStartSLOduration=98.766777928 podStartE2EDuration="1m38.766777928s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:25:41.764357799 +0000 UTC m=+122.613128239" watchObservedRunningTime="2025-11-26 00:25:41.766777928 +0000 UTC m=+122.615548358" Nov 26 00:25:42 crc kubenswrapper[4766]: I1126 00:25:42.826620 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:42 crc kubenswrapper[4766]: I1126 00:25:42.826627 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:42 crc kubenswrapper[4766]: I1126 00:25:42.826630 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:42 crc kubenswrapper[4766]: E1126 00:25:42.827417 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:42 crc kubenswrapper[4766]: E1126 00:25:42.827211 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:42 crc kubenswrapper[4766]: I1126 00:25:42.826714 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:42 crc kubenswrapper[4766]: E1126 00:25:42.827555 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:42 crc kubenswrapper[4766]: E1126 00:25:42.827756 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:44 crc kubenswrapper[4766]: I1126 00:25:44.825753 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:44 crc kubenswrapper[4766]: I1126 00:25:44.825819 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:44 crc kubenswrapper[4766]: I1126 00:25:44.825819 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:44 crc kubenswrapper[4766]: I1126 00:25:44.825785 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:44 crc kubenswrapper[4766]: E1126 00:25:44.825919 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:44 crc kubenswrapper[4766]: E1126 00:25:44.826075 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:44 crc kubenswrapper[4766]: E1126 00:25:44.826169 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:44 crc kubenswrapper[4766]: E1126 00:25:44.826266 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:44 crc kubenswrapper[4766]: E1126 00:25:44.913115 4766 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 00:25:46 crc kubenswrapper[4766]: I1126 00:25:46.826374 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:46 crc kubenswrapper[4766]: E1126 00:25:46.826966 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:46 crc kubenswrapper[4766]: I1126 00:25:46.826561 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:46 crc kubenswrapper[4766]: E1126 00:25:46.827163 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:46 crc kubenswrapper[4766]: I1126 00:25:46.826546 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:46 crc kubenswrapper[4766]: E1126 00:25:46.827338 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:46 crc kubenswrapper[4766]: I1126 00:25:46.826587 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:46 crc kubenswrapper[4766]: E1126 00:25:46.827505 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:48 crc kubenswrapper[4766]: I1126 00:25:48.826469 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:48 crc kubenswrapper[4766]: I1126 00:25:48.826533 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:48 crc kubenswrapper[4766]: I1126 00:25:48.826568 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:48 crc kubenswrapper[4766]: I1126 00:25:48.826462 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:48 crc kubenswrapper[4766]: E1126 00:25:48.826629 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:48 crc kubenswrapper[4766]: E1126 00:25:48.826799 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:48 crc kubenswrapper[4766]: E1126 00:25:48.826975 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:48 crc kubenswrapper[4766]: E1126 00:25:48.827026 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:49 crc kubenswrapper[4766]: I1126 00:25:49.828914 4766 scope.go:117] "RemoveContainer" containerID="58d961f26d62b958e729b2bd5129790335de3019e968595091099e5e416353d1" Nov 26 00:25:49 crc kubenswrapper[4766]: E1126 00:25:49.913538 4766 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 00:25:50 crc kubenswrapper[4766]: I1126 00:25:50.765227 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-prqrp_234068f7-4097-4e6b-aac3-4bafe425c1f2/kube-multus/1.log" Nov 26 00:25:50 crc kubenswrapper[4766]: I1126 00:25:50.765314 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-prqrp" event={"ID":"234068f7-4097-4e6b-aac3-4bafe425c1f2","Type":"ContainerStarted","Data":"c0c13369c6a2c56514a4088ae25d890e67379d75c2709a3f3041b422947fbcbd"} Nov 26 00:25:50 crc kubenswrapper[4766]: I1126 00:25:50.826130 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:50 crc kubenswrapper[4766]: I1126 00:25:50.826222 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:50 crc kubenswrapper[4766]: E1126 00:25:50.826315 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:50 crc kubenswrapper[4766]: I1126 00:25:50.826510 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:50 crc kubenswrapper[4766]: I1126 00:25:50.826538 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:50 crc kubenswrapper[4766]: E1126 00:25:50.826600 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:50 crc kubenswrapper[4766]: E1126 00:25:50.826754 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:50 crc kubenswrapper[4766]: E1126 00:25:50.826908 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:52 crc kubenswrapper[4766]: I1126 00:25:52.826719 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:52 crc kubenswrapper[4766]: I1126 00:25:52.826784 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:52 crc kubenswrapper[4766]: E1126 00:25:52.826956 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:52 crc kubenswrapper[4766]: E1126 00:25:52.827119 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:52 crc kubenswrapper[4766]: I1126 00:25:52.827465 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:52 crc kubenswrapper[4766]: E1126 00:25:52.827626 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:52 crc kubenswrapper[4766]: I1126 00:25:52.828793 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:52 crc kubenswrapper[4766]: E1126 00:25:52.829058 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:54 crc kubenswrapper[4766]: I1126 00:25:54.826222 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:54 crc kubenswrapper[4766]: I1126 00:25:54.826274 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:54 crc kubenswrapper[4766]: I1126 00:25:54.826323 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:54 crc kubenswrapper[4766]: I1126 00:25:54.826230 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:54 crc kubenswrapper[4766]: E1126 00:25:54.826441 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 00:25:54 crc kubenswrapper[4766]: E1126 00:25:54.826355 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 00:25:54 crc kubenswrapper[4766]: E1126 00:25:54.826582 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 00:25:54 crc kubenswrapper[4766]: E1126 00:25:54.826642 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-5z9ll" podUID="b6600efa-4858-41aa-a890-44ce262b63dd" Nov 26 00:25:56 crc kubenswrapper[4766]: I1126 00:25:56.825850 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:25:56 crc kubenswrapper[4766]: I1126 00:25:56.825910 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:25:56 crc kubenswrapper[4766]: I1126 00:25:56.825910 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:25:56 crc kubenswrapper[4766]: I1126 00:25:56.826003 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:25:56 crc kubenswrapper[4766]: I1126 00:25:56.828691 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 26 00:25:56 crc kubenswrapper[4766]: I1126 00:25:56.829216 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 26 00:25:56 crc kubenswrapper[4766]: I1126 00:25:56.829603 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 26 00:25:56 crc kubenswrapper[4766]: I1126 00:25:56.829610 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 26 00:25:56 crc kubenswrapper[4766]: I1126 00:25:56.830746 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 26 00:25:56 crc kubenswrapper[4766]: I1126 00:25:56.831874 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.010343 4766 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.055705 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.056155 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.059710 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.060041 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.065071 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.065506 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.067714 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-x2vlb"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.068157 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.072079 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-x2vlb" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.079179 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.079179 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.085202 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.085768 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.086235 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.086358 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.086491 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.086797 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.086881 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.087071 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.087121 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.087320 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.087424 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.087810 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.086803 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.093157 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.093576 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.093894 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.096593 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/11353a4e-daac-46a8-97de-160ca1272ea1-audit-policies\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.096642 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/11353a4e-daac-46a8-97de-160ca1272ea1-encryption-config\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.096688 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-serving-cert\") pod \"route-controller-manager-6576b87f9c-dsmkn\" (UID: \"2e6903aa-3f06-44f0-807c-4098ba5fa6d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.096719 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/11353a4e-daac-46a8-97de-160ca1272ea1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.096748 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqtzl\" (UniqueName: \"kubernetes.io/projected/11353a4e-daac-46a8-97de-160ca1272ea1-kube-api-access-lqtzl\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.096769 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/11353a4e-daac-46a8-97de-160ca1272ea1-audit-dir\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.096808 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11353a4e-daac-46a8-97de-160ca1272ea1-serving-cert\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.096827 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-client-ca\") pod \"route-controller-manager-6576b87f9c-dsmkn\" (UID: \"2e6903aa-3f06-44f0-807c-4098ba5fa6d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.096850 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-config\") pod \"route-controller-manager-6576b87f9c-dsmkn\" (UID: \"2e6903aa-3f06-44f0-807c-4098ba5fa6d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.096869 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/11353a4e-daac-46a8-97de-160ca1272ea1-etcd-client\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.096891 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/11353a4e-daac-46a8-97de-160ca1272ea1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.096917 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwt9w\" (UniqueName: \"kubernetes.io/projected/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-kube-api-access-wwt9w\") pod \"route-controller-manager-6576b87f9c-dsmkn\" (UID: \"2e6903aa-3f06-44f0-807c-4098ba5fa6d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.098955 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-blhhp"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.099318 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-blhhp" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.102235 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.102744 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.103283 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.103489 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.103941 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.105970 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-bqmvb"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.106483 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zndxf"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.106894 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zndxf" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.107268 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bqmvb" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.107451 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-hqj86"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.108005 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.109943 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.112288 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8sgwh"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.112708 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8zjx6"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.113034 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-59zt8"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.113345 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-59zt8" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.113854 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.114290 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8sgwh" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.120232 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.121240 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.121392 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.121549 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.121674 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.121777 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.121875 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.122486 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-tgrk8"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.122913 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-bz6r8"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.123194 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-pruner-29401920-82g4w"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.123538 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29401920-82g4w" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.123920 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.124266 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-bz6r8" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.129617 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4vmx"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.130171 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4vmx" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.131693 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-cbzmc"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.132116 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-cbzmc" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.132709 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-m8qgq"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.133082 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.134277 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.134720 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bnkxg"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.135215 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-z22jf"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.135882 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8whf2"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.136212 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bnkxg" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.136408 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.136970 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-z22jf" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.141965 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-x2vlb"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.142517 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.142609 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.142805 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.142808 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.142919 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.147236 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mjvdt"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.150586 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.152006 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.152247 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.152401 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.154220 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zndxf"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.156894 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.157120 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.157156 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.157687 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.157828 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.158071 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.158205 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.158232 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.158275 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rj29g"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.158545 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.158777 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.159087 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.159321 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rj29g" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.159862 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.159975 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.160083 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.160216 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.160221 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.160350 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.160644 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.160842 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.160965 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"serviceca" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.161014 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.161121 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.161195 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.161555 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.161590 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.161754 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.161804 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.160969 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.162352 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.162368 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.162485 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.176385 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.176633 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.179820 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.179946 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.179967 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.180077 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.180106 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.180183 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.181362 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xbr6t"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.182009 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7z6k"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.182319 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.182374 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.182494 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.182580 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.182686 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.182981 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7z6k" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.182045 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xbr6t" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.183268 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5thz5"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.183614 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5thz5" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.185224 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.185463 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.185478 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.185545 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.186921 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-mq7cp"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.187456 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mq7cp" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.185804 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.189527 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.190604 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.194523 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.198415 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/11353a4e-daac-46a8-97de-160ca1272ea1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.198467 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwt9w\" (UniqueName: \"kubernetes.io/projected/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-kube-api-access-wwt9w\") pod \"route-controller-manager-6576b87f9c-dsmkn\" (UID: \"2e6903aa-3f06-44f0-807c-4098ba5fa6d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.198511 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/11353a4e-daac-46a8-97de-160ca1272ea1-audit-policies\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.198552 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/11353a4e-daac-46a8-97de-160ca1272ea1-encryption-config\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.198577 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-serving-cert\") pod \"route-controller-manager-6576b87f9c-dsmkn\" (UID: \"2e6903aa-3f06-44f0-807c-4098ba5fa6d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.198607 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/11353a4e-daac-46a8-97de-160ca1272ea1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.198644 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqtzl\" (UniqueName: \"kubernetes.io/projected/11353a4e-daac-46a8-97de-160ca1272ea1-kube-api-access-lqtzl\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.198687 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/11353a4e-daac-46a8-97de-160ca1272ea1-audit-dir\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.198718 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11353a4e-daac-46a8-97de-160ca1272ea1-serving-cert\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.198742 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-config\") pod \"route-controller-manager-6576b87f9c-dsmkn\" (UID: \"2e6903aa-3f06-44f0-807c-4098ba5fa6d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.198769 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-client-ca\") pod \"route-controller-manager-6576b87f9c-dsmkn\" (UID: \"2e6903aa-3f06-44f0-807c-4098ba5fa6d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.198794 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/11353a4e-daac-46a8-97de-160ca1272ea1-etcd-client\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.199389 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.203679 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tnhkb"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.204091 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7kxvg"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.204430 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-dt2nb"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.206380 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.206900 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"pruner-dockercfg-p7bcw" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.210557 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.211092 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.211290 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.211502 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/11353a4e-daac-46a8-97de-160ca1272ea1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.211918 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/11353a4e-daac-46a8-97de-160ca1272ea1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.212483 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/11353a4e-daac-46a8-97de-160ca1272ea1-audit-policies\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.212913 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-config\") pod \"route-controller-manager-6576b87f9c-dsmkn\" (UID: \"2e6903aa-3f06-44f0-807c-4098ba5fa6d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.213292 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-p5mt6"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.213612 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-client-ca\") pod \"route-controller-manager-6576b87f9c-dsmkn\" (UID: \"2e6903aa-3f06-44f0-807c-4098ba5fa6d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.213800 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p5mt6" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.213946 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/11353a4e-daac-46a8-97de-160ca1272ea1-audit-dir\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.214019 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tnhkb" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.214132 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7kxvg" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.214255 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dt2nb" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.216096 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/11353a4e-daac-46a8-97de-160ca1272ea1-encryption-config\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.217387 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.218291 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-45dcj"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.233809 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-45dcj" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.234575 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-serving-cert\") pod \"route-controller-manager-6576b87f9c-dsmkn\" (UID: \"2e6903aa-3f06-44f0-807c-4098ba5fa6d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.234958 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.235090 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/11353a4e-daac-46a8-97de-160ca1272ea1-etcd-client\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.235136 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11353a4e-daac-46a8-97de-160ca1272ea1-serving-cert\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.236638 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.236822 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.238041 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.238047 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.242837 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2vsl7"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.244156 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.245848 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.246144 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-z5kzz"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.249371 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-z5kzz" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.253287 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vvlgv"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.254473 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vvlgv" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.255959 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pm5j5"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.256510 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pm5j5" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.258016 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.259059 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4qsp7"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.259751 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qsp7" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.260831 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-w62xw"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.261236 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-w62xw" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.262398 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.262866 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.264025 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rf9kz"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.264466 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rf9kz" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.265965 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8sgwh"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.268428 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-bz6r8"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.270006 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-59zt8"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.271563 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29401920-82g4w"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.273104 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4vmx"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.274771 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8zjx6"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.275877 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.277284 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-m8qgq"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.278231 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-hqj86"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.280097 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mjvdt"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.280122 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.281038 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bnkxg"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.282495 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5thz5"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.286377 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rj29g"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.286428 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-z22jf"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.286443 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-cbzmc"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.291042 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7z6k"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.291076 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-tgrk8"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.294831 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xbr6t"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.294892 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8whf2"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.295938 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-w62xw"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.297010 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-8kdj9"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.297796 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-8kdj9" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.298073 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.299422 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.300894 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-blhhp"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.303262 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-zbsq2"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.304018 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-zbsq2" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.305173 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-z5kzz"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.306917 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tnhkb"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.308994 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-mq7cp"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.311128 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pm5j5"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.312495 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-g4bjk"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.313613 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.314798 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-p5mt6"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.316910 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.318100 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2vsl7"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.318625 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.319793 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-dt2nb"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.320975 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7kxvg"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.322304 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4qsp7"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.324129 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vvlgv"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.328195 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-g4bjk"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.330015 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-8kdj9"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.331452 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rf9kz"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.332727 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-k2shv"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.333532 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-k2shv" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.334241 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-k2shv"] Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.338166 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.358183 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.377816 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.398733 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.418264 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.442253 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.458574 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.478740 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.498370 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.518754 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.538069 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.558564 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.578452 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.598140 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.618423 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.638542 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.680873 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.697982 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.719018 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.738468 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.757638 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.778813 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.797787 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.818234 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.838356 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.858578 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.879304 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.897848 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.919964 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.939003 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.959353 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 26 00:25:58 crc kubenswrapper[4766]: I1126 00:25:58.979844 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.045726 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwt9w\" (UniqueName: \"kubernetes.io/projected/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-kube-api-access-wwt9w\") pod \"route-controller-manager-6576b87f9c-dsmkn\" (UID: \"2e6903aa-3f06-44f0-807c-4098ba5fa6d0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.059592 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqtzl\" (UniqueName: \"kubernetes.io/projected/11353a4e-daac-46a8-97de-160ca1272ea1-kube-api-access-lqtzl\") pod \"apiserver-7bbb656c7d-f4gkn\" (UID: \"11353a4e-daac-46a8-97de-160ca1272ea1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.060018 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.079238 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.100079 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.118939 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.147639 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.158744 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.178779 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.198814 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.217179 4766 request.go:700] Waited for 1.002712925s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.219224 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.239234 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.258336 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.279085 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.299164 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.306852 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.319456 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.321792 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.339425 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.361466 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.379076 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.398907 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.419470 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.439221 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.459051 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.479181 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.498698 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.535889 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn"] Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.536136 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.538232 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.549322 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn"] Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.558870 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.579180 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.598111 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.618036 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.638580 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.658973 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.679716 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.698991 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.718621 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.738348 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.757902 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.777868 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.797767 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" event={"ID":"11353a4e-daac-46a8-97de-160ca1272ea1","Type":"ContainerDied","Data":"102ddfebe6e31d2610df013694b70ce61a1dbfd1c453082badd912b9c34d94af"} Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.797955 4766 generic.go:334] "Generic (PLEG): container finished" podID="11353a4e-daac-46a8-97de-160ca1272ea1" containerID="102ddfebe6e31d2610df013694b70ce61a1dbfd1c453082badd912b9c34d94af" exitCode=0 Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.798039 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.798201 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" event={"ID":"11353a4e-daac-46a8-97de-160ca1272ea1","Type":"ContainerStarted","Data":"8539bf25e243dec395587feb8e37404c60772fe2b08d0c0e9d95c40fe6ceeef3"} Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.800670 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" event={"ID":"2e6903aa-3f06-44f0-807c-4098ba5fa6d0","Type":"ContainerStarted","Data":"0075cab8c7badd77fdfb6505d91198e3e29dcd7ba3fd761756ac9efca6dd1ba1"} Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.800713 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" event={"ID":"2e6903aa-3f06-44f0-807c-4098ba5fa6d0","Type":"ContainerStarted","Data":"ad4d720b99f936bd05c9b42318d94b4dbf0397839577c56bdfd3994cbb69713b"} Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.800987 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.804872 4766 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-dsmkn container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.804991 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" podUID="2e6903aa-3f06-44f0-807c-4098ba5fa6d0" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.818765 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.838450 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.859409 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.878403 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.893231 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.899597 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.918447 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.938881 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.958320 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.978106 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 26 00:25:59 crc kubenswrapper[4766]: I1126 00:25:59.997908 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.018840 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.039247 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.058156 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.079064 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.099401 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.118749 4766 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.137848 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.158050 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.178314 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.198046 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.218786 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.236504 4766 request.go:700] Waited for 1.57689201s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/persistentvolumes/pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.320334 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1e392e29-fdb8-450d-ac24-ba91e802139a-registry-tls\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.320436 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-console-oauth-config\") pod \"console-f9d7485db-hqj86\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.320484 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e00b34f6-3577-4658-b819-43ada470fbe7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-bnkxg\" (UID: \"e00b34f6-3577-4658-b819-43ada470fbe7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bnkxg" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.320528 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/734b50e7-a98e-4999-bab8-ae2547f08c0c-serving-cert\") pod \"console-operator-58897d9998-cbzmc\" (UID: \"734b50e7-a98e-4999-bab8-ae2547f08c0c\") " pod="openshift-console-operator/console-operator-58897d9998-cbzmc" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.320573 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28b3189b-79c5-4da2-9ce5-d067f66a47c2-config\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.320597 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9c103959-be49-4b6a-9689-0a58451bc2b7-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8sgwh\" (UID: \"9c103959-be49-4b6a-9689-0a58451bc2b7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8sgwh" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.320640 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gml6x\" (UniqueName: \"kubernetes.io/projected/284847a7-94c0-4f03-a54d-6d22fe77353f-kube-api-access-gml6x\") pod \"image-pruner-29401920-82g4w\" (UID: \"284847a7-94c0-4f03-a54d-6d22fe77353f\") " pod="openshift-image-registry/image-pruner-29401920-82g4w" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.320694 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1e392e29-fdb8-450d-ac24-ba91e802139a-trusted-ca\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.320743 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b3321717-6860-48e4-91a4-99245add293d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-n4vmx\" (UID: \"b3321717-6860-48e4-91a4-99245add293d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4vmx" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.320767 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/28b3189b-79c5-4da2-9ce5-d067f66a47c2-encryption-config\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.320809 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-oauth-serving-cert\") pod \"console-f9d7485db-hqj86\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.320834 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9aaa727c-565e-4819-801d-9f1787118d20-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-mjvdt\" (UID: \"9aaa727c-565e-4819-801d-9f1787118d20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.320896 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28b3189b-79c5-4da2-9ce5-d067f66a47c2-serving-cert\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.320923 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9aaa727c-565e-4819-801d-9f1787118d20-client-ca\") pod \"controller-manager-879f6c89f-mjvdt\" (UID: \"9aaa727c-565e-4819-801d-9f1787118d20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.320966 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/20bb15a2-ef87-40cf-b450-f2a50a290325-metrics-tls\") pod \"dns-operator-744455d44c-z22jf\" (UID: \"20bb15a2-ef87-40cf-b450-f2a50a290325\") " pod="openshift-dns-operator/dns-operator-744455d44c-z22jf" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.320992 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/cc0815f7-09db-4adf-b525-205c8fd8e64e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-59zt8\" (UID: \"cc0815f7-09db-4adf-b525-205c8fd8e64e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-59zt8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.321056 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b3321717-6860-48e4-91a4-99245add293d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-n4vmx\" (UID: \"b3321717-6860-48e4-91a4-99245add293d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4vmx" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.321078 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/28b3189b-79c5-4da2-9ce5-d067f66a47c2-audit\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.321126 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4ngl\" (UniqueName: \"kubernetes.io/projected/cb710ab4-0d33-4e91-b749-68169fc723c2-kube-api-access-c4ngl\") pod \"machine-approver-56656f9798-bqmvb\" (UID: \"cb710ab4-0d33-4e91-b749-68169fc723c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bqmvb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.321148 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-console-serving-cert\") pod \"console-f9d7485db-hqj86\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.321206 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64bb8da2-3238-49d0-9cb2-b582701fb35d-config\") pod \"etcd-operator-b45778765-8whf2\" (UID: \"64bb8da2-3238-49d0-9cb2-b582701fb35d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.321228 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/28b3189b-79c5-4da2-9ce5-d067f66a47c2-node-pullsecrets\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.321272 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.321296 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/28b3189b-79c5-4da2-9ce5-d067f66a47c2-etcd-serving-ca\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.321351 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cb710ab4-0d33-4e91-b749-68169fc723c2-auth-proxy-config\") pod \"machine-approver-56656f9798-bqmvb\" (UID: \"cb710ab4-0d33-4e91-b749-68169fc723c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bqmvb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.321393 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8260737-caa6-42f1-8854-ad2b97a1ee42-config\") pod \"machine-api-operator-5694c8668f-x2vlb\" (UID: \"c8260737-caa6-42f1-8854-ad2b97a1ee42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x2vlb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.321451 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/64bb8da2-3238-49d0-9cb2-b582701fb35d-etcd-service-ca\") pod \"etcd-operator-b45778765-8whf2\" (UID: \"64bb8da2-3238-49d0-9cb2-b582701fb35d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.321474 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e00b34f6-3577-4658-b819-43ada470fbe7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-bnkxg\" (UID: \"e00b34f6-3577-4658-b819-43ada470fbe7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bnkxg" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.321518 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgsrx\" (UniqueName: \"kubernetes.io/projected/b3321717-6860-48e4-91a4-99245add293d-kube-api-access-tgsrx\") pod \"cluster-image-registry-operator-dc59b4c8b-n4vmx\" (UID: \"b3321717-6860-48e4-91a4-99245add293d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4vmx" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.321562 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk7cf\" (UniqueName: \"kubernetes.io/projected/9c103959-be49-4b6a-9689-0a58451bc2b7-kube-api-access-qk7cf\") pod \"cluster-samples-operator-665b6dd947-8sgwh\" (UID: \"9c103959-be49-4b6a-9689-0a58451bc2b7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8sgwh" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.321613 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcv6t\" (UniqueName: \"kubernetes.io/projected/6f7e21ec-878c-4f90-9f5f-51e95d71e373-kube-api-access-jcv6t\") pod \"openshift-apiserver-operator-796bbdcf4f-zndxf\" (UID: \"6f7e21ec-878c-4f90-9f5f-51e95d71e373\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zndxf" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.321724 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-trusted-ca-bundle\") pod \"console-f9d7485db-hqj86\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.321798 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/64bb8da2-3238-49d0-9cb2-b582701fb35d-etcd-ca\") pod \"etcd-operator-b45778765-8whf2\" (UID: \"64bb8da2-3238-49d0-9cb2-b582701fb35d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.321843 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/28b3189b-79c5-4da2-9ce5-d067f66a47c2-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.321887 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js6sh\" (UniqueName: \"kubernetes.io/projected/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-kube-api-access-js6sh\") pod \"console-f9d7485db-hqj86\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.321932 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45bb08cb-d545-4568-9a64-780628c87786-config\") pod \"authentication-operator-69f744f599-blhhp\" (UID: \"45bb08cb-d545-4568-9a64-780628c87786\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-blhhp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.322027 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1e392e29-fdb8-450d-ac24-ba91e802139a-bound-sa-token\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.322076 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/64bb8da2-3238-49d0-9cb2-b582701fb35d-etcd-client\") pod \"etcd-operator-b45778765-8whf2\" (UID: \"64bb8da2-3238-49d0-9cb2-b582701fb35d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.322126 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/45bb08cb-d545-4568-9a64-780628c87786-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-blhhp\" (UID: \"45bb08cb-d545-4568-9a64-780628c87786\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-blhhp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.322173 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk9v8\" (UniqueName: \"kubernetes.io/projected/c8260737-caa6-42f1-8854-ad2b97a1ee42-kube-api-access-mk9v8\") pod \"machine-api-operator-5694c8668f-x2vlb\" (UID: \"c8260737-caa6-42f1-8854-ad2b97a1ee42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x2vlb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.322216 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9aaa727c-565e-4819-801d-9f1787118d20-serving-cert\") pod \"controller-manager-879f6c89f-mjvdt\" (UID: \"9aaa727c-565e-4819-801d-9f1787118d20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.322259 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/cb710ab4-0d33-4e91-b749-68169fc723c2-machine-approver-tls\") pod \"machine-approver-56656f9798-bqmvb\" (UID: \"cb710ab4-0d33-4e91-b749-68169fc723c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bqmvb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.322333 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-audit-policies\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.322374 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c91e0334-6829-4a8f-9239-3930e84e65c4-audit-dir\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.322418 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.322466 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zflng\" (UniqueName: \"kubernetes.io/projected/cc0815f7-09db-4adf-b525-205c8fd8e64e-kube-api-access-zflng\") pod \"openshift-config-operator-7777fb866f-59zt8\" (UID: \"cc0815f7-09db-4adf-b525-205c8fd8e64e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-59zt8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.322517 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.322601 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67wjp\" (UniqueName: \"kubernetes.io/projected/d94c7fca-2c24-4005-9839-0578e1b39283-kube-api-access-67wjp\") pod \"downloads-7954f5f757-bz6r8\" (UID: \"d94c7fca-2c24-4005-9839-0578e1b39283\") " pod="openshift-console/downloads-7954f5f757-bz6r8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.322678 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9z2xg\" (UniqueName: \"kubernetes.io/projected/45bb08cb-d545-4568-9a64-780628c87786-kube-api-access-9z2xg\") pod \"authentication-operator-69f744f599-blhhp\" (UID: \"45bb08cb-d545-4568-9a64-780628c87786\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-blhhp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.322722 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c8260737-caa6-42f1-8854-ad2b97a1ee42-images\") pod \"machine-api-operator-5694c8668f-x2vlb\" (UID: \"c8260737-caa6-42f1-8854-ad2b97a1ee42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x2vlb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.322770 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.322819 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.322865 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbztm\" (UniqueName: \"kubernetes.io/projected/c91e0334-6829-4a8f-9239-3930e84e65c4-kube-api-access-sbztm\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.322961 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.323036 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.323082 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.323117 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.323169 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45bb08cb-d545-4568-9a64-780628c87786-serving-cert\") pod \"authentication-operator-69f744f599-blhhp\" (UID: \"45bb08cb-d545-4568-9a64-780628c87786\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-blhhp" Nov 26 00:26:00 crc kubenswrapper[4766]: E1126 00:26:00.324239 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:00.824223502 +0000 UTC m=+141.672993932 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.329763 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1e392e29-fdb8-450d-ac24-ba91e802139a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.329809 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6j56\" (UniqueName: \"kubernetes.io/projected/9aaa727c-565e-4819-801d-9f1787118d20-kube-api-access-r6j56\") pod \"controller-manager-879f6c89f-mjvdt\" (UID: \"9aaa727c-565e-4819-801d-9f1787118d20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.329826 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/734b50e7-a98e-4999-bab8-ae2547f08c0c-trusted-ca\") pod \"console-operator-58897d9998-cbzmc\" (UID: \"734b50e7-a98e-4999-bab8-ae2547f08c0c\") " pod="openshift-console-operator/console-operator-58897d9998-cbzmc" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.329849 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/28b3189b-79c5-4da2-9ce5-d067f66a47c2-image-import-ca\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.329885 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kb9mk\" (UniqueName: \"kubernetes.io/projected/20bb15a2-ef87-40cf-b450-f2a50a290325-kube-api-access-kb9mk\") pod \"dns-operator-744455d44c-z22jf\" (UID: \"20bb15a2-ef87-40cf-b450-f2a50a290325\") " pod="openshift-dns-operator/dns-operator-744455d44c-z22jf" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.329912 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/45bb08cb-d545-4568-9a64-780628c87786-service-ca-bundle\") pod \"authentication-operator-69f744f599-blhhp\" (UID: \"45bb08cb-d545-4568-9a64-780628c87786\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-blhhp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.329938 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs6k7\" (UniqueName: \"kubernetes.io/projected/e00b34f6-3577-4658-b819-43ada470fbe7-kube-api-access-fs6k7\") pod \"openshift-controller-manager-operator-756b6f6bc6-bnkxg\" (UID: \"e00b34f6-3577-4658-b819-43ada470fbe7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bnkxg" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.329961 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/28b3189b-79c5-4da2-9ce5-d067f66a47c2-audit-dir\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330039 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/00869536-c66a-4fae-91e7-0c6d271ab8cd-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rj29g\" (UID: \"00869536-c66a-4fae-91e7-0c6d271ab8cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rj29g" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330072 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b3321717-6860-48e4-91a4-99245add293d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-n4vmx\" (UID: \"b3321717-6860-48e4-91a4-99245add293d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4vmx" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330099 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/28b3189b-79c5-4da2-9ce5-d067f66a47c2-etcd-client\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330145 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f7e21ec-878c-4f90-9f5f-51e95d71e373-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zndxf\" (UID: \"6f7e21ec-878c-4f90-9f5f-51e95d71e373\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zndxf" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330171 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-service-ca\") pod \"console-f9d7485db-hqj86\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330221 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00869536-c66a-4fae-91e7-0c6d271ab8cd-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rj29g\" (UID: \"00869536-c66a-4fae-91e7-0c6d271ab8cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rj29g" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330245 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330273 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64bb8da2-3238-49d0-9cb2-b582701fb35d-serving-cert\") pod \"etcd-operator-b45778765-8whf2\" (UID: \"64bb8da2-3238-49d0-9cb2-b582701fb35d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330298 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwjt5\" (UniqueName: \"kubernetes.io/projected/64bb8da2-3238-49d0-9cb2-b582701fb35d-kube-api-access-rwjt5\") pod \"etcd-operator-b45778765-8whf2\" (UID: \"64bb8da2-3238-49d0-9cb2-b582701fb35d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330321 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc0815f7-09db-4adf-b525-205c8fd8e64e-serving-cert\") pod \"openshift-config-operator-7777fb866f-59zt8\" (UID: \"cc0815f7-09db-4adf-b525-205c8fd8e64e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-59zt8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330346 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c8260737-caa6-42f1-8854-ad2b97a1ee42-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-x2vlb\" (UID: \"c8260737-caa6-42f1-8854-ad2b97a1ee42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x2vlb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330375 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/284847a7-94c0-4f03-a54d-6d22fe77353f-serviceca\") pod \"image-pruner-29401920-82g4w\" (UID: \"284847a7-94c0-4f03-a54d-6d22fe77353f\") " pod="openshift-image-registry/image-pruner-29401920-82g4w" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330400 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-console-config\") pod \"console-f9d7485db-hqj86\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330438 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvt7w\" (UniqueName: \"kubernetes.io/projected/1e392e29-fdb8-450d-ac24-ba91e802139a-kube-api-access-mvt7w\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330462 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330488 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aaa727c-565e-4819-801d-9f1787118d20-config\") pod \"controller-manager-879f6c89f-mjvdt\" (UID: \"9aaa727c-565e-4819-801d-9f1787118d20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330512 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/734b50e7-a98e-4999-bab8-ae2547f08c0c-config\") pod \"console-operator-58897d9998-cbzmc\" (UID: \"734b50e7-a98e-4999-bab8-ae2547f08c0c\") " pod="openshift-console-operator/console-operator-58897d9998-cbzmc" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330562 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1e392e29-fdb8-450d-ac24-ba91e802139a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330587 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330609 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjp9n\" (UniqueName: \"kubernetes.io/projected/28b3189b-79c5-4da2-9ce5-d067f66a47c2-kube-api-access-pjp9n\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330632 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb710ab4-0d33-4e91-b749-68169fc723c2-config\") pod \"machine-approver-56656f9798-bqmvb\" (UID: \"cb710ab4-0d33-4e91-b749-68169fc723c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bqmvb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330676 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1e392e29-fdb8-450d-ac24-ba91e802139a-registry-certificates\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330699 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00869536-c66a-4fae-91e7-0c6d271ab8cd-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rj29g\" (UID: \"00869536-c66a-4fae-91e7-0c6d271ab8cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rj29g" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330821 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx6tx\" (UniqueName: \"kubernetes.io/projected/734b50e7-a98e-4999-bab8-ae2547f08c0c-kube-api-access-wx6tx\") pod \"console-operator-58897d9998-cbzmc\" (UID: \"734b50e7-a98e-4999-bab8-ae2547f08c0c\") " pod="openshift-console-operator/console-operator-58897d9998-cbzmc" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.330857 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f7e21ec-878c-4f90-9f5f-51e95d71e373-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zndxf\" (UID: \"6f7e21ec-878c-4f90-9f5f-51e95d71e373\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zndxf" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.431350 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:00 crc kubenswrapper[4766]: E1126 00:26:00.431591 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:00.931544044 +0000 UTC m=+141.780314474 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.431732 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.431802 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwtf5\" (UniqueName: \"kubernetes.io/projected/abf46edd-34b9-42e7-94ca-a3066047cb53-kube-api-access-bwtf5\") pod \"csi-hostpathplugin-g4bjk\" (UID: \"abf46edd-34b9-42e7-94ca-a3066047cb53\") " pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.431847 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.431883 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1e392e29-fdb8-450d-ac24-ba91e802139a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.431910 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6j56\" (UniqueName: \"kubernetes.io/projected/9aaa727c-565e-4819-801d-9f1787118d20-kube-api-access-r6j56\") pod \"controller-manager-879f6c89f-mjvdt\" (UID: \"9aaa727c-565e-4819-801d-9f1787118d20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.431940 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/734b50e7-a98e-4999-bab8-ae2547f08c0c-trusted-ca\") pod \"console-operator-58897d9998-cbzmc\" (UID: \"734b50e7-a98e-4999-bab8-ae2547f08c0c\") " pod="openshift-console-operator/console-operator-58897d9998-cbzmc" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.431974 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvqdb\" (UniqueName: \"kubernetes.io/projected/28602273-98a9-4649-80f3-4e8848b16ee3-kube-api-access-kvqdb\") pod \"package-server-manager-789f6589d5-vvlgv\" (UID: \"28602273-98a9-4649-80f3-4e8848b16ee3\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vvlgv" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432000 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7196e744-19f7-4295-9d97-20ddfe1c44b8-signing-key\") pod \"service-ca-9c57cc56f-w62xw\" (UID: \"7196e744-19f7-4295-9d97-20ddfe1c44b8\") " pod="openshift-service-ca/service-ca-9c57cc56f-w62xw" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432030 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/28b3189b-79c5-4da2-9ce5-d067f66a47c2-image-import-ca\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432062 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs6k7\" (UniqueName: \"kubernetes.io/projected/e00b34f6-3577-4658-b819-43ada470fbe7-kube-api-access-fs6k7\") pod \"openshift-controller-manager-operator-756b6f6bc6-bnkxg\" (UID: \"e00b34f6-3577-4658-b819-43ada470fbe7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bnkxg" Nov 26 00:26:00 crc kubenswrapper[4766]: E1126 00:26:00.432173 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:00.932154199 +0000 UTC m=+141.780924869 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432207 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/28b3189b-79c5-4da2-9ce5-d067f66a47c2-etcd-client\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432234 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/00869536-c66a-4fae-91e7-0c6d271ab8cd-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rj29g\" (UID: \"00869536-c66a-4fae-91e7-0c6d271ab8cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rj29g" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432254 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b3321717-6860-48e4-91a4-99245add293d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-n4vmx\" (UID: \"b3321717-6860-48e4-91a4-99245add293d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4vmx" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432276 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432308 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/abf46edd-34b9-42e7-94ca-a3066047cb53-plugins-dir\") pod \"csi-hostpathplugin-g4bjk\" (UID: \"abf46edd-34b9-42e7-94ca-a3066047cb53\") " pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432329 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4xmb\" (UniqueName: \"kubernetes.io/projected/4e7b24da-83fa-446d-b2a9-97bf382ff2c5-kube-api-access-v4xmb\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7z6k\" (UID: \"4e7b24da-83fa-446d-b2a9-97bf382ff2c5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7z6k" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432349 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6714d5a2-a6be-4841-8268-66de19d238ec-profile-collector-cert\") pod \"olm-operator-6b444d44fb-tnhkb\" (UID: \"6714d5a2-a6be-4841-8268-66de19d238ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tnhkb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432374 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwjt5\" (UniqueName: \"kubernetes.io/projected/64bb8da2-3238-49d0-9cb2-b582701fb35d-kube-api-access-rwjt5\") pod \"etcd-operator-b45778765-8whf2\" (UID: \"64bb8da2-3238-49d0-9cb2-b582701fb35d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432470 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fb2e8401-8c07-4083-8b0f-821cd21b5858-webhook-cert\") pod \"packageserver-d55dfcdfc-bs5t5\" (UID: \"fb2e8401-8c07-4083-8b0f-821cd21b5858\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432446 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1e392e29-fdb8-450d-ac24-ba91e802139a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432617 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc2ff29c-9c87-44f9-95f9-7043db18ffe3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5thz5\" (UID: \"bc2ff29c-9c87-44f9-95f9-7043db18ffe3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5thz5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432670 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dea338f1-f870-484b-ac43-7da254eac47a-config-volume\") pod \"collect-profiles-29401935-hhtv9\" (UID: \"dea338f1-f870-484b-ac43-7da254eac47a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432707 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvt7w\" (UniqueName: \"kubernetes.io/projected/1e392e29-fdb8-450d-ac24-ba91e802139a-kube-api-access-mvt7w\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432732 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432758 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62hpk\" (UniqueName: \"kubernetes.io/projected/752819e1-d0ad-4820-979a-bf6ebfc59b4d-kube-api-access-62hpk\") pod \"control-plane-machine-set-operator-78cbb6b69f-pm5j5\" (UID: \"752819e1-d0ad-4820-979a-bf6ebfc59b4d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pm5j5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432781 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/02a5e11b-33ad-4c1f-8073-2558ba723cd3-trusted-ca\") pod \"ingress-operator-5b745b69d9-7kxvg\" (UID: \"02a5e11b-33ad-4c1f-8073-2558ba723cd3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7kxvg" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432807 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1e392e29-fdb8-450d-ac24-ba91e802139a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432832 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb710ab4-0d33-4e91-b749-68169fc723c2-config\") pod \"machine-approver-56656f9798-bqmvb\" (UID: \"cb710ab4-0d33-4e91-b749-68169fc723c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bqmvb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432856 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00869536-c66a-4fae-91e7-0c6d271ab8cd-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rj29g\" (UID: \"00869536-c66a-4fae-91e7-0c6d271ab8cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rj29g" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432879 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dea338f1-f870-484b-ac43-7da254eac47a-secret-volume\") pod \"collect-profiles-29401935-hhtv9\" (UID: \"dea338f1-f870-484b-ac43-7da254eac47a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432908 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc2ff29c-9c87-44f9-95f9-7043db18ffe3-config\") pod \"kube-apiserver-operator-766d6c64bb-5thz5\" (UID: \"bc2ff29c-9c87-44f9-95f9-7043db18ffe3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5thz5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432931 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5jmt\" (UniqueName: \"kubernetes.io/projected/40ba7989-d314-4273-b2e1-aaf05d4b50a4-kube-api-access-p5jmt\") pod \"multus-admission-controller-857f4d67dd-z5kzz\" (UID: \"40ba7989-d314-4273-b2e1-aaf05d4b50a4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z5kzz" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432958 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1e392e29-fdb8-450d-ac24-ba91e802139a-registry-tls\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.432984 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e00b34f6-3577-4658-b819-43ada470fbe7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-bnkxg\" (UID: \"e00b34f6-3577-4658-b819-43ada470fbe7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bnkxg" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.433125 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c71045e6-eb17-417d-b9fe-12d3a21f4a56-serving-cert\") pod \"service-ca-operator-777779d784-4qsp7\" (UID: \"c71045e6-eb17-417d-b9fe-12d3a21f4a56\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qsp7" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.433171 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/abf46edd-34b9-42e7-94ca-a3066047cb53-socket-dir\") pod \"csi-hostpathplugin-g4bjk\" (UID: \"abf46edd-34b9-42e7-94ca-a3066047cb53\") " pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.433590 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb710ab4-0d33-4e91-b749-68169fc723c2-config\") pod \"machine-approver-56656f9798-bqmvb\" (UID: \"cb710ab4-0d33-4e91-b749-68169fc723c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bqmvb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.433663 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1e392e29-fdb8-450d-ac24-ba91e802139a-trusted-ca\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.433696 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9c103959-be49-4b6a-9689-0a58451bc2b7-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8sgwh\" (UID: \"9c103959-be49-4b6a-9689-0a58451bc2b7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8sgwh" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.433713 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gml6x\" (UniqueName: \"kubernetes.io/projected/284847a7-94c0-4f03-a54d-6d22fe77353f-kube-api-access-gml6x\") pod \"image-pruner-29401920-82g4w\" (UID: \"284847a7-94c0-4f03-a54d-6d22fe77353f\") " pod="openshift-image-registry/image-pruner-29401920-82g4w" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.433730 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/28b3189b-79c5-4da2-9ce5-d067f66a47c2-image-import-ca\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.433751 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-oauth-serving-cert\") pod \"console-f9d7485db-hqj86\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.434033 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/734b50e7-a98e-4999-bab8-ae2547f08c0c-trusted-ca\") pod \"console-operator-58897d9998-cbzmc\" (UID: \"734b50e7-a98e-4999-bab8-ae2547f08c0c\") " pod="openshift-console-operator/console-operator-58897d9998-cbzmc" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.434132 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/00486d12-48b6-4c9b-b844-78ce031ff80e-proxy-tls\") pod \"machine-config-controller-84d6567774-mq7cp\" (UID: \"00486d12-48b6-4c9b-b844-78ce031ff80e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mq7cp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.434170 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d69vj\" (UniqueName: \"kubernetes.io/projected/5fe4c6f0-f223-41fe-adaf-58434d4f66d8-kube-api-access-d69vj\") pod \"migrator-59844c95c7-dt2nb\" (UID: \"5fe4c6f0-f223-41fe-adaf-58434d4f66d8\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dt2nb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.434201 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28b3189b-79c5-4da2-9ce5-d067f66a47c2-serving-cert\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.434226 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/28602273-98a9-4649-80f3-4e8848b16ee3-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-vvlgv\" (UID: \"28602273-98a9-4649-80f3-4e8848b16ee3\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vvlgv" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.434279 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/645b66a3-584a-4791-8558-64b73b38e96b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-rf9kz\" (UID: \"645b66a3-584a-4791-8558-64b73b38e96b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rf9kz" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.434309 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9aaa727c-565e-4819-801d-9f1787118d20-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-mjvdt\" (UID: \"9aaa727c-565e-4819-801d-9f1787118d20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.434335 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9aaa727c-565e-4819-801d-9f1787118d20-client-ca\") pod \"controller-manager-879f6c89f-mjvdt\" (UID: \"9aaa727c-565e-4819-801d-9f1787118d20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.434480 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-oauth-serving-cert\") pod \"console-f9d7485db-hqj86\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.434479 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.434529 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00869536-c66a-4fae-91e7-0c6d271ab8cd-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rj29g\" (UID: \"00869536-c66a-4fae-91e7-0c6d271ab8cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rj29g" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.434704 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/752819e1-d0ad-4820-979a-bf6ebfc59b4d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-pm5j5\" (UID: \"752819e1-d0ad-4820-979a-bf6ebfc59b4d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pm5j5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.434737 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b3321717-6860-48e4-91a4-99245add293d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-n4vmx\" (UID: \"b3321717-6860-48e4-91a4-99245add293d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4vmx" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.434784 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4ngl\" (UniqueName: \"kubernetes.io/projected/cb710ab4-0d33-4e91-b749-68169fc723c2-kube-api-access-c4ngl\") pod \"machine-approver-56656f9798-bqmvb\" (UID: \"cb710ab4-0d33-4e91-b749-68169fc723c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bqmvb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.434810 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/02a5e11b-33ad-4c1f-8073-2558ba723cd3-metrics-tls\") pod \"ingress-operator-5b745b69d9-7kxvg\" (UID: \"02a5e11b-33ad-4c1f-8073-2558ba723cd3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7kxvg" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.434834 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64bb8da2-3238-49d0-9cb2-b582701fb35d-config\") pod \"etcd-operator-b45778765-8whf2\" (UID: \"64bb8da2-3238-49d0-9cb2-b582701fb35d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.434858 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/28b3189b-79c5-4da2-9ce5-d067f66a47c2-etcd-serving-ca\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.434922 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cb710ab4-0d33-4e91-b749-68169fc723c2-auth-proxy-config\") pod \"machine-approver-56656f9798-bqmvb\" (UID: \"cb710ab4-0d33-4e91-b749-68169fc723c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bqmvb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.434946 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.434975 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8260737-caa6-42f1-8854-ad2b97a1ee42-config\") pod \"machine-api-operator-5694c8668f-x2vlb\" (UID: \"c8260737-caa6-42f1-8854-ad2b97a1ee42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x2vlb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.435002 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bds6d\" (UniqueName: \"kubernetes.io/projected/f1e5e774-51ea-47e8-8b02-1bde1728da35-kube-api-access-bds6d\") pod \"marketplace-operator-79b997595-2vsl7\" (UID: \"f1e5e774-51ea-47e8-8b02-1bde1728da35\") " pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.435030 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7196e744-19f7-4295-9d97-20ddfe1c44b8-signing-cabundle\") pod \"service-ca-9c57cc56f-w62xw\" (UID: \"7196e744-19f7-4295-9d97-20ddfe1c44b8\") " pod="openshift-service-ca/service-ca-9c57cc56f-w62xw" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.435104 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfpd2\" (UniqueName: \"kubernetes.io/projected/7196e744-19f7-4295-9d97-20ddfe1c44b8-kube-api-access-qfpd2\") pod \"service-ca-9c57cc56f-w62xw\" (UID: \"7196e744-19f7-4295-9d97-20ddfe1c44b8\") " pod="openshift-service-ca/service-ca-9c57cc56f-w62xw" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.435139 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e00b34f6-3577-4658-b819-43ada470fbe7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-bnkxg\" (UID: \"e00b34f6-3577-4658-b819-43ada470fbe7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bnkxg" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.435162 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fb2e8401-8c07-4083-8b0f-821cd21b5858-apiservice-cert\") pod \"packageserver-d55dfcdfc-bs5t5\" (UID: \"fb2e8401-8c07-4083-8b0f-821cd21b5858\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.435189 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/64bb8da2-3238-49d0-9cb2-b582701fb35d-etcd-service-ca\") pod \"etcd-operator-b45778765-8whf2\" (UID: \"64bb8da2-3238-49d0-9cb2-b582701fb35d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.435213 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgsrx\" (UniqueName: \"kubernetes.io/projected/b3321717-6860-48e4-91a4-99245add293d-kube-api-access-tgsrx\") pod \"cluster-image-registry-operator-dc59b4c8b-n4vmx\" (UID: \"b3321717-6860-48e4-91a4-99245add293d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4vmx" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.435230 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcv6t\" (UniqueName: \"kubernetes.io/projected/6f7e21ec-878c-4f90-9f5f-51e95d71e373-kube-api-access-jcv6t\") pod \"openshift-apiserver-operator-796bbdcf4f-zndxf\" (UID: \"6f7e21ec-878c-4f90-9f5f-51e95d71e373\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zndxf" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.435798 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/28b3189b-79c5-4da2-9ce5-d067f66a47c2-etcd-serving-ca\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.439804 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1e392e29-fdb8-450d-ac24-ba91e802139a-trusted-ca\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.440244 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9aaa727c-565e-4819-801d-9f1787118d20-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-mjvdt\" (UID: \"9aaa727c-565e-4819-801d-9f1787118d20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.441151 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.441162 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b3321717-6860-48e4-91a4-99245add293d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-n4vmx\" (UID: \"b3321717-6860-48e4-91a4-99245add293d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4vmx" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.441504 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.441676 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1e392e29-fdb8-450d-ac24-ba91e802139a-registry-tls\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.442247 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/28b3189b-79c5-4da2-9ce5-d067f66a47c2-etcd-client\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.442274 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64bb8da2-3238-49d0-9cb2-b582701fb35d-config\") pod \"etcd-operator-b45778765-8whf2\" (UID: \"64bb8da2-3238-49d0-9cb2-b582701fb35d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.442866 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cb710ab4-0d33-4e91-b749-68169fc723c2-auth-proxy-config\") pod \"machine-approver-56656f9798-bqmvb\" (UID: \"cb710ab4-0d33-4e91-b749-68169fc723c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bqmvb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.443970 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8260737-caa6-42f1-8854-ad2b97a1ee42-config\") pod \"machine-api-operator-5694c8668f-x2vlb\" (UID: \"c8260737-caa6-42f1-8854-ad2b97a1ee42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x2vlb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.444375 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9c103959-be49-4b6a-9689-0a58451bc2b7-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8sgwh\" (UID: \"9c103959-be49-4b6a-9689-0a58451bc2b7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8sgwh" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.444472 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-trusted-ca-bundle\") pod \"console-f9d7485db-hqj86\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.444526 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/abf46edd-34b9-42e7-94ca-a3066047cb53-mountpoint-dir\") pod \"csi-hostpathplugin-g4bjk\" (UID: \"abf46edd-34b9-42e7-94ca-a3066047cb53\") " pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.444565 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/28b3189b-79c5-4da2-9ce5-d067f66a47c2-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.444597 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/abf46edd-34b9-42e7-94ca-a3066047cb53-registration-dir\") pod \"csi-hostpathplugin-g4bjk\" (UID: \"abf46edd-34b9-42e7-94ca-a3066047cb53\") " pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.444757 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/64bb8da2-3238-49d0-9cb2-b582701fb35d-etcd-ca\") pod \"etcd-operator-b45778765-8whf2\" (UID: \"64bb8da2-3238-49d0-9cb2-b582701fb35d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.444849 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45bb08cb-d545-4568-9a64-780628c87786-config\") pod \"authentication-operator-69f744f599-blhhp\" (UID: \"45bb08cb-d545-4568-9a64-780628c87786\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-blhhp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.445910 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-trusted-ca-bundle\") pod \"console-f9d7485db-hqj86\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.446022 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.446447 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/64bb8da2-3238-49d0-9cb2-b582701fb35d-etcd-service-ca\") pod \"etcd-operator-b45778765-8whf2\" (UID: \"64bb8da2-3238-49d0-9cb2-b582701fb35d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.446542 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk9v8\" (UniqueName: \"kubernetes.io/projected/c8260737-caa6-42f1-8854-ad2b97a1ee42-kube-api-access-mk9v8\") pod \"machine-api-operator-5694c8668f-x2vlb\" (UID: \"c8260737-caa6-42f1-8854-ad2b97a1ee42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x2vlb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.446606 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/64bb8da2-3238-49d0-9cb2-b582701fb35d-etcd-client\") pod \"etcd-operator-b45778765-8whf2\" (UID: \"64bb8da2-3238-49d0-9cb2-b582701fb35d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.446746 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f1e5e774-51ea-47e8-8b02-1bde1728da35-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2vsl7\" (UID: \"f1e5e774-51ea-47e8-8b02-1bde1728da35\") " pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.446755 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45bb08cb-d545-4568-9a64-780628c87786-config\") pod \"authentication-operator-69f744f599-blhhp\" (UID: \"45bb08cb-d545-4568-9a64-780628c87786\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-blhhp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.446838 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/64bb8da2-3238-49d0-9cb2-b582701fb35d-etcd-ca\") pod \"etcd-operator-b45778765-8whf2\" (UID: \"64bb8da2-3238-49d0-9cb2-b582701fb35d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.446921 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e00b34f6-3577-4658-b819-43ada470fbe7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-bnkxg\" (UID: \"e00b34f6-3577-4658-b819-43ada470fbe7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bnkxg" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.446929 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/26e7d7f8-eccb-479a-bd88-69e526049c18-certs\") pod \"machine-config-server-zbsq2\" (UID: \"26e7d7f8-eccb-479a-bd88-69e526049c18\") " pod="openshift-machine-config-operator/machine-config-server-zbsq2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.447127 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9aaa727c-565e-4819-801d-9f1787118d20-serving-cert\") pod \"controller-manager-879f6c89f-mjvdt\" (UID: \"9aaa727c-565e-4819-801d-9f1787118d20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.447167 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/cb710ab4-0d33-4e91-b749-68169fc723c2-machine-approver-tls\") pod \"machine-approver-56656f9798-bqmvb\" (UID: \"cb710ab4-0d33-4e91-b749-68169fc723c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bqmvb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.447819 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zflng\" (UniqueName: \"kubernetes.io/projected/cc0815f7-09db-4adf-b525-205c8fd8e64e-kube-api-access-zflng\") pod \"openshift-config-operator-7777fb866f-59zt8\" (UID: \"cc0815f7-09db-4adf-b525-205c8fd8e64e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-59zt8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.447897 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/06530996-dd63-4ed5-adb3-790ecc93d11e-metrics-certs\") pod \"router-default-5444994796-45dcj\" (UID: \"06530996-dd63-4ed5-adb3-790ecc93d11e\") " pod="openshift-ingress/router-default-5444994796-45dcj" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.447935 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vt5bm\" (UniqueName: \"kubernetes.io/projected/06530996-dd63-4ed5-adb3-790ecc93d11e-kube-api-access-vt5bm\") pod \"router-default-5444994796-45dcj\" (UID: \"06530996-dd63-4ed5-adb3-790ecc93d11e\") " pod="openshift-ingress/router-default-5444994796-45dcj" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.447985 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67wjp\" (UniqueName: \"kubernetes.io/projected/d94c7fca-2c24-4005-9839-0578e1b39283-kube-api-access-67wjp\") pod \"downloads-7954f5f757-bz6r8\" (UID: \"d94c7fca-2c24-4005-9839-0578e1b39283\") " pod="openshift-console/downloads-7954f5f757-bz6r8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.448019 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj595\" (UniqueName: \"kubernetes.io/projected/c71045e6-eb17-417d-b9fe-12d3a21f4a56-kube-api-access-gj595\") pod \"service-ca-operator-777779d784-4qsp7\" (UID: \"c71045e6-eb17-417d-b9fe-12d3a21f4a56\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qsp7" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.448078 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9aaa727c-565e-4819-801d-9f1787118d20-client-ca\") pod \"controller-manager-879f6c89f-mjvdt\" (UID: \"9aaa727c-565e-4819-801d-9f1787118d20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.448171 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpp5v\" (UniqueName: \"kubernetes.io/projected/36e16640-f72c-4e52-9c0d-61d37c1f91d8-kube-api-access-dpp5v\") pod \"ingress-canary-8kdj9\" (UID: \"36e16640-f72c-4e52-9c0d-61d37c1f91d8\") " pod="openshift-ingress-canary/ingress-canary-8kdj9" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.448237 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.448314 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.448505 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbztm\" (UniqueName: \"kubernetes.io/projected/c91e0334-6829-4a8f-9239-3930e84e65c4-kube-api-access-sbztm\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.448908 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.448954 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1e392e29-fdb8-450d-ac24-ba91e802139a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.449075 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.449122 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/53fb45a4-5f96-4e9c-a231-7e840bd5c79f-srv-cert\") pod \"catalog-operator-68c6474976-xbr6t\" (UID: \"53fb45a4-5f96-4e9c-a231-7e840bd5c79f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xbr6t" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.449271 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45bb08cb-d545-4568-9a64-780628c87786-serving-cert\") pod \"authentication-operator-69f744f599-blhhp\" (UID: \"45bb08cb-d545-4568-9a64-780628c87786\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-blhhp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.449329 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mn9b\" (UniqueName: \"kubernetes.io/projected/c3dad121-17b2-4c92-a99f-34439a678fa9-kube-api-access-8mn9b\") pod \"dns-default-k2shv\" (UID: \"c3dad121-17b2-4c92-a99f-34439a678fa9\") " pod="openshift-dns/dns-default-k2shv" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.449339 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.449361 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2692q\" (UniqueName: \"kubernetes.io/projected/26e7d7f8-eccb-479a-bd88-69e526049c18-kube-api-access-2692q\") pod \"machine-config-server-zbsq2\" (UID: \"26e7d7f8-eccb-479a-bd88-69e526049c18\") " pod="openshift-machine-config-operator/machine-config-server-zbsq2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.449496 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/45bb08cb-d545-4568-9a64-780628c87786-service-ca-bundle\") pod \"authentication-operator-69f744f599-blhhp\" (UID: \"45bb08cb-d545-4568-9a64-780628c87786\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-blhhp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.449524 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06530996-dd63-4ed5-adb3-790ecc93d11e-service-ca-bundle\") pod \"router-default-5444994796-45dcj\" (UID: \"06530996-dd63-4ed5-adb3-790ecc93d11e\") " pod="openshift-ingress/router-default-5444994796-45dcj" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.449558 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kb9mk\" (UniqueName: \"kubernetes.io/projected/20bb15a2-ef87-40cf-b450-f2a50a290325-kube-api-access-kb9mk\") pod \"dns-operator-744455d44c-z22jf\" (UID: \"20bb15a2-ef87-40cf-b450-f2a50a290325\") " pod="openshift-dns-operator/dns-operator-744455d44c-z22jf" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.449594 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/02a5e11b-33ad-4c1f-8073-2558ba723cd3-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7kxvg\" (UID: \"02a5e11b-33ad-4c1f-8073-2558ba723cd3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7kxvg" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.449625 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/28b3189b-79c5-4da2-9ce5-d067f66a47c2-audit-dir\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.449670 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00869536-c66a-4fae-91e7-0c6d271ab8cd-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rj29g\" (UID: \"00869536-c66a-4fae-91e7-0c6d271ab8cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rj29g" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.449699 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64bb8da2-3238-49d0-9cb2-b582701fb35d-serving-cert\") pod \"etcd-operator-b45778765-8whf2\" (UID: \"64bb8da2-3238-49d0-9cb2-b582701fb35d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.449734 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f7e21ec-878c-4f90-9f5f-51e95d71e373-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zndxf\" (UID: \"6f7e21ec-878c-4f90-9f5f-51e95d71e373\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zndxf" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.449762 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-service-ca\") pod \"console-f9d7485db-hqj86\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.449792 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfn5j\" (UniqueName: \"kubernetes.io/projected/6714d5a2-a6be-4841-8268-66de19d238ec-kube-api-access-sfn5j\") pod \"olm-operator-6b444d44fb-tnhkb\" (UID: \"6714d5a2-a6be-4841-8268-66de19d238ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tnhkb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.449838 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f1e5e774-51ea-47e8-8b02-1bde1728da35-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2vsl7\" (UID: \"f1e5e774-51ea-47e8-8b02-1bde1728da35\") " pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.449868 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/26e7d7f8-eccb-479a-bd88-69e526049c18-node-bootstrap-token\") pod \"machine-config-server-zbsq2\" (UID: \"26e7d7f8-eccb-479a-bd88-69e526049c18\") " pod="openshift-machine-config-operator/machine-config-server-zbsq2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.449933 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c3dad121-17b2-4c92-a99f-34439a678fa9-config-volume\") pod \"dns-default-k2shv\" (UID: \"c3dad121-17b2-4c92-a99f-34439a678fa9\") " pod="openshift-dns/dns-default-k2shv" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.449969 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/284847a7-94c0-4f03-a54d-6d22fe77353f-serviceca\") pod \"image-pruner-29401920-82g4w\" (UID: \"284847a7-94c0-4f03-a54d-6d22fe77353f\") " pod="openshift-image-registry/image-pruner-29401920-82g4w" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450000 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc0815f7-09db-4adf-b525-205c8fd8e64e-serving-cert\") pod \"openshift-config-operator-7777fb866f-59zt8\" (UID: \"cc0815f7-09db-4adf-b525-205c8fd8e64e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-59zt8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450031 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c8260737-caa6-42f1-8854-ad2b97a1ee42-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-x2vlb\" (UID: \"c8260737-caa6-42f1-8854-ad2b97a1ee42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x2vlb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450061 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-console-config\") pod \"console-f9d7485db-hqj86\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450090 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc2ff29c-9c87-44f9-95f9-7043db18ffe3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5thz5\" (UID: \"bc2ff29c-9c87-44f9-95f9-7043db18ffe3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5thz5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450120 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aaa727c-565e-4819-801d-9f1787118d20-config\") pod \"controller-manager-879f6c89f-mjvdt\" (UID: \"9aaa727c-565e-4819-801d-9f1787118d20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450150 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/734b50e7-a98e-4999-bab8-ae2547f08c0c-config\") pod \"console-operator-58897d9998-cbzmc\" (UID: \"734b50e7-a98e-4999-bab8-ae2547f08c0c\") " pod="openshift-console-operator/console-operator-58897d9998-cbzmc" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450182 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/fb2e8401-8c07-4083-8b0f-821cd21b5858-tmpfs\") pod \"packageserver-d55dfcdfc-bs5t5\" (UID: \"fb2e8401-8c07-4083-8b0f-821cd21b5858\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450207 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/abf46edd-34b9-42e7-94ca-a3066047cb53-csi-data-dir\") pod \"csi-hostpathplugin-g4bjk\" (UID: \"abf46edd-34b9-42e7-94ca-a3066047cb53\") " pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450239 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c3dad121-17b2-4c92-a99f-34439a678fa9-metrics-tls\") pod \"dns-default-k2shv\" (UID: \"c3dad121-17b2-4c92-a99f-34439a678fa9\") " pod="openshift-dns/dns-default-k2shv" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450270 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450300 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjp9n\" (UniqueName: \"kubernetes.io/projected/28b3189b-79c5-4da2-9ce5-d067f66a47c2-kube-api-access-pjp9n\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450333 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1e392e29-fdb8-450d-ac24-ba91e802139a-registry-certificates\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450339 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28b3189b-79c5-4da2-9ce5-d067f66a47c2-serving-cert\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450357 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx6tx\" (UniqueName: \"kubernetes.io/projected/734b50e7-a98e-4999-bab8-ae2547f08c0c-kube-api-access-wx6tx\") pod \"console-operator-58897d9998-cbzmc\" (UID: \"734b50e7-a98e-4999-bab8-ae2547f08c0c\") " pod="openshift-console-operator/console-operator-58897d9998-cbzmc" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450462 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f7e21ec-878c-4f90-9f5f-51e95d71e373-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zndxf\" (UID: \"6f7e21ec-878c-4f90-9f5f-51e95d71e373\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zndxf" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450529 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/28b3189b-79c5-4da2-9ce5-d067f66a47c2-audit-dir\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450611 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88bpw\" (UniqueName: \"kubernetes.io/projected/fb2e8401-8c07-4083-8b0f-821cd21b5858-kube-api-access-88bpw\") pod \"packageserver-d55dfcdfc-bs5t5\" (UID: \"fb2e8401-8c07-4083-8b0f-821cd21b5858\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450678 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-console-oauth-config\") pod \"console-f9d7485db-hqj86\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450732 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bz8z\" (UniqueName: \"kubernetes.io/projected/00486d12-48b6-4c9b-b844-78ce031ff80e-kube-api-access-4bz8z\") pod \"machine-config-controller-84d6567774-mq7cp\" (UID: \"00486d12-48b6-4c9b-b844-78ce031ff80e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mq7cp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450764 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44q2z\" (UniqueName: \"kubernetes.io/projected/53fb45a4-5f96-4e9c-a231-7e840bd5c79f-kube-api-access-44q2z\") pod \"catalog-operator-68c6474976-xbr6t\" (UID: \"53fb45a4-5f96-4e9c-a231-7e840bd5c79f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xbr6t" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450812 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/734b50e7-a98e-4999-bab8-ae2547f08c0c-serving-cert\") pod \"console-operator-58897d9998-cbzmc\" (UID: \"734b50e7-a98e-4999-bab8-ae2547f08c0c\") " pod="openshift-console-operator/console-operator-58897d9998-cbzmc" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450845 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28b3189b-79c5-4da2-9ce5-d067f66a47c2-config\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450895 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b3321717-6860-48e4-91a4-99245add293d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-n4vmx\" (UID: \"b3321717-6860-48e4-91a4-99245add293d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4vmx" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450926 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/28b3189b-79c5-4da2-9ce5-d067f66a47c2-encryption-config\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450972 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/40ba7989-d314-4273-b2e1-aaf05d4b50a4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-z5kzz\" (UID: \"40ba7989-d314-4273-b2e1-aaf05d4b50a4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z5kzz" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.450996 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/20bb15a2-ef87-40cf-b450-f2a50a290325-metrics-tls\") pod \"dns-operator-744455d44c-z22jf\" (UID: \"20bb15a2-ef87-40cf-b450-f2a50a290325\") " pod="openshift-dns-operator/dns-operator-744455d44c-z22jf" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451022 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/cc0815f7-09db-4adf-b525-205c8fd8e64e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-59zt8\" (UID: \"cc0815f7-09db-4adf-b525-205c8fd8e64e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-59zt8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451044 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/28b3189b-79c5-4da2-9ce5-d067f66a47c2-audit\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451066 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-console-serving-cert\") pod \"console-f9d7485db-hqj86\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451104 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/06530996-dd63-4ed5-adb3-790ecc93d11e-default-certificate\") pod \"router-default-5444994796-45dcj\" (UID: \"06530996-dd63-4ed5-adb3-790ecc93d11e\") " pod="openshift-ingress/router-default-5444994796-45dcj" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451128 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/06530996-dd63-4ed5-adb3-790ecc93d11e-stats-auth\") pod \"router-default-5444994796-45dcj\" (UID: \"06530996-dd63-4ed5-adb3-790ecc93d11e\") " pod="openshift-ingress/router-default-5444994796-45dcj" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451158 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4e7b24da-83fa-446d-b2a9-97bf382ff2c5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7z6k\" (UID: \"4e7b24da-83fa-446d-b2a9-97bf382ff2c5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7z6k" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451200 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/28b3189b-79c5-4da2-9ce5-d067f66a47c2-node-pullsecrets\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451231 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e7b24da-83fa-446d-b2a9-97bf382ff2c5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7z6k\" (UID: \"4e7b24da-83fa-446d-b2a9-97bf382ff2c5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7z6k" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451257 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk7cf\" (UniqueName: \"kubernetes.io/projected/9c103959-be49-4b6a-9689-0a58451bc2b7-kube-api-access-qk7cf\") pod \"cluster-samples-operator-665b6dd947-8sgwh\" (UID: \"9c103959-be49-4b6a-9689-0a58451bc2b7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8sgwh" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451282 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7np2\" (UniqueName: \"kubernetes.io/projected/c69cf39a-28f6-491f-8563-6063ffef11aa-kube-api-access-b7np2\") pod \"machine-config-operator-74547568cd-p5mt6\" (UID: \"c69cf39a-28f6-491f-8563-6063ffef11aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p5mt6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451310 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz52x\" (UniqueName: \"kubernetes.io/projected/02a5e11b-33ad-4c1f-8073-2558ba723cd3-kube-api-access-pz52x\") pod \"ingress-operator-5b745b69d9-7kxvg\" (UID: \"02a5e11b-33ad-4c1f-8073-2558ba723cd3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7kxvg" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451340 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/645b66a3-584a-4791-8558-64b73b38e96b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-rf9kz\" (UID: \"645b66a3-584a-4791-8558-64b73b38e96b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rf9kz" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451374 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js6sh\" (UniqueName: \"kubernetes.io/projected/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-kube-api-access-js6sh\") pod \"console-f9d7485db-hqj86\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451405 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1e392e29-fdb8-450d-ac24-ba91e802139a-bound-sa-token\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451433 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/45bb08cb-d545-4568-9a64-780628c87786-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-blhhp\" (UID: \"45bb08cb-d545-4568-9a64-780628c87786\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-blhhp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451466 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/36e16640-f72c-4e52-9c0d-61d37c1f91d8-cert\") pod \"ingress-canary-8kdj9\" (UID: \"36e16640-f72c-4e52-9c0d-61d37c1f91d8\") " pod="openshift-ingress-canary/ingress-canary-8kdj9" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451499 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c91e0334-6829-4a8f-9239-3930e84e65c4-audit-dir\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451536 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451557 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4vzv\" (UniqueName: \"kubernetes.io/projected/dea338f1-f870-484b-ac43-7da254eac47a-kube-api-access-k4vzv\") pod \"collect-profiles-29401935-hhtv9\" (UID: \"dea338f1-f870-484b-ac43-7da254eac47a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451579 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c69cf39a-28f6-491f-8563-6063ffef11aa-proxy-tls\") pod \"machine-config-operator-74547568cd-p5mt6\" (UID: \"c69cf39a-28f6-491f-8563-6063ffef11aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p5mt6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451600 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c69cf39a-28f6-491f-8563-6063ffef11aa-auth-proxy-config\") pod \"machine-config-operator-74547568cd-p5mt6\" (UID: \"c69cf39a-28f6-491f-8563-6063ffef11aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p5mt6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451626 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-audit-policies\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451663 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/00486d12-48b6-4c9b-b844-78ce031ff80e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-mq7cp\" (UID: \"00486d12-48b6-4c9b-b844-78ce031ff80e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mq7cp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451685 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6714d5a2-a6be-4841-8268-66de19d238ec-srv-cert\") pod \"olm-operator-6b444d44fb-tnhkb\" (UID: \"6714d5a2-a6be-4841-8268-66de19d238ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tnhkb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451710 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c69cf39a-28f6-491f-8563-6063ffef11aa-images\") pod \"machine-config-operator-74547568cd-p5mt6\" (UID: \"c69cf39a-28f6-491f-8563-6063ffef11aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p5mt6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451735 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/53fb45a4-5f96-4e9c-a231-7e840bd5c79f-profile-collector-cert\") pod \"catalog-operator-68c6474976-xbr6t\" (UID: \"53fb45a4-5f96-4e9c-a231-7e840bd5c79f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xbr6t" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451763 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451785 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9z2xg\" (UniqueName: \"kubernetes.io/projected/45bb08cb-d545-4568-9a64-780628c87786-kube-api-access-9z2xg\") pod \"authentication-operator-69f744f599-blhhp\" (UID: \"45bb08cb-d545-4568-9a64-780628c87786\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-blhhp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451809 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c8260737-caa6-42f1-8854-ad2b97a1ee42-images\") pod \"machine-api-operator-5694c8668f-x2vlb\" (UID: \"c8260737-caa6-42f1-8854-ad2b97a1ee42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x2vlb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451843 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c71045e6-eb17-417d-b9fe-12d3a21f4a56-config\") pod \"service-ca-operator-777779d784-4qsp7\" (UID: \"c71045e6-eb17-417d-b9fe-12d3a21f4a56\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qsp7" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451885 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451908 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/645b66a3-584a-4791-8558-64b73b38e96b-config\") pod \"kube-controller-manager-operator-78b949d7b-rf9kz\" (UID: \"645b66a3-584a-4791-8558-64b73b38e96b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rf9kz" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.452490 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.452543 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e00b34f6-3577-4658-b819-43ada470fbe7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-bnkxg\" (UID: \"e00b34f6-3577-4658-b819-43ada470fbe7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bnkxg" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.452697 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f7e21ec-878c-4f90-9f5f-51e95d71e373-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zndxf\" (UID: \"6f7e21ec-878c-4f90-9f5f-51e95d71e373\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zndxf" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.451397 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9aaa727c-565e-4819-801d-9f1787118d20-serving-cert\") pod \"controller-manager-879f6c89f-mjvdt\" (UID: \"9aaa727c-565e-4819-801d-9f1787118d20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.453344 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-console-config\") pod \"console-f9d7485db-hqj86\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.453883 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/45bb08cb-d545-4568-9a64-780628c87786-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-blhhp\" (UID: \"45bb08cb-d545-4568-9a64-780628c87786\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-blhhp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.453943 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c91e0334-6829-4a8f-9239-3930e84e65c4-audit-dir\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.454398 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/28b3189b-79c5-4da2-9ce5-d067f66a47c2-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.454765 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aaa727c-565e-4819-801d-9f1787118d20-config\") pod \"controller-manager-879f6c89f-mjvdt\" (UID: \"9aaa727c-565e-4819-801d-9f1787118d20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.454894 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-console-oauth-config\") pod \"console-f9d7485db-hqj86\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.455489 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/cb710ab4-0d33-4e91-b749-68169fc723c2-machine-approver-tls\") pod \"machine-approver-56656f9798-bqmvb\" (UID: \"cb710ab4-0d33-4e91-b749-68169fc723c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bqmvb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.456191 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28b3189b-79c5-4da2-9ce5-d067f66a47c2-config\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.456319 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/28b3189b-79c5-4da2-9ce5-d067f66a47c2-node-pullsecrets\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.456562 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00869536-c66a-4fae-91e7-0c6d271ab8cd-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rj29g\" (UID: \"00869536-c66a-4fae-91e7-0c6d271ab8cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rj29g" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.456583 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/734b50e7-a98e-4999-bab8-ae2547f08c0c-config\") pod \"console-operator-58897d9998-cbzmc\" (UID: \"734b50e7-a98e-4999-bab8-ae2547f08c0c\") " pod="openshift-console-operator/console-operator-58897d9998-cbzmc" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.456720 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/284847a7-94c0-4f03-a54d-6d22fe77353f-serviceca\") pod \"image-pruner-29401920-82g4w\" (UID: \"284847a7-94c0-4f03-a54d-6d22fe77353f\") " pod="openshift-image-registry/image-pruner-29401920-82g4w" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.456912 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/28b3189b-79c5-4da2-9ce5-d067f66a47c2-audit\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.457366 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-audit-policies\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.457372 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/cc0815f7-09db-4adf-b525-205c8fd8e64e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-59zt8\" (UID: \"cc0815f7-09db-4adf-b525-205c8fd8e64e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-59zt8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.458086 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.459192 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c8260737-caa6-42f1-8854-ad2b97a1ee42-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-x2vlb\" (UID: \"c8260737-caa6-42f1-8854-ad2b97a1ee42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x2vlb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.460675 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b3321717-6860-48e4-91a4-99245add293d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-n4vmx\" (UID: \"b3321717-6860-48e4-91a4-99245add293d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4vmx" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.460705 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c8260737-caa6-42f1-8854-ad2b97a1ee42-images\") pod \"machine-api-operator-5694c8668f-x2vlb\" (UID: \"c8260737-caa6-42f1-8854-ad2b97a1ee42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x2vlb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.460713 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/734b50e7-a98e-4999-bab8-ae2547f08c0c-serving-cert\") pod \"console-operator-58897d9998-cbzmc\" (UID: \"734b50e7-a98e-4999-bab8-ae2547f08c0c\") " pod="openshift-console-operator/console-operator-58897d9998-cbzmc" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.461770 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-console-serving-cert\") pod \"console-f9d7485db-hqj86\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.461871 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/64bb8da2-3238-49d0-9cb2-b582701fb35d-etcd-client\") pod \"etcd-operator-b45778765-8whf2\" (UID: \"64bb8da2-3238-49d0-9cb2-b582701fb35d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.462075 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/45bb08cb-d545-4568-9a64-780628c87786-service-ca-bundle\") pod \"authentication-operator-69f744f599-blhhp\" (UID: \"45bb08cb-d545-4568-9a64-780628c87786\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-blhhp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.462128 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-service-ca\") pod \"console-f9d7485db-hqj86\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.463030 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1e392e29-fdb8-450d-ac24-ba91e802139a-registry-certificates\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.463710 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.464801 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc0815f7-09db-4adf-b525-205c8fd8e64e-serving-cert\") pod \"openshift-config-operator-7777fb866f-59zt8\" (UID: \"cc0815f7-09db-4adf-b525-205c8fd8e64e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-59zt8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.464987 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/20bb15a2-ef87-40cf-b450-f2a50a290325-metrics-tls\") pod \"dns-operator-744455d44c-z22jf\" (UID: \"20bb15a2-ef87-40cf-b450-f2a50a290325\") " pod="openshift-dns-operator/dns-operator-744455d44c-z22jf" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.465817 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f7e21ec-878c-4f90-9f5f-51e95d71e373-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zndxf\" (UID: \"6f7e21ec-878c-4f90-9f5f-51e95d71e373\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zndxf" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.466679 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.467050 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/28b3189b-79c5-4da2-9ce5-d067f66a47c2-encryption-config\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.468426 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45bb08cb-d545-4568-9a64-780628c87786-serving-cert\") pod \"authentication-operator-69f744f599-blhhp\" (UID: \"45bb08cb-d545-4568-9a64-780628c87786\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-blhhp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.470183 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64bb8da2-3238-49d0-9cb2-b582701fb35d-serving-cert\") pod \"etcd-operator-b45778765-8whf2\" (UID: \"64bb8da2-3238-49d0-9cb2-b582701fb35d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.472704 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.480577 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6j56\" (UniqueName: \"kubernetes.io/projected/9aaa727c-565e-4819-801d-9f1787118d20-kube-api-access-r6j56\") pod \"controller-manager-879f6c89f-mjvdt\" (UID: \"9aaa727c-565e-4819-801d-9f1787118d20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.496019 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs6k7\" (UniqueName: \"kubernetes.io/projected/e00b34f6-3577-4658-b819-43ada470fbe7-kube-api-access-fs6k7\") pod \"openshift-controller-manager-operator-756b6f6bc6-bnkxg\" (UID: \"e00b34f6-3577-4658-b819-43ada470fbe7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bnkxg" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.515462 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b3321717-6860-48e4-91a4-99245add293d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-n4vmx\" (UID: \"b3321717-6860-48e4-91a4-99245add293d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4vmx" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.536452 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/00869536-c66a-4fae-91e7-0c6d271ab8cd-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rj29g\" (UID: \"00869536-c66a-4fae-91e7-0c6d271ab8cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rj29g" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.552135 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwjt5\" (UniqueName: \"kubernetes.io/projected/64bb8da2-3238-49d0-9cb2-b582701fb35d-kube-api-access-rwjt5\") pod \"etcd-operator-b45778765-8whf2\" (UID: \"64bb8da2-3238-49d0-9cb2-b582701fb35d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" Nov 26 00:26:00 crc kubenswrapper[4766]: E1126 00:26:00.552533 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:01.052511832 +0000 UTC m=+141.901282272 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.552368 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.552971 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc2ff29c-9c87-44f9-95f9-7043db18ffe3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5thz5\" (UID: \"bc2ff29c-9c87-44f9-95f9-7043db18ffe3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5thz5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553017 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/fb2e8401-8c07-4083-8b0f-821cd21b5858-tmpfs\") pod \"packageserver-d55dfcdfc-bs5t5\" (UID: \"fb2e8401-8c07-4083-8b0f-821cd21b5858\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553048 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/abf46edd-34b9-42e7-94ca-a3066047cb53-csi-data-dir\") pod \"csi-hostpathplugin-g4bjk\" (UID: \"abf46edd-34b9-42e7-94ca-a3066047cb53\") " pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553068 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c3dad121-17b2-4c92-a99f-34439a678fa9-metrics-tls\") pod \"dns-default-k2shv\" (UID: \"c3dad121-17b2-4c92-a99f-34439a678fa9\") " pod="openshift-dns/dns-default-k2shv" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553108 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88bpw\" (UniqueName: \"kubernetes.io/projected/fb2e8401-8c07-4083-8b0f-821cd21b5858-kube-api-access-88bpw\") pod \"packageserver-d55dfcdfc-bs5t5\" (UID: \"fb2e8401-8c07-4083-8b0f-821cd21b5858\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553140 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bz8z\" (UniqueName: \"kubernetes.io/projected/00486d12-48b6-4c9b-b844-78ce031ff80e-kube-api-access-4bz8z\") pod \"machine-config-controller-84d6567774-mq7cp\" (UID: \"00486d12-48b6-4c9b-b844-78ce031ff80e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mq7cp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553166 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44q2z\" (UniqueName: \"kubernetes.io/projected/53fb45a4-5f96-4e9c-a231-7e840bd5c79f-kube-api-access-44q2z\") pod \"catalog-operator-68c6474976-xbr6t\" (UID: \"53fb45a4-5f96-4e9c-a231-7e840bd5c79f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xbr6t" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553187 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/40ba7989-d314-4273-b2e1-aaf05d4b50a4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-z5kzz\" (UID: \"40ba7989-d314-4273-b2e1-aaf05d4b50a4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z5kzz" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553210 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/06530996-dd63-4ed5-adb3-790ecc93d11e-default-certificate\") pod \"router-default-5444994796-45dcj\" (UID: \"06530996-dd63-4ed5-adb3-790ecc93d11e\") " pod="openshift-ingress/router-default-5444994796-45dcj" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553232 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/06530996-dd63-4ed5-adb3-790ecc93d11e-stats-auth\") pod \"router-default-5444994796-45dcj\" (UID: \"06530996-dd63-4ed5-adb3-790ecc93d11e\") " pod="openshift-ingress/router-default-5444994796-45dcj" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553255 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4e7b24da-83fa-446d-b2a9-97bf382ff2c5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7z6k\" (UID: \"4e7b24da-83fa-446d-b2a9-97bf382ff2c5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7z6k" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553289 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e7b24da-83fa-446d-b2a9-97bf382ff2c5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7z6k\" (UID: \"4e7b24da-83fa-446d-b2a9-97bf382ff2c5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7z6k" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553290 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/abf46edd-34b9-42e7-94ca-a3066047cb53-csi-data-dir\") pod \"csi-hostpathplugin-g4bjk\" (UID: \"abf46edd-34b9-42e7-94ca-a3066047cb53\") " pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553313 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7np2\" (UniqueName: \"kubernetes.io/projected/c69cf39a-28f6-491f-8563-6063ffef11aa-kube-api-access-b7np2\") pod \"machine-config-operator-74547568cd-p5mt6\" (UID: \"c69cf39a-28f6-491f-8563-6063ffef11aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p5mt6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553445 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz52x\" (UniqueName: \"kubernetes.io/projected/02a5e11b-33ad-4c1f-8073-2558ba723cd3-kube-api-access-pz52x\") pod \"ingress-operator-5b745b69d9-7kxvg\" (UID: \"02a5e11b-33ad-4c1f-8073-2558ba723cd3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7kxvg" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553523 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/645b66a3-584a-4791-8558-64b73b38e96b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-rf9kz\" (UID: \"645b66a3-584a-4791-8558-64b73b38e96b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rf9kz" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553545 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/fb2e8401-8c07-4083-8b0f-821cd21b5858-tmpfs\") pod \"packageserver-d55dfcdfc-bs5t5\" (UID: \"fb2e8401-8c07-4083-8b0f-821cd21b5858\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553611 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/36e16640-f72c-4e52-9c0d-61d37c1f91d8-cert\") pod \"ingress-canary-8kdj9\" (UID: \"36e16640-f72c-4e52-9c0d-61d37c1f91d8\") " pod="openshift-ingress-canary/ingress-canary-8kdj9" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553700 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4vzv\" (UniqueName: \"kubernetes.io/projected/dea338f1-f870-484b-ac43-7da254eac47a-kube-api-access-k4vzv\") pod \"collect-profiles-29401935-hhtv9\" (UID: \"dea338f1-f870-484b-ac43-7da254eac47a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553789 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c69cf39a-28f6-491f-8563-6063ffef11aa-proxy-tls\") pod \"machine-config-operator-74547568cd-p5mt6\" (UID: \"c69cf39a-28f6-491f-8563-6063ffef11aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p5mt6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553838 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c69cf39a-28f6-491f-8563-6063ffef11aa-auth-proxy-config\") pod \"machine-config-operator-74547568cd-p5mt6\" (UID: \"c69cf39a-28f6-491f-8563-6063ffef11aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p5mt6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553891 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/00486d12-48b6-4c9b-b844-78ce031ff80e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-mq7cp\" (UID: \"00486d12-48b6-4c9b-b844-78ce031ff80e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mq7cp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.553954 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6714d5a2-a6be-4841-8268-66de19d238ec-srv-cert\") pod \"olm-operator-6b444d44fb-tnhkb\" (UID: \"6714d5a2-a6be-4841-8268-66de19d238ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tnhkb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.554004 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c69cf39a-28f6-491f-8563-6063ffef11aa-images\") pod \"machine-config-operator-74547568cd-p5mt6\" (UID: \"c69cf39a-28f6-491f-8563-6063ffef11aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p5mt6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.554050 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/53fb45a4-5f96-4e9c-a231-7e840bd5c79f-profile-collector-cert\") pod \"catalog-operator-68c6474976-xbr6t\" (UID: \"53fb45a4-5f96-4e9c-a231-7e840bd5c79f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xbr6t" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.554102 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c71045e6-eb17-417d-b9fe-12d3a21f4a56-config\") pod \"service-ca-operator-777779d784-4qsp7\" (UID: \"c71045e6-eb17-417d-b9fe-12d3a21f4a56\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qsp7" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.554150 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/645b66a3-584a-4791-8558-64b73b38e96b-config\") pod \"kube-controller-manager-operator-78b949d7b-rf9kz\" (UID: \"645b66a3-584a-4791-8558-64b73b38e96b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rf9kz" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.554194 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.554229 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwtf5\" (UniqueName: \"kubernetes.io/projected/abf46edd-34b9-42e7-94ca-a3066047cb53-kube-api-access-bwtf5\") pod \"csi-hostpathplugin-g4bjk\" (UID: \"abf46edd-34b9-42e7-94ca-a3066047cb53\") " pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.554281 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvqdb\" (UniqueName: \"kubernetes.io/projected/28602273-98a9-4649-80f3-4e8848b16ee3-kube-api-access-kvqdb\") pod \"package-server-manager-789f6589d5-vvlgv\" (UID: \"28602273-98a9-4649-80f3-4e8848b16ee3\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vvlgv" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.555539 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7196e744-19f7-4295-9d97-20ddfe1c44b8-signing-key\") pod \"service-ca-9c57cc56f-w62xw\" (UID: \"7196e744-19f7-4295-9d97-20ddfe1c44b8\") " pod="openshift-service-ca/service-ca-9c57cc56f-w62xw" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.555591 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/abf46edd-34b9-42e7-94ca-a3066047cb53-plugins-dir\") pod \"csi-hostpathplugin-g4bjk\" (UID: \"abf46edd-34b9-42e7-94ca-a3066047cb53\") " pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.555628 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fb2e8401-8c07-4083-8b0f-821cd21b5858-webhook-cert\") pod \"packageserver-d55dfcdfc-bs5t5\" (UID: \"fb2e8401-8c07-4083-8b0f-821cd21b5858\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.555690 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4xmb\" (UniqueName: \"kubernetes.io/projected/4e7b24da-83fa-446d-b2a9-97bf382ff2c5-kube-api-access-v4xmb\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7z6k\" (UID: \"4e7b24da-83fa-446d-b2a9-97bf382ff2c5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7z6k" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.555724 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6714d5a2-a6be-4841-8268-66de19d238ec-profile-collector-cert\") pod \"olm-operator-6b444d44fb-tnhkb\" (UID: \"6714d5a2-a6be-4841-8268-66de19d238ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tnhkb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.555758 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc2ff29c-9c87-44f9-95f9-7043db18ffe3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5thz5\" (UID: \"bc2ff29c-9c87-44f9-95f9-7043db18ffe3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5thz5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.555800 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dea338f1-f870-484b-ac43-7da254eac47a-config-volume\") pod \"collect-profiles-29401935-hhtv9\" (UID: \"dea338f1-f870-484b-ac43-7da254eac47a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.555871 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62hpk\" (UniqueName: \"kubernetes.io/projected/752819e1-d0ad-4820-979a-bf6ebfc59b4d-kube-api-access-62hpk\") pod \"control-plane-machine-set-operator-78cbb6b69f-pm5j5\" (UID: \"752819e1-d0ad-4820-979a-bf6ebfc59b4d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pm5j5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.555906 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/02a5e11b-33ad-4c1f-8073-2558ba723cd3-trusted-ca\") pod \"ingress-operator-5b745b69d9-7kxvg\" (UID: \"02a5e11b-33ad-4c1f-8073-2558ba723cd3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7kxvg" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.555938 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dea338f1-f870-484b-ac43-7da254eac47a-secret-volume\") pod \"collect-profiles-29401935-hhtv9\" (UID: \"dea338f1-f870-484b-ac43-7da254eac47a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.555971 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc2ff29c-9c87-44f9-95f9-7043db18ffe3-config\") pod \"kube-apiserver-operator-766d6c64bb-5thz5\" (UID: \"bc2ff29c-9c87-44f9-95f9-7043db18ffe3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5thz5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.556004 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5jmt\" (UniqueName: \"kubernetes.io/projected/40ba7989-d314-4273-b2e1-aaf05d4b50a4-kube-api-access-p5jmt\") pod \"multus-admission-controller-857f4d67dd-z5kzz\" (UID: \"40ba7989-d314-4273-b2e1-aaf05d4b50a4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z5kzz" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.556037 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c71045e6-eb17-417d-b9fe-12d3a21f4a56-serving-cert\") pod \"service-ca-operator-777779d784-4qsp7\" (UID: \"c71045e6-eb17-417d-b9fe-12d3a21f4a56\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qsp7" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.556068 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/abf46edd-34b9-42e7-94ca-a3066047cb53-socket-dir\") pod \"csi-hostpathplugin-g4bjk\" (UID: \"abf46edd-34b9-42e7-94ca-a3066047cb53\") " pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.556165 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d69vj\" (UniqueName: \"kubernetes.io/projected/5fe4c6f0-f223-41fe-adaf-58434d4f66d8-kube-api-access-d69vj\") pod \"migrator-59844c95c7-dt2nb\" (UID: \"5fe4c6f0-f223-41fe-adaf-58434d4f66d8\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dt2nb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.556201 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/00486d12-48b6-4c9b-b844-78ce031ff80e-proxy-tls\") pod \"machine-config-controller-84d6567774-mq7cp\" (UID: \"00486d12-48b6-4c9b-b844-78ce031ff80e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mq7cp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.556239 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/28602273-98a9-4649-80f3-4e8848b16ee3-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-vvlgv\" (UID: \"28602273-98a9-4649-80f3-4e8848b16ee3\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vvlgv" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.556272 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/645b66a3-584a-4791-8558-64b73b38e96b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-rf9kz\" (UID: \"645b66a3-584a-4791-8558-64b73b38e96b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rf9kz" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.556322 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/02a5e11b-33ad-4c1f-8073-2558ba723cd3-metrics-tls\") pod \"ingress-operator-5b745b69d9-7kxvg\" (UID: \"02a5e11b-33ad-4c1f-8073-2558ba723cd3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7kxvg" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.556369 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/752819e1-d0ad-4820-979a-bf6ebfc59b4d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-pm5j5\" (UID: \"752819e1-d0ad-4820-979a-bf6ebfc59b4d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pm5j5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.556429 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bds6d\" (UniqueName: \"kubernetes.io/projected/f1e5e774-51ea-47e8-8b02-1bde1728da35-kube-api-access-bds6d\") pod \"marketplace-operator-79b997595-2vsl7\" (UID: \"f1e5e774-51ea-47e8-8b02-1bde1728da35\") " pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.556477 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7196e744-19f7-4295-9d97-20ddfe1c44b8-signing-cabundle\") pod \"service-ca-9c57cc56f-w62xw\" (UID: \"7196e744-19f7-4295-9d97-20ddfe1c44b8\") " pod="openshift-service-ca/service-ca-9c57cc56f-w62xw" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.556527 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfpd2\" (UniqueName: \"kubernetes.io/projected/7196e744-19f7-4295-9d97-20ddfe1c44b8-kube-api-access-qfpd2\") pod \"service-ca-9c57cc56f-w62xw\" (UID: \"7196e744-19f7-4295-9d97-20ddfe1c44b8\") " pod="openshift-service-ca/service-ca-9c57cc56f-w62xw" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.556581 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fb2e8401-8c07-4083-8b0f-821cd21b5858-apiservice-cert\") pod \"packageserver-d55dfcdfc-bs5t5\" (UID: \"fb2e8401-8c07-4083-8b0f-821cd21b5858\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.556732 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/06530996-dd63-4ed5-adb3-790ecc93d11e-stats-auth\") pod \"router-default-5444994796-45dcj\" (UID: \"06530996-dd63-4ed5-adb3-790ecc93d11e\") " pod="openshift-ingress/router-default-5444994796-45dcj" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.556782 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/abf46edd-34b9-42e7-94ca-a3066047cb53-mountpoint-dir\") pod \"csi-hostpathplugin-g4bjk\" (UID: \"abf46edd-34b9-42e7-94ca-a3066047cb53\") " pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.556866 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/abf46edd-34b9-42e7-94ca-a3066047cb53-registration-dir\") pod \"csi-hostpathplugin-g4bjk\" (UID: \"abf46edd-34b9-42e7-94ca-a3066047cb53\") " pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.556963 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f1e5e774-51ea-47e8-8b02-1bde1728da35-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2vsl7\" (UID: \"f1e5e774-51ea-47e8-8b02-1bde1728da35\") " pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.557008 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/26e7d7f8-eccb-479a-bd88-69e526049c18-certs\") pod \"machine-config-server-zbsq2\" (UID: \"26e7d7f8-eccb-479a-bd88-69e526049c18\") " pod="openshift-machine-config-operator/machine-config-server-zbsq2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.557138 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/06530996-dd63-4ed5-adb3-790ecc93d11e-metrics-certs\") pod \"router-default-5444994796-45dcj\" (UID: \"06530996-dd63-4ed5-adb3-790ecc93d11e\") " pod="openshift-ingress/router-default-5444994796-45dcj" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.557237 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vt5bm\" (UniqueName: \"kubernetes.io/projected/06530996-dd63-4ed5-adb3-790ecc93d11e-kube-api-access-vt5bm\") pod \"router-default-5444994796-45dcj\" (UID: \"06530996-dd63-4ed5-adb3-790ecc93d11e\") " pod="openshift-ingress/router-default-5444994796-45dcj" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.557418 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj595\" (UniqueName: \"kubernetes.io/projected/c71045e6-eb17-417d-b9fe-12d3a21f4a56-kube-api-access-gj595\") pod \"service-ca-operator-777779d784-4qsp7\" (UID: \"c71045e6-eb17-417d-b9fe-12d3a21f4a56\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qsp7" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.557470 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpp5v\" (UniqueName: \"kubernetes.io/projected/36e16640-f72c-4e52-9c0d-61d37c1f91d8-kube-api-access-dpp5v\") pod \"ingress-canary-8kdj9\" (UID: \"36e16640-f72c-4e52-9c0d-61d37c1f91d8\") " pod="openshift-ingress-canary/ingress-canary-8kdj9" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.557526 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/53fb45a4-5f96-4e9c-a231-7e840bd5c79f-srv-cert\") pod \"catalog-operator-68c6474976-xbr6t\" (UID: \"53fb45a4-5f96-4e9c-a231-7e840bd5c79f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xbr6t" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.557572 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mn9b\" (UniqueName: \"kubernetes.io/projected/c3dad121-17b2-4c92-a99f-34439a678fa9-kube-api-access-8mn9b\") pod \"dns-default-k2shv\" (UID: \"c3dad121-17b2-4c92-a99f-34439a678fa9\") " pod="openshift-dns/dns-default-k2shv" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.557620 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2692q\" (UniqueName: \"kubernetes.io/projected/26e7d7f8-eccb-479a-bd88-69e526049c18-kube-api-access-2692q\") pod \"machine-config-server-zbsq2\" (UID: \"26e7d7f8-eccb-479a-bd88-69e526049c18\") " pod="openshift-machine-config-operator/machine-config-server-zbsq2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.557664 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/06530996-dd63-4ed5-adb3-790ecc93d11e-default-certificate\") pod \"router-default-5444994796-45dcj\" (UID: \"06530996-dd63-4ed5-adb3-790ecc93d11e\") " pod="openshift-ingress/router-default-5444994796-45dcj" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.557695 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06530996-dd63-4ed5-adb3-790ecc93d11e-service-ca-bundle\") pod \"router-default-5444994796-45dcj\" (UID: \"06530996-dd63-4ed5-adb3-790ecc93d11e\") " pod="openshift-ingress/router-default-5444994796-45dcj" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.557733 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/02a5e11b-33ad-4c1f-8073-2558ba723cd3-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7kxvg\" (UID: \"02a5e11b-33ad-4c1f-8073-2558ba723cd3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7kxvg" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.557776 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfn5j\" (UniqueName: \"kubernetes.io/projected/6714d5a2-a6be-4841-8268-66de19d238ec-kube-api-access-sfn5j\") pod \"olm-operator-6b444d44fb-tnhkb\" (UID: \"6714d5a2-a6be-4841-8268-66de19d238ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tnhkb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.557823 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/26e7d7f8-eccb-479a-bd88-69e526049c18-node-bootstrap-token\") pod \"machine-config-server-zbsq2\" (UID: \"26e7d7f8-eccb-479a-bd88-69e526049c18\") " pod="openshift-machine-config-operator/machine-config-server-zbsq2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.557859 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c3dad121-17b2-4c92-a99f-34439a678fa9-config-volume\") pod \"dns-default-k2shv\" (UID: \"c3dad121-17b2-4c92-a99f-34439a678fa9\") " pod="openshift-dns/dns-default-k2shv" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.557904 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f1e5e774-51ea-47e8-8b02-1bde1728da35-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2vsl7\" (UID: \"f1e5e774-51ea-47e8-8b02-1bde1728da35\") " pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.558903 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/40ba7989-d314-4273-b2e1-aaf05d4b50a4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-z5kzz\" (UID: \"40ba7989-d314-4273-b2e1-aaf05d4b50a4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z5kzz" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.559599 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4e7b24da-83fa-446d-b2a9-97bf382ff2c5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7z6k\" (UID: \"4e7b24da-83fa-446d-b2a9-97bf382ff2c5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7z6k" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.559916 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/00486d12-48b6-4c9b-b844-78ce031ff80e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-mq7cp\" (UID: \"00486d12-48b6-4c9b-b844-78ce031ff80e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mq7cp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.559968 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6714d5a2-a6be-4841-8268-66de19d238ec-profile-collector-cert\") pod \"olm-operator-6b444d44fb-tnhkb\" (UID: \"6714d5a2-a6be-4841-8268-66de19d238ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tnhkb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.560434 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c69cf39a-28f6-491f-8563-6063ffef11aa-auth-proxy-config\") pod \"machine-config-operator-74547568cd-p5mt6\" (UID: \"c69cf39a-28f6-491f-8563-6063ffef11aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p5mt6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.560415 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c3dad121-17b2-4c92-a99f-34439a678fa9-metrics-tls\") pod \"dns-default-k2shv\" (UID: \"c3dad121-17b2-4c92-a99f-34439a678fa9\") " pod="openshift-dns/dns-default-k2shv" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.560854 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c71045e6-eb17-417d-b9fe-12d3a21f4a56-config\") pod \"service-ca-operator-777779d784-4qsp7\" (UID: \"c71045e6-eb17-417d-b9fe-12d3a21f4a56\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qsp7" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.561439 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6714d5a2-a6be-4841-8268-66de19d238ec-srv-cert\") pod \"olm-operator-6b444d44fb-tnhkb\" (UID: \"6714d5a2-a6be-4841-8268-66de19d238ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tnhkb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.561582 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/645b66a3-584a-4791-8558-64b73b38e96b-config\") pod \"kube-controller-manager-operator-78b949d7b-rf9kz\" (UID: \"645b66a3-584a-4791-8558-64b73b38e96b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rf9kz" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.561633 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dea338f1-f870-484b-ac43-7da254eac47a-config-volume\") pod \"collect-profiles-29401935-hhtv9\" (UID: \"dea338f1-f870-484b-ac43-7da254eac47a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9" Nov 26 00:26:00 crc kubenswrapper[4766]: E1126 00:26:00.561925 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:01.061906703 +0000 UTC m=+141.910677253 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.561989 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc2ff29c-9c87-44f9-95f9-7043db18ffe3-config\") pod \"kube-apiserver-operator-766d6c64bb-5thz5\" (UID: \"bc2ff29c-9c87-44f9-95f9-7043db18ffe3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5thz5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.563020 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/53fb45a4-5f96-4e9c-a231-7e840bd5c79f-profile-collector-cert\") pod \"catalog-operator-68c6474976-xbr6t\" (UID: \"53fb45a4-5f96-4e9c-a231-7e840bd5c79f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xbr6t" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.563549 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c69cf39a-28f6-491f-8563-6063ffef11aa-proxy-tls\") pod \"machine-config-operator-74547568cd-p5mt6\" (UID: \"c69cf39a-28f6-491f-8563-6063ffef11aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p5mt6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.563923 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/36e16640-f72c-4e52-9c0d-61d37c1f91d8-cert\") pod \"ingress-canary-8kdj9\" (UID: \"36e16640-f72c-4e52-9c0d-61d37c1f91d8\") " pod="openshift-ingress-canary/ingress-canary-8kdj9" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.564272 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c71045e6-eb17-417d-b9fe-12d3a21f4a56-serving-cert\") pod \"service-ca-operator-777779d784-4qsp7\" (UID: \"c71045e6-eb17-417d-b9fe-12d3a21f4a56\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qsp7" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.564502 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/abf46edd-34b9-42e7-94ca-a3066047cb53-socket-dir\") pod \"csi-hostpathplugin-g4bjk\" (UID: \"abf46edd-34b9-42e7-94ca-a3066047cb53\") " pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.564597 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/02a5e11b-33ad-4c1f-8073-2558ba723cd3-trusted-ca\") pod \"ingress-operator-5b745b69d9-7kxvg\" (UID: \"02a5e11b-33ad-4c1f-8073-2558ba723cd3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7kxvg" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.564893 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e7b24da-83fa-446d-b2a9-97bf382ff2c5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7z6k\" (UID: \"4e7b24da-83fa-446d-b2a9-97bf382ff2c5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7z6k" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.565050 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fb2e8401-8c07-4083-8b0f-821cd21b5858-apiservice-cert\") pod \"packageserver-d55dfcdfc-bs5t5\" (UID: \"fb2e8401-8c07-4083-8b0f-821cd21b5858\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.567000 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/28602273-98a9-4649-80f3-4e8848b16ee3-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-vvlgv\" (UID: \"28602273-98a9-4649-80f3-4e8848b16ee3\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vvlgv" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.567329 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/53fb45a4-5f96-4e9c-a231-7e840bd5c79f-srv-cert\") pod \"catalog-operator-68c6474976-xbr6t\" (UID: \"53fb45a4-5f96-4e9c-a231-7e840bd5c79f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xbr6t" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.567550 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dea338f1-f870-484b-ac43-7da254eac47a-secret-volume\") pod \"collect-profiles-29401935-hhtv9\" (UID: \"dea338f1-f870-484b-ac43-7da254eac47a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.567705 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/02a5e11b-33ad-4c1f-8073-2558ba723cd3-metrics-tls\") pod \"ingress-operator-5b745b69d9-7kxvg\" (UID: \"02a5e11b-33ad-4c1f-8073-2558ba723cd3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7kxvg" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.567838 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/abf46edd-34b9-42e7-94ca-a3066047cb53-plugins-dir\") pod \"csi-hostpathplugin-g4bjk\" (UID: \"abf46edd-34b9-42e7-94ca-a3066047cb53\") " pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.567961 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/abf46edd-34b9-42e7-94ca-a3066047cb53-registration-dir\") pod \"csi-hostpathplugin-g4bjk\" (UID: \"abf46edd-34b9-42e7-94ca-a3066047cb53\") " pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.568005 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/abf46edd-34b9-42e7-94ca-a3066047cb53-mountpoint-dir\") pod \"csi-hostpathplugin-g4bjk\" (UID: \"abf46edd-34b9-42e7-94ca-a3066047cb53\") " pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.568419 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06530996-dd63-4ed5-adb3-790ecc93d11e-service-ca-bundle\") pod \"router-default-5444994796-45dcj\" (UID: \"06530996-dd63-4ed5-adb3-790ecc93d11e\") " pod="openshift-ingress/router-default-5444994796-45dcj" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.569281 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/752819e1-d0ad-4820-979a-bf6ebfc59b4d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-pm5j5\" (UID: \"752819e1-d0ad-4820-979a-bf6ebfc59b4d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pm5j5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.570011 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c3dad121-17b2-4c92-a99f-34439a678fa9-config-volume\") pod \"dns-default-k2shv\" (UID: \"c3dad121-17b2-4c92-a99f-34439a678fa9\") " pod="openshift-dns/dns-default-k2shv" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.570572 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f1e5e774-51ea-47e8-8b02-1bde1728da35-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2vsl7\" (UID: \"f1e5e774-51ea-47e8-8b02-1bde1728da35\") " pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.570623 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7196e744-19f7-4295-9d97-20ddfe1c44b8-signing-cabundle\") pod \"service-ca-9c57cc56f-w62xw\" (UID: \"7196e744-19f7-4295-9d97-20ddfe1c44b8\") " pod="openshift-service-ca/service-ca-9c57cc56f-w62xw" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.570728 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c69cf39a-28f6-491f-8563-6063ffef11aa-images\") pod \"machine-config-operator-74547568cd-p5mt6\" (UID: \"c69cf39a-28f6-491f-8563-6063ffef11aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p5mt6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.571042 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f1e5e774-51ea-47e8-8b02-1bde1728da35-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2vsl7\" (UID: \"f1e5e774-51ea-47e8-8b02-1bde1728da35\") " pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.571260 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fb2e8401-8c07-4083-8b0f-821cd21b5858-webhook-cert\") pod \"packageserver-d55dfcdfc-bs5t5\" (UID: \"fb2e8401-8c07-4083-8b0f-821cd21b5858\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.571339 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/26e7d7f8-eccb-479a-bd88-69e526049c18-node-bootstrap-token\") pod \"machine-config-server-zbsq2\" (UID: \"26e7d7f8-eccb-479a-bd88-69e526049c18\") " pod="openshift-machine-config-operator/machine-config-server-zbsq2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.571760 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/00486d12-48b6-4c9b-b844-78ce031ff80e-proxy-tls\") pod \"machine-config-controller-84d6567774-mq7cp\" (UID: \"00486d12-48b6-4c9b-b844-78ce031ff80e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mq7cp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.572055 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc2ff29c-9c87-44f9-95f9-7043db18ffe3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5thz5\" (UID: \"bc2ff29c-9c87-44f9-95f9-7043db18ffe3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5thz5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.572162 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/26e7d7f8-eccb-479a-bd88-69e526049c18-certs\") pod \"machine-config-server-zbsq2\" (UID: \"26e7d7f8-eccb-479a-bd88-69e526049c18\") " pod="openshift-machine-config-operator/machine-config-server-zbsq2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.572836 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7196e744-19f7-4295-9d97-20ddfe1c44b8-signing-key\") pod \"service-ca-9c57cc56f-w62xw\" (UID: \"7196e744-19f7-4295-9d97-20ddfe1c44b8\") " pod="openshift-service-ca/service-ca-9c57cc56f-w62xw" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.573154 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/645b66a3-584a-4791-8558-64b73b38e96b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-rf9kz\" (UID: \"645b66a3-584a-4791-8558-64b73b38e96b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rf9kz" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.573423 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/06530996-dd63-4ed5-adb3-790ecc93d11e-metrics-certs\") pod \"router-default-5444994796-45dcj\" (UID: \"06530996-dd63-4ed5-adb3-790ecc93d11e\") " pod="openshift-ingress/router-default-5444994796-45dcj" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.574469 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvt7w\" (UniqueName: \"kubernetes.io/projected/1e392e29-fdb8-450d-ac24-ba91e802139a-kube-api-access-mvt7w\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.591421 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gml6x\" (UniqueName: \"kubernetes.io/projected/284847a7-94c0-4f03-a54d-6d22fe77353f-kube-api-access-gml6x\") pod \"image-pruner-29401920-82g4w\" (UID: \"284847a7-94c0-4f03-a54d-6d22fe77353f\") " pod="openshift-image-registry/image-pruner-29401920-82g4w" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.613728 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4ngl\" (UniqueName: \"kubernetes.io/projected/cb710ab4-0d33-4e91-b749-68169fc723c2-kube-api-access-c4ngl\") pod \"machine-approver-56656f9798-bqmvb\" (UID: \"cb710ab4-0d33-4e91-b749-68169fc723c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bqmvb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.642378 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgsrx\" (UniqueName: \"kubernetes.io/projected/b3321717-6860-48e4-91a4-99245add293d-kube-api-access-tgsrx\") pod \"cluster-image-registry-operator-dc59b4c8b-n4vmx\" (UID: \"b3321717-6860-48e4-91a4-99245add293d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4vmx" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.658944 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:00 crc kubenswrapper[4766]: E1126 00:26:00.659114 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:01.159092576 +0000 UTC m=+142.007863006 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.659358 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcv6t\" (UniqueName: \"kubernetes.io/projected/6f7e21ec-878c-4f90-9f5f-51e95d71e373-kube-api-access-jcv6t\") pod \"openshift-apiserver-operator-796bbdcf4f-zndxf\" (UID: \"6f7e21ec-878c-4f90-9f5f-51e95d71e373\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zndxf" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.659446 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: E1126 00:26:00.659805 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:01.159794043 +0000 UTC m=+142.008564473 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.692499 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk9v8\" (UniqueName: \"kubernetes.io/projected/c8260737-caa6-42f1-8854-ad2b97a1ee42-kube-api-access-mk9v8\") pod \"machine-api-operator-5694c8668f-x2vlb\" (UID: \"c8260737-caa6-42f1-8854-ad2b97a1ee42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x2vlb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.696274 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29401920-82g4w" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.711205 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67wjp\" (UniqueName: \"kubernetes.io/projected/d94c7fca-2c24-4005-9839-0578e1b39283-kube-api-access-67wjp\") pod \"downloads-7954f5f757-bz6r8\" (UID: \"d94c7fca-2c24-4005-9839-0578e1b39283\") " pod="openshift-console/downloads-7954f5f757-bz6r8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.714011 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-bz6r8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.723335 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4vmx" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.734627 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zflng\" (UniqueName: \"kubernetes.io/projected/cc0815f7-09db-4adf-b525-205c8fd8e64e-kube-api-access-zflng\") pod \"openshift-config-operator-7777fb866f-59zt8\" (UID: \"cc0815f7-09db-4adf-b525-205c8fd8e64e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-59zt8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.744760 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bnkxg" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.752341 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.753486 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbztm\" (UniqueName: \"kubernetes.io/projected/c91e0334-6829-4a8f-9239-3930e84e65c4-kube-api-access-sbztm\") pod \"oauth-openshift-558db77b4-tgrk8\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.761642 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:00 crc kubenswrapper[4766]: E1126 00:26:00.761920 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:01.261885487 +0000 UTC m=+142.110655927 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.762537 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: E1126 00:26:00.763110 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:01.263085886 +0000 UTC m=+142.111856356 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.768099 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.773403 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rj29g" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.776583 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kb9mk\" (UniqueName: \"kubernetes.io/projected/20bb15a2-ef87-40cf-b450-f2a50a290325-kube-api-access-kb9mk\") pod \"dns-operator-744455d44c-z22jf\" (UID: \"20bb15a2-ef87-40cf-b450-f2a50a290325\") " pod="openshift-dns-operator/dns-operator-744455d44c-z22jf" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.802910 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx6tx\" (UniqueName: \"kubernetes.io/projected/734b50e7-a98e-4999-bab8-ae2547f08c0c-kube-api-access-wx6tx\") pod \"console-operator-58897d9998-cbzmc\" (UID: \"734b50e7-a98e-4999-bab8-ae2547f08c0c\") " pod="openshift-console-operator/console-operator-58897d9998-cbzmc" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.830567 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js6sh\" (UniqueName: \"kubernetes.io/projected/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-kube-api-access-js6sh\") pod \"console-f9d7485db-hqj86\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.839525 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-x2vlb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.839802 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1e392e29-fdb8-450d-ac24-ba91e802139a-bound-sa-token\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.846380 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" event={"ID":"11353a4e-daac-46a8-97de-160ca1272ea1","Type":"ContainerStarted","Data":"e00f921f23b52bbf8d611b60ce9e6e8a80c5adcdd53c5e87d49f2696f163fe82"} Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.860350 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjp9n\" (UniqueName: \"kubernetes.io/projected/28b3189b-79c5-4da2-9ce5-d067f66a47c2-kube-api-access-pjp9n\") pod \"apiserver-76f77b778f-8zjx6\" (UID: \"28b3189b-79c5-4da2-9ce5-d067f66a47c2\") " pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.863884 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:00 crc kubenswrapper[4766]: E1126 00:26:00.863995 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:01.36397505 +0000 UTC m=+142.212745480 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.864537 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:00 crc kubenswrapper[4766]: E1126 00:26:00.864838 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:01.364828231 +0000 UTC m=+142.213598661 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.880208 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zndxf" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.898369 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9z2xg\" (UniqueName: \"kubernetes.io/projected/45bb08cb-d545-4568-9a64-780628c87786-kube-api-access-9z2xg\") pod \"authentication-operator-69f744f599-blhhp\" (UID: \"45bb08cb-d545-4568-9a64-780628c87786\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-blhhp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.900336 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk7cf\" (UniqueName: \"kubernetes.io/projected/9c103959-be49-4b6a-9689-0a58451bc2b7-kube-api-access-qk7cf\") pod \"cluster-samples-operator-665b6dd947-8sgwh\" (UID: \"9c103959-be49-4b6a-9689-0a58451bc2b7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8sgwh" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.912187 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bqmvb" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.913380 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc2ff29c-9c87-44f9-95f9-7043db18ffe3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5thz5\" (UID: \"bc2ff29c-9c87-44f9-95f9-7043db18ffe3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5thz5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.941918 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29401920-82g4w"] Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.948170 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7np2\" (UniqueName: \"kubernetes.io/projected/c69cf39a-28f6-491f-8563-6063ffef11aa-kube-api-access-b7np2\") pod \"machine-config-operator-74547568cd-p5mt6\" (UID: \"c69cf39a-28f6-491f-8563-6063ffef11aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p5mt6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.956078 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.956197 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.964348 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-59zt8" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.966471 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:00 crc kubenswrapper[4766]: E1126 00:26:00.966975 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:01.466959065 +0000 UTC m=+142.315729495 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.972970 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.982044 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bz8z\" (UniqueName: \"kubernetes.io/projected/00486d12-48b6-4c9b-b844-78ce031ff80e-kube-api-access-4bz8z\") pod \"machine-config-controller-84d6567774-mq7cp\" (UID: \"00486d12-48b6-4c9b-b844-78ce031ff80e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mq7cp" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.982523 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz52x\" (UniqueName: \"kubernetes.io/projected/02a5e11b-33ad-4c1f-8073-2558ba723cd3-kube-api-access-pz52x\") pod \"ingress-operator-5b745b69d9-7kxvg\" (UID: \"02a5e11b-33ad-4c1f-8073-2558ba723cd3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7kxvg" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.991582 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88bpw\" (UniqueName: \"kubernetes.io/projected/fb2e8401-8c07-4083-8b0f-821cd21b5858-kube-api-access-88bpw\") pod \"packageserver-d55dfcdfc-bs5t5\" (UID: \"fb2e8401-8c07-4083-8b0f-821cd21b5858\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5" Nov 26 00:26:00 crc kubenswrapper[4766]: I1126 00:26:00.996782 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8sgwh" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.006300 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.016782 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44q2z\" (UniqueName: \"kubernetes.io/projected/53fb45a4-5f96-4e9c-a231-7e840bd5c79f-kube-api-access-44q2z\") pod \"catalog-operator-68c6474976-xbr6t\" (UID: \"53fb45a4-5f96-4e9c-a231-7e840bd5c79f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xbr6t" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.032061 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-cbzmc" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.044602 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4vzv\" (UniqueName: \"kubernetes.io/projected/dea338f1-f870-484b-ac43-7da254eac47a-kube-api-access-k4vzv\") pod \"collect-profiles-29401935-hhtv9\" (UID: \"dea338f1-f870-484b-ac43-7da254eac47a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.070991 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-z22jf" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.071792 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:01 crc kubenswrapper[4766]: E1126 00:26:01.072169 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:01.572155075 +0000 UTC m=+142.420925505 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.087135 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xbr6t" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.094716 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5thz5" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.097258 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d69vj\" (UniqueName: \"kubernetes.io/projected/5fe4c6f0-f223-41fe-adaf-58434d4f66d8-kube-api-access-d69vj\") pod \"migrator-59844c95c7-dt2nb\" (UID: \"5fe4c6f0-f223-41fe-adaf-58434d4f66d8\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dt2nb" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.106012 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mq7cp" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.109491 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5jmt\" (UniqueName: \"kubernetes.io/projected/40ba7989-d314-4273-b2e1-aaf05d4b50a4-kube-api-access-p5jmt\") pod \"multus-admission-controller-857f4d67dd-z5kzz\" (UID: \"40ba7989-d314-4273-b2e1-aaf05d4b50a4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z5kzz" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.111952 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p5mt6" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.136611 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dt2nb" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.141479 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62hpk\" (UniqueName: \"kubernetes.io/projected/752819e1-d0ad-4820-979a-bf6ebfc59b4d-kube-api-access-62hpk\") pod \"control-plane-machine-set-operator-78cbb6b69f-pm5j5\" (UID: \"752819e1-d0ad-4820-979a-bf6ebfc59b4d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pm5j5" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.146327 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-blhhp" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.149311 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvqdb\" (UniqueName: \"kubernetes.io/projected/28602273-98a9-4649-80f3-4e8848b16ee3-kube-api-access-kvqdb\") pod \"package-server-manager-789f6589d5-vvlgv\" (UID: \"28602273-98a9-4649-80f3-4e8848b16ee3\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vvlgv" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.152395 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwtf5\" (UniqueName: \"kubernetes.io/projected/abf46edd-34b9-42e7-94ca-a3066047cb53-kube-api-access-bwtf5\") pod \"csi-hostpathplugin-g4bjk\" (UID: \"abf46edd-34b9-42e7-94ca-a3066047cb53\") " pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.155698 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.155816 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/02a5e11b-33ad-4c1f-8073-2558ba723cd3-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7kxvg\" (UID: \"02a5e11b-33ad-4c1f-8073-2558ba723cd3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7kxvg" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.169295 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-z5kzz" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.172488 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:01 crc kubenswrapper[4766]: E1126 00:26:01.172935 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:01.672920005 +0000 UTC m=+142.521690435 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.174979 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vvlgv" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.181100 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/645b66a3-584a-4791-8558-64b73b38e96b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-rf9kz\" (UID: \"645b66a3-584a-4791-8558-64b73b38e96b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rf9kz" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.182574 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pm5j5" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.198193 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mn9b\" (UniqueName: \"kubernetes.io/projected/c3dad121-17b2-4c92-a99f-34439a678fa9-kube-api-access-8mn9b\") pod \"dns-default-k2shv\" (UID: \"c3dad121-17b2-4c92-a99f-34439a678fa9\") " pod="openshift-dns/dns-default-k2shv" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.200095 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-bz6r8"] Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.205907 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.212005 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rf9kz" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.212890 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2692q\" (UniqueName: \"kubernetes.io/projected/26e7d7f8-eccb-479a-bd88-69e526049c18-kube-api-access-2692q\") pod \"machine-config-server-zbsq2\" (UID: \"26e7d7f8-eccb-479a-bd88-69e526049c18\") " pod="openshift-machine-config-operator/machine-config-server-zbsq2" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.215134 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4vmx"] Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.227230 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-zbsq2" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.235441 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj595\" (UniqueName: \"kubernetes.io/projected/c71045e6-eb17-417d-b9fe-12d3a21f4a56-kube-api-access-gj595\") pod \"service-ca-operator-777779d784-4qsp7\" (UID: \"c71045e6-eb17-417d-b9fe-12d3a21f4a56\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qsp7" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.253483 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.257492 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vt5bm\" (UniqueName: \"kubernetes.io/projected/06530996-dd63-4ed5-adb3-790ecc93d11e-kube-api-access-vt5bm\") pod \"router-default-5444994796-45dcj\" (UID: \"06530996-dd63-4ed5-adb3-790ecc93d11e\") " pod="openshift-ingress/router-default-5444994796-45dcj" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.258449 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-k2shv" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.274760 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:01 crc kubenswrapper[4766]: E1126 00:26:01.275188 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:01.775166973 +0000 UTC m=+142.623937433 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.278431 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpp5v\" (UniqueName: \"kubernetes.io/projected/36e16640-f72c-4e52-9c0d-61d37c1f91d8-kube-api-access-dpp5v\") pod \"ingress-canary-8kdj9\" (UID: \"36e16640-f72c-4e52-9c0d-61d37c1f91d8\") " pod="openshift-ingress-canary/ingress-canary-8kdj9" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.300029 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4xmb\" (UniqueName: \"kubernetes.io/projected/4e7b24da-83fa-446d-b2a9-97bf382ff2c5-kube-api-access-v4xmb\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7z6k\" (UID: \"4e7b24da-83fa-446d-b2a9-97bf382ff2c5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7z6k" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.323506 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfpd2\" (UniqueName: \"kubernetes.io/projected/7196e744-19f7-4295-9d97-20ddfe1c44b8-kube-api-access-qfpd2\") pod \"service-ca-9c57cc56f-w62xw\" (UID: \"7196e744-19f7-4295-9d97-20ddfe1c44b8\") " pod="openshift-service-ca/service-ca-9c57cc56f-w62xw" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.344678 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8whf2"] Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.351448 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bds6d\" (UniqueName: \"kubernetes.io/projected/f1e5e774-51ea-47e8-8b02-1bde1728da35-kube-api-access-bds6d\") pod \"marketplace-operator-79b997595-2vsl7\" (UID: \"f1e5e774-51ea-47e8-8b02-1bde1728da35\") " pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.372038 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfn5j\" (UniqueName: \"kubernetes.io/projected/6714d5a2-a6be-4841-8268-66de19d238ec-kube-api-access-sfn5j\") pod \"olm-operator-6b444d44fb-tnhkb\" (UID: \"6714d5a2-a6be-4841-8268-66de19d238ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tnhkb" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.375863 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:01 crc kubenswrapper[4766]: E1126 00:26:01.376230 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:01.87621292 +0000 UTC m=+142.724983340 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.380412 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7z6k" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.395006 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rj29g"] Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.415377 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bnkxg"] Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.417952 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tnhkb" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.428216 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7kxvg" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.447542 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-45dcj" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.464090 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.477131 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:01 crc kubenswrapper[4766]: E1126 00:26:01.477532 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:01.977512944 +0000 UTC m=+142.826283374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.494042 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qsp7" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.497364 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-w62xw" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.518585 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-8kdj9" Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.567575 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mjvdt"] Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.576717 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-x2vlb"] Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.579074 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:01 crc kubenswrapper[4766]: E1126 00:26:01.579422 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:02.079408082 +0000 UTC m=+142.928178512 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.592726 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zndxf"] Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.679987 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:01 crc kubenswrapper[4766]: E1126 00:26:01.680303 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:02.180291796 +0000 UTC m=+143.029062226 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.709225 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" podStartSLOduration=117.709210778 podStartE2EDuration="1m57.709210778s" podCreationTimestamp="2025-11-26 00:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:01.708449819 +0000 UTC m=+142.557220249" watchObservedRunningTime="2025-11-26 00:26:01.709210778 +0000 UTC m=+142.557981208" Nov 26 00:26:01 crc kubenswrapper[4766]: W1126 00:26:01.736282 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd94c7fca_2c24_4005_9839_0578e1b39283.slice/crio-ed08ca9ff51a245ab68d45c68725aee8c2aa5cd205b9d0fd32913b705ab7f913 WatchSource:0}: Error finding container ed08ca9ff51a245ab68d45c68725aee8c2aa5cd205b9d0fd32913b705ab7f913: Status 404 returned error can't find the container with id ed08ca9ff51a245ab68d45c68725aee8c2aa5cd205b9d0fd32913b705ab7f913 Nov 26 00:26:01 crc kubenswrapper[4766]: W1126 00:26:01.748845 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3321717_6860_48e4_91a4_99245add293d.slice/crio-9e08968fc69cfc9a080ebaed6dc537a130db9b2d1d0b3a78f95897170f1022bc WatchSource:0}: Error finding container 9e08968fc69cfc9a080ebaed6dc537a130db9b2d1d0b3a78f95897170f1022bc: Status 404 returned error can't find the container with id 9e08968fc69cfc9a080ebaed6dc537a130db9b2d1d0b3a78f95897170f1022bc Nov 26 00:26:01 crc kubenswrapper[4766]: W1126 00:26:01.750416 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64bb8da2_3238_49d0_9cb2_b582701fb35d.slice/crio-8fecf3d1b254bff4bf1315a8c25a0eaa082b67782d1df47304d7a3c0a73985c1 WatchSource:0}: Error finding container 8fecf3d1b254bff4bf1315a8c25a0eaa082b67782d1df47304d7a3c0a73985c1: Status 404 returned error can't find the container with id 8fecf3d1b254bff4bf1315a8c25a0eaa082b67782d1df47304d7a3c0a73985c1 Nov 26 00:26:01 crc kubenswrapper[4766]: W1126 00:26:01.752770 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00869536_c66a_4fae_91e7_0c6d271ab8cd.slice/crio-6b31f5a91d23744f6898ba40560ae2e785bd208581990e2b546e7cd82b06cd1c WatchSource:0}: Error finding container 6b31f5a91d23744f6898ba40560ae2e785bd208581990e2b546e7cd82b06cd1c: Status 404 returned error can't find the container with id 6b31f5a91d23744f6898ba40560ae2e785bd208581990e2b546e7cd82b06cd1c Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.757506 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-59zt8"] Nov 26 00:26:01 crc kubenswrapper[4766]: W1126 00:26:01.762983 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode00b34f6_3577_4658_b819_43ada470fbe7.slice/crio-69c88bef98aa1b04fd478c4d250c3f1b6c37c90f3d76e6c8866d8bd4edbdc54f WatchSource:0}: Error finding container 69c88bef98aa1b04fd478c4d250c3f1b6c37c90f3d76e6c8866d8bd4edbdc54f: Status 404 returned error can't find the container with id 69c88bef98aa1b04fd478c4d250c3f1b6c37c90f3d76e6c8866d8bd4edbdc54f Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.769275 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-hqj86"] Nov 26 00:26:01 crc kubenswrapper[4766]: W1126 00:26:01.772289 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9aaa727c_565e_4819_801d_9f1787118d20.slice/crio-456addb0ffb294f97e1e8e7b98ebaa82a4171c9db8aa4baf100f0e30643e4309 WatchSource:0}: Error finding container 456addb0ffb294f97e1e8e7b98ebaa82a4171c9db8aa4baf100f0e30643e4309: Status 404 returned error can't find the container with id 456addb0ffb294f97e1e8e7b98ebaa82a4171c9db8aa4baf100f0e30643e4309 Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.783450 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:01 crc kubenswrapper[4766]: E1126 00:26:01.783631 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:02.283597149 +0000 UTC m=+143.132367599 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.784020 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:01 crc kubenswrapper[4766]: E1126 00:26:01.784432 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:02.284416229 +0000 UTC m=+143.133186679 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.814877 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xbr6t"] Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.818344 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-tgrk8"] Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.818772 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8zjx6"] Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.885155 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:01 crc kubenswrapper[4766]: E1126 00:26:01.885307 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:02.385284452 +0000 UTC m=+143.234054872 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.885498 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:01 crc kubenswrapper[4766]: E1126 00:26:01.885779 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:02.385765144 +0000 UTC m=+143.234535574 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.897900 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8sgwh"] Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.897945 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-cbzmc"] Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.904551 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29401920-82g4w" event={"ID":"284847a7-94c0-4f03-a54d-6d22fe77353f","Type":"ContainerStarted","Data":"33242ffa5a061764e610c7855f70ed79b90fb80b3a9d9300ca24769ab2d50bfb"} Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.914306 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rj29g" event={"ID":"00869536-c66a-4fae-91e7-0c6d271ab8cd","Type":"ContainerStarted","Data":"6b31f5a91d23744f6898ba40560ae2e785bd208581990e2b546e7cd82b06cd1c"} Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.915135 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" event={"ID":"64bb8da2-3238-49d0-9cb2-b582701fb35d","Type":"ContainerStarted","Data":"8fecf3d1b254bff4bf1315a8c25a0eaa082b67782d1df47304d7a3c0a73985c1"} Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.915726 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-bz6r8" event={"ID":"d94c7fca-2c24-4005-9839-0578e1b39283","Type":"ContainerStarted","Data":"ed08ca9ff51a245ab68d45c68725aee8c2aa5cd205b9d0fd32913b705ab7f913"} Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.916315 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zndxf" event={"ID":"6f7e21ec-878c-4f90-9f5f-51e95d71e373","Type":"ContainerStarted","Data":"bccbad381ee445e7d736d19a25b57a956fa237b8558ce94c62561491e381d225"} Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.918371 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-x2vlb" event={"ID":"c8260737-caa6-42f1-8854-ad2b97a1ee42","Type":"ContainerStarted","Data":"914a8a8a0efa1f65e70e0037767b84d173f9a51d3496b7e7b0d6289c4eff821d"} Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.919993 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bnkxg" event={"ID":"e00b34f6-3577-4658-b819-43ada470fbe7","Type":"ContainerStarted","Data":"69c88bef98aa1b04fd478c4d250c3f1b6c37c90f3d76e6c8866d8bd4edbdc54f"} Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.920621 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bqmvb" event={"ID":"cb710ab4-0d33-4e91-b749-68169fc723c2","Type":"ContainerStarted","Data":"b058eb817aec9ab76fad463b3362854882b064699a739c982b1b809dbe5adb24"} Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.921688 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4vmx" event={"ID":"b3321717-6860-48e4-91a4-99245add293d","Type":"ContainerStarted","Data":"9e08968fc69cfc9a080ebaed6dc537a130db9b2d1d0b3a78f95897170f1022bc"} Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.923361 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" event={"ID":"9aaa727c-565e-4819-801d-9f1787118d20","Type":"ContainerStarted","Data":"456addb0ffb294f97e1e8e7b98ebaa82a4171c9db8aa4baf100f0e30643e4309"} Nov 26 00:26:01 crc kubenswrapper[4766]: W1126 00:26:01.937801 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4d6938a_1f6b_4b90_8c85_4616c4def2e5.slice/crio-015c21f4b5a4534086c0aacb2fae8d63deaa0912687654f8ff65c424ecc17f3a WatchSource:0}: Error finding container 015c21f4b5a4534086c0aacb2fae8d63deaa0912687654f8ff65c424ecc17f3a: Status 404 returned error can't find the container with id 015c21f4b5a4534086c0aacb2fae8d63deaa0912687654f8ff65c424ecc17f3a Nov 26 00:26:01 crc kubenswrapper[4766]: W1126 00:26:01.938104 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc0815f7_09db_4adf_b525_205c8fd8e64e.slice/crio-e21d687eadc305b9ac95c6ecb7b20043af70c1b37a88c3120f504d4541099154 WatchSource:0}: Error finding container e21d687eadc305b9ac95c6ecb7b20043af70c1b37a88c3120f504d4541099154: Status 404 returned error can't find the container with id e21d687eadc305b9ac95c6ecb7b20043af70c1b37a88c3120f504d4541099154 Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.980491 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-blhhp"] Nov 26 00:26:01 crc kubenswrapper[4766]: I1126 00:26:01.987146 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:01 crc kubenswrapper[4766]: E1126 00:26:01.989697 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:02.489673192 +0000 UTC m=+143.338443672 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.014488 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-z22jf"] Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.088756 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:02 crc kubenswrapper[4766]: E1126 00:26:02.089112 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:02.58909987 +0000 UTC m=+143.437870290 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.164961 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5thz5"] Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.191081 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:02 crc kubenswrapper[4766]: E1126 00:26:02.191481 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:02.69145847 +0000 UTC m=+143.540228900 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.196027 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5"] Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.217475 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-dt2nb"] Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.249580 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-mq7cp"] Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.292418 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:02 crc kubenswrapper[4766]: E1126 00:26:02.292744 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:02.792732093 +0000 UTC m=+143.641502513 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.299022 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pm5j5"] Nov 26 00:26:02 crc kubenswrapper[4766]: W1126 00:26:02.318219 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20bb15a2_ef87_40cf_b450_f2a50a290325.slice/crio-3314556868ff7096b6a7895f9ad2a10d3b33fe86fae620d78270e83fb5562c47 WatchSource:0}: Error finding container 3314556868ff7096b6a7895f9ad2a10d3b33fe86fae620d78270e83fb5562c47: Status 404 returned error can't find the container with id 3314556868ff7096b6a7895f9ad2a10d3b33fe86fae620d78270e83fb5562c47 Nov 26 00:26:02 crc kubenswrapper[4766]: W1126 00:26:02.352704 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb2e8401_8c07_4083_8b0f_821cd21b5858.slice/crio-f953f759110f4af1d91816b6ac6ed34f00982360d7236794b0824df271e4ab87 WatchSource:0}: Error finding container f953f759110f4af1d91816b6ac6ed34f00982360d7236794b0824df271e4ab87: Status 404 returned error can't find the container with id f953f759110f4af1d91816b6ac6ed34f00982360d7236794b0824df271e4ab87 Nov 26 00:26:02 crc kubenswrapper[4766]: W1126 00:26:02.366614 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe4c6f0_f223_41fe_adaf_58434d4f66d8.slice/crio-86fadf983f60d79b582b1552a25e7cc7ea2af20c11ca41554c2cb6ea1827f44a WatchSource:0}: Error finding container 86fadf983f60d79b582b1552a25e7cc7ea2af20c11ca41554c2cb6ea1827f44a: Status 404 returned error can't find the container with id 86fadf983f60d79b582b1552a25e7cc7ea2af20c11ca41554c2cb6ea1827f44a Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.380854 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-p5mt6"] Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.393945 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:02 crc kubenswrapper[4766]: E1126 00:26:02.394118 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:02.894089238 +0000 UTC m=+143.742859668 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.394249 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:02 crc kubenswrapper[4766]: E1126 00:26:02.394526 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:02.894513489 +0000 UTC m=+143.743283919 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.496169 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:02 crc kubenswrapper[4766]: E1126 00:26:02.496781 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:02.996765636 +0000 UTC m=+143.845536066 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.514406 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7z6k"] Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.597591 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:02 crc kubenswrapper[4766]: E1126 00:26:02.598217 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:03.098204503 +0000 UTC m=+143.946974933 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:02 crc kubenswrapper[4766]: W1126 00:26:02.633744 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e7b24da_83fa_446d_b2a9_97bf382ff2c5.slice/crio-05e6ce79e1ea763114b73dc7758fd0378913be86ed26b5cc8523142f7ca9f254 WatchSource:0}: Error finding container 05e6ce79e1ea763114b73dc7758fd0378913be86ed26b5cc8523142f7ca9f254: Status 404 returned error can't find the container with id 05e6ce79e1ea763114b73dc7758fd0378913be86ed26b5cc8523142f7ca9f254 Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.703854 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:02 crc kubenswrapper[4766]: E1126 00:26:02.704102 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:03.204077159 +0000 UTC m=+144.052847589 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.704214 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:02 crc kubenswrapper[4766]: E1126 00:26:02.704627 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:03.204620233 +0000 UTC m=+144.053390663 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.804891 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:02 crc kubenswrapper[4766]: E1126 00:26:02.805044 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:03.305019224 +0000 UTC m=+144.153789654 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.805150 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:02 crc kubenswrapper[4766]: E1126 00:26:02.805481 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:03.305474026 +0000 UTC m=+144.154244456 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.883452 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-g4bjk"] Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.897875 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-z5kzz"] Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.910527 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:02 crc kubenswrapper[4766]: E1126 00:26:02.910870 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:03.41083999 +0000 UTC m=+144.259610460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.911079 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:02 crc kubenswrapper[4766]: E1126 00:26:02.911798 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:03.411777393 +0000 UTC m=+144.260547873 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.927829 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pm5j5" event={"ID":"752819e1-d0ad-4820-979a-bf6ebfc59b4d","Type":"ContainerStarted","Data":"2327dac75b7f92340fe74d1169fe5723679dea68bb483c5b4326ea65a5f9b240"} Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.929063 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5" event={"ID":"fb2e8401-8c07-4083-8b0f-821cd21b5858","Type":"ContainerStarted","Data":"f953f759110f4af1d91816b6ac6ed34f00982360d7236794b0824df271e4ab87"} Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.930009 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" event={"ID":"28b3189b-79c5-4da2-9ce5-d067f66a47c2","Type":"ContainerStarted","Data":"c4837ab305c674043c0cb8e171b07f5c300967feb9f4be24f96e5756e1d50012"} Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.930752 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xbr6t" event={"ID":"53fb45a4-5f96-4e9c-a231-7e840bd5c79f","Type":"ContainerStarted","Data":"e621083315163d5c6111b058e5506f5f6f60d1bd949d0c4f5623704e0414e3bc"} Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.932040 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29401920-82g4w" event={"ID":"284847a7-94c0-4f03-a54d-6d22fe77353f","Type":"ContainerStarted","Data":"c6b8df07f4c9fc0d3940bf06cd60a70cb5b80db504d71e9a09de043504fb206e"} Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.933111 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-cbzmc" event={"ID":"734b50e7-a98e-4999-bab8-ae2547f08c0c","Type":"ContainerStarted","Data":"c64e24228ead2442e6f46e51d85fff2c70f35d0e8fb53cf3921ce30f687289de"} Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.934607 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bqmvb" event={"ID":"cb710ab4-0d33-4e91-b749-68169fc723c2","Type":"ContainerStarted","Data":"13afc66c185f4681fc8e0026535db69c91b3fbcce096e54153cc51a2d6fe6ceb"} Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.935547 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-hqj86" event={"ID":"e4d6938a-1f6b-4b90-8c85-4616c4def2e5","Type":"ContainerStarted","Data":"015c21f4b5a4534086c0aacb2fae8d63deaa0912687654f8ff65c424ecc17f3a"} Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.936980 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-blhhp" event={"ID":"45bb08cb-d545-4568-9a64-780628c87786","Type":"ContainerStarted","Data":"63366b383a7293ed50cabda5673e2e4a7a4add177202537e078e75595a525712"} Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.938667 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dt2nb" event={"ID":"5fe4c6f0-f223-41fe-adaf-58434d4f66d8","Type":"ContainerStarted","Data":"86fadf983f60d79b582b1552a25e7cc7ea2af20c11ca41554c2cb6ea1827f44a"} Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.940039 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-59zt8" event={"ID":"cc0815f7-09db-4adf-b525-205c8fd8e64e","Type":"ContainerStarted","Data":"e21d687eadc305b9ac95c6ecb7b20043af70c1b37a88c3120f504d4541099154"} Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.940874 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mq7cp" event={"ID":"00486d12-48b6-4c9b-b844-78ce031ff80e","Type":"ContainerStarted","Data":"23952811a4fc2bd4eae51535efcc30dcc325a4def391500ddb833c17df273874"} Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.941693 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5thz5" event={"ID":"bc2ff29c-9c87-44f9-95f9-7043db18ffe3","Type":"ContainerStarted","Data":"04088930546b7af7482c89489858bc1f4a36ba7ddaca75019f93a4f2b40fabff"} Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.943275 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-zbsq2" event={"ID":"26e7d7f8-eccb-479a-bd88-69e526049c18","Type":"ContainerStarted","Data":"c61bfb41727c56cc0977aac014373bf5424371e8d12c056caefee1fe730a1096"} Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.944131 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7z6k" event={"ID":"4e7b24da-83fa-446d-b2a9-97bf382ff2c5","Type":"ContainerStarted","Data":"05e6ce79e1ea763114b73dc7758fd0378913be86ed26b5cc8523142f7ca9f254"} Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.944901 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" event={"ID":"c91e0334-6829-4a8f-9239-3930e84e65c4","Type":"ContainerStarted","Data":"34e8e4dbbc1668d50f4d50aa8d3603a15dbeb2018309e8cb329660751898cc06"} Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.945624 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-z22jf" event={"ID":"20bb15a2-ef87-40cf-b450-f2a50a290325","Type":"ContainerStarted","Data":"3314556868ff7096b6a7895f9ad2a10d3b33fe86fae620d78270e83fb5562c47"} Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.946889 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-45dcj" event={"ID":"06530996-dd63-4ed5-adb3-790ecc93d11e","Type":"ContainerStarted","Data":"090f35b391b686d32289146375b4fcd6e251cd6d3fa90c034fadee8dc33ab564"} Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.947562 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p5mt6" event={"ID":"c69cf39a-28f6-491f-8563-6063ffef11aa","Type":"ContainerStarted","Data":"ab5441733d9acef08182b5242dc6d147abe4f5c151b25d770e041f38d7eb472c"} Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.951353 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" podStartSLOduration=118.951341327 podStartE2EDuration="1m58.951341327s" podCreationTimestamp="2025-11-26 00:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:02.950495476 +0000 UTC m=+143.799265906" watchObservedRunningTime="2025-11-26 00:26:02.951341327 +0000 UTC m=+143.800111757" Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.951577 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8sgwh" event={"ID":"9c103959-be49-4b6a-9689-0a58451bc2b7","Type":"ContainerStarted","Data":"5551a6a172ac77130851ce47d99f68bc0595aaa9cf9ad3e072277fd944115811"} Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.993779 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-k2shv"] Nov 26 00:26:02 crc kubenswrapper[4766]: I1126 00:26:02.997364 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tnhkb"] Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.011619 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vvlgv"] Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.012047 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:03 crc kubenswrapper[4766]: E1126 00:26:03.012200 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:03.512166274 +0000 UTC m=+144.360936704 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.012264 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:03 crc kubenswrapper[4766]: E1126 00:26:03.012611 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:03.512598965 +0000 UTC m=+144.361369395 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:03 crc kubenswrapper[4766]: W1126 00:26:03.085141 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podabf46edd_34b9_42e7_94ca_a3066047cb53.slice/crio-e2c6b9b681be675739a6ce9bd9653762c346ea8f97410b3282a3b1b36d6a3aff WatchSource:0}: Error finding container e2c6b9b681be675739a6ce9bd9653762c346ea8f97410b3282a3b1b36d6a3aff: Status 404 returned error can't find the container with id e2c6b9b681be675739a6ce9bd9653762c346ea8f97410b3282a3b1b36d6a3aff Nov 26 00:26:03 crc kubenswrapper[4766]: W1126 00:26:03.104071 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40ba7989_d314_4273_b2e1_aaf05d4b50a4.slice/crio-9913e88638c2fee366b4054a09fdf9616c15c174b59ccae81ffa27ec94f79005 WatchSource:0}: Error finding container 9913e88638c2fee366b4054a09fdf9616c15c174b59ccae81ffa27ec94f79005: Status 404 returned error can't find the container with id 9913e88638c2fee366b4054a09fdf9616c15c174b59ccae81ffa27ec94f79005 Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.113288 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:03 crc kubenswrapper[4766]: E1126 00:26:03.113491 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:03.613467048 +0000 UTC m=+144.462237478 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.113756 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:03 crc kubenswrapper[4766]: W1126 00:26:03.113809 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6714d5a2_a6be_4841_8268_66de19d238ec.slice/crio-3af7e1b0207cc26437eeadd8c0dcc467a09bf53a9139a551c592105d86574d09 WatchSource:0}: Error finding container 3af7e1b0207cc26437eeadd8c0dcc467a09bf53a9139a551c592105d86574d09: Status 404 returned error can't find the container with id 3af7e1b0207cc26437eeadd8c0dcc467a09bf53a9139a551c592105d86574d09 Nov 26 00:26:03 crc kubenswrapper[4766]: W1126 00:26:03.114044 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28602273_98a9_4649_80f3_4e8848b16ee3.slice/crio-00dd91172623bd999bbb2a6f8541be53f6b1fb3ee0d1bc5d47d735b5362010aa WatchSource:0}: Error finding container 00dd91172623bd999bbb2a6f8541be53f6b1fb3ee0d1bc5d47d735b5362010aa: Status 404 returned error can't find the container with id 00dd91172623bd999bbb2a6f8541be53f6b1fb3ee0d1bc5d47d735b5362010aa Nov 26 00:26:03 crc kubenswrapper[4766]: E1126 00:26:03.114112 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:03.614104823 +0000 UTC m=+144.462875253 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.204562 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-w62xw"] Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.206729 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-8kdj9"] Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.208151 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4qsp7"] Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.215005 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:03 crc kubenswrapper[4766]: E1126 00:26:03.215234 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:03.715204832 +0000 UTC m=+144.563975282 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.215312 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:03 crc kubenswrapper[4766]: E1126 00:26:03.215695 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:03.715675894 +0000 UTC m=+144.564446374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.263322 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7kxvg"] Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.265226 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rf9kz"] Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.270238 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2vsl7"] Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.282394 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9"] Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.316193 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:03 crc kubenswrapper[4766]: E1126 00:26:03.316751 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:03.816731862 +0000 UTC m=+144.665502292 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.418204 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:03 crc kubenswrapper[4766]: E1126 00:26:03.418543 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:03.918531768 +0000 UTC m=+144.767302198 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:03 crc kubenswrapper[4766]: W1126 00:26:03.483360 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02a5e11b_33ad_4c1f_8073_2558ba723cd3.slice/crio-1785925b2c25f47dffaf45a37058bdca548e92149c37cc7bdd92baec175559ac WatchSource:0}: Error finding container 1785925b2c25f47dffaf45a37058bdca548e92149c37cc7bdd92baec175559ac: Status 404 returned error can't find the container with id 1785925b2c25f47dffaf45a37058bdca548e92149c37cc7bdd92baec175559ac Nov 26 00:26:03 crc kubenswrapper[4766]: W1126 00:26:03.497140 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod645b66a3_584a_4791_8558_64b73b38e96b.slice/crio-5a62e4a16a932d4825453e954c24efc3ed51f21661c33318ff9addcb35732076 WatchSource:0}: Error finding container 5a62e4a16a932d4825453e954c24efc3ed51f21661c33318ff9addcb35732076: Status 404 returned error can't find the container with id 5a62e4a16a932d4825453e954c24efc3ed51f21661c33318ff9addcb35732076 Nov 26 00:26:03 crc kubenswrapper[4766]: W1126 00:26:03.506066 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1e5e774_51ea_47e8_8b02_1bde1728da35.slice/crio-a4a489bbb6a0d52c03da5c3b0e4e949f975dd481d0e735b14dfd4acfd638ceac WatchSource:0}: Error finding container a4a489bbb6a0d52c03da5c3b0e4e949f975dd481d0e735b14dfd4acfd638ceac: Status 404 returned error can't find the container with id a4a489bbb6a0d52c03da5c3b0e4e949f975dd481d0e735b14dfd4acfd638ceac Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.519584 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:03 crc kubenswrapper[4766]: E1126 00:26:03.519890 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:04.019874943 +0000 UTC m=+144.868645373 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.621910 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:03 crc kubenswrapper[4766]: E1126 00:26:03.622260 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:04.122244273 +0000 UTC m=+144.971014703 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.723427 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:03 crc kubenswrapper[4766]: E1126 00:26:03.724210 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:04.224190123 +0000 UTC m=+145.072960553 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.825877 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:03 crc kubenswrapper[4766]: E1126 00:26:03.826297 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:04.326281256 +0000 UTC m=+145.175051686 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.929403 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:03 crc kubenswrapper[4766]: E1126 00:26:03.929941 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:04.429922666 +0000 UTC m=+145.278693106 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.962894 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4vmx" event={"ID":"b3321717-6860-48e4-91a4-99245add293d","Type":"ContainerStarted","Data":"dd73c92317ff28e4a6c689a1dba979243cf034fa40875f77af3f2602beba642b"} Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.966473 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-cbzmc" event={"ID":"734b50e7-a98e-4999-bab8-ae2547f08c0c","Type":"ContainerStarted","Data":"dbb46a3d428c39ec3d7e281d92d919570f57616c76c91869e3ff703369c6be3b"} Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.966823 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-cbzmc" Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.968780 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-z5kzz" event={"ID":"40ba7989-d314-4273-b2e1-aaf05d4b50a4","Type":"ContainerStarted","Data":"9913e88638c2fee366b4054a09fdf9616c15c174b59ccae81ffa27ec94f79005"} Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.970091 4766 patch_prober.go:28] interesting pod/console-operator-58897d9998-cbzmc container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.970129 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-cbzmc" podUID="734b50e7-a98e-4999-bab8-ae2547f08c0c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.990922 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" event={"ID":"abf46edd-34b9-42e7-94ca-a3066047cb53","Type":"ContainerStarted","Data":"e2c6b9b681be675739a6ce9bd9653762c346ea8f97410b3282a3b1b36d6a3aff"} Nov 26 00:26:03 crc kubenswrapper[4766]: I1126 00:26:03.992369 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-n4vmx" podStartSLOduration=120.992346403 podStartE2EDuration="2m0.992346403s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:03.98652079 +0000 UTC m=+144.835291230" watchObservedRunningTime="2025-11-26 00:26:03.992346403 +0000 UTC m=+144.841116833" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.010134 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-cbzmc" podStartSLOduration=121.010114661 podStartE2EDuration="2m1.010114661s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:04.005030585 +0000 UTC m=+144.853801015" watchObservedRunningTime="2025-11-26 00:26:04.010114661 +0000 UTC m=+144.858885091" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.010806 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8sgwh" event={"ID":"9c103959-be49-4b6a-9689-0a58451bc2b7","Type":"ContainerStarted","Data":"4919ef3230e55e1fd7a3caafb6ff55cfc8552cce7ac6e42215df8a05fb2ec7b7"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.014032 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rf9kz" event={"ID":"645b66a3-584a-4791-8558-64b73b38e96b","Type":"ContainerStarted","Data":"5a62e4a16a932d4825453e954c24efc3ed51f21661c33318ff9addcb35732076"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.017885 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rj29g" event={"ID":"00869536-c66a-4fae-91e7-0c6d271ab8cd","Type":"ContainerStarted","Data":"46c7b6c287abde50bd99553b5a6d3f7c9a4a0fdd8cb2c3556e0f6abbe5f8a77f"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.023694 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-w62xw" event={"ID":"7196e744-19f7-4295-9d97-20ddfe1c44b8","Type":"ContainerStarted","Data":"1e65ac0dd985a2ca8e020d01d376927a5b7b74fe4c58a37116b34d6727b93b89"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.028337 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8kdj9" event={"ID":"36e16640-f72c-4e52-9c0d-61d37c1f91d8","Type":"ContainerStarted","Data":"54f7e868400fa865287bbe6e1333707b8ece6b65c050c2440aba3014aab0ee7c"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.031483 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:04 crc kubenswrapper[4766]: E1126 00:26:04.032647 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:04.532631575 +0000 UTC m=+145.381402005 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.038280 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rj29g" podStartSLOduration=120.038262764 podStartE2EDuration="2m0.038262764s" podCreationTimestamp="2025-11-26 00:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:04.035594708 +0000 UTC m=+144.884365138" watchObservedRunningTime="2025-11-26 00:26:04.038262764 +0000 UTC m=+144.887033194" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.045267 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" event={"ID":"f1e5e774-51ea-47e8-8b02-1bde1728da35","Type":"ContainerStarted","Data":"a4a489bbb6a0d52c03da5c3b0e4e949f975dd481d0e735b14dfd4acfd638ceac"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.054843 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" event={"ID":"64bb8da2-3238-49d0-9cb2-b582701fb35d","Type":"ContainerStarted","Data":"123bf9d45b7a85a8fa9a3cde13ef54b6d706f21446f9ced6a210f6697fab50ed"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.059449 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qsp7" event={"ID":"c71045e6-eb17-417d-b9fe-12d3a21f4a56","Type":"ContainerStarted","Data":"7d60725fcbe882a1a1ec019a56007e77c828d166260715a4c7f5ac7806ec6212"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.064363 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xbr6t" event={"ID":"53fb45a4-5f96-4e9c-a231-7e840bd5c79f","Type":"ContainerStarted","Data":"f007ddbf2b07552d0b392ea2671fd5dfc9bf358f6c87f4d9e440cb8424349b39"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.065537 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xbr6t" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.075026 4766 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-xbr6t container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.075084 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xbr6t" podUID="53fb45a4-5f96-4e9c-a231-7e840bd5c79f" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.081521 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7z6k" event={"ID":"4e7b24da-83fa-446d-b2a9-97bf382ff2c5","Type":"ContainerStarted","Data":"65a52024a8764d3533a838204625926c7c77377c8cbd102108431a9969e76d7d"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.091440 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5" event={"ID":"fb2e8401-8c07-4083-8b0f-821cd21b5858","Type":"ContainerStarted","Data":"bba90b9eae51cd34dcf0fc1b6d3f2557a834bc0d8fc368f8620e9e1b25bcfe36"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.092474 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.101485 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xbr6t" podStartSLOduration=120.10146675 podStartE2EDuration="2m0.10146675s" podCreationTimestamp="2025-11-26 00:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:04.100477755 +0000 UTC m=+144.949248195" watchObservedRunningTime="2025-11-26 00:26:04.10146675 +0000 UTC m=+144.950237180" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.101549 4766 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-bs5t5 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" start-of-body= Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.101613 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5" podUID="fb2e8401-8c07-4083-8b0f-821cd21b5858" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.103037 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-8whf2" podStartSLOduration=121.103030428 podStartE2EDuration="2m1.103030428s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:04.076243479 +0000 UTC m=+144.925013929" watchObservedRunningTime="2025-11-26 00:26:04.103030428 +0000 UTC m=+144.951800858" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.110515 4766 generic.go:334] "Generic (PLEG): container finished" podID="cc0815f7-09db-4adf-b525-205c8fd8e64e" containerID="0fa2948fc125d57f1c30c7d88f41f458a2244bb64b2efe5002a05f851d368850" exitCode=0 Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.110738 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-59zt8" event={"ID":"cc0815f7-09db-4adf-b525-205c8fd8e64e","Type":"ContainerDied","Data":"0fa2948fc125d57f1c30c7d88f41f458a2244bb64b2efe5002a05f851d368850"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.115869 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zndxf" event={"ID":"6f7e21ec-878c-4f90-9f5f-51e95d71e373","Type":"ContainerStarted","Data":"a194e56e070ce85f94533eb6d45d5f016496005e889340d2764ce60b292d907c"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.121707 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-x2vlb" event={"ID":"c8260737-caa6-42f1-8854-ad2b97a1ee42","Type":"ContainerStarted","Data":"91c8ec6d87f810ed7d947456dba1a5f0ff63688f495d2c619b08da8dd1855ff8"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.123870 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7z6k" podStartSLOduration=120.123855581 podStartE2EDuration="2m0.123855581s" podCreationTimestamp="2025-11-26 00:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:04.122953288 +0000 UTC m=+144.971723728" watchObservedRunningTime="2025-11-26 00:26:04.123855581 +0000 UTC m=+144.972626011" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.128286 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" event={"ID":"9aaa727c-565e-4819-801d-9f1787118d20","Type":"ContainerStarted","Data":"04db4c3f77a9f02e92b110390dbb87ad59837534850bb190c39f1178091e033e"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.129393 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.132166 4766 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-mjvdt container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.132203 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" podUID="9aaa727c-565e-4819-801d-9f1787118d20" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.134120 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:04 crc kubenswrapper[4766]: E1126 00:26:04.137091 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:04.637070186 +0000 UTC m=+145.485840616 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.147728 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5" podStartSLOduration=120.147703608 podStartE2EDuration="2m0.147703608s" podCreationTimestamp="2025-11-26 00:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:04.146520889 +0000 UTC m=+144.995291319" watchObservedRunningTime="2025-11-26 00:26:04.147703608 +0000 UTC m=+144.996474038" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.170260 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" podStartSLOduration=121.170239493 podStartE2EDuration="2m1.170239493s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:04.168072579 +0000 UTC m=+145.016843009" watchObservedRunningTime="2025-11-26 00:26:04.170239493 +0000 UTC m=+145.019009923" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.194066 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p5mt6" event={"ID":"c69cf39a-28f6-491f-8563-6063ffef11aa","Type":"ContainerStarted","Data":"5d936e5b375cf414142afc2e885248ab489ea27daa57be49b6e50b00cfe474d4"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.205198 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zndxf" podStartSLOduration=121.205171823 podStartE2EDuration="2m1.205171823s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:04.199240847 +0000 UTC m=+145.048011277" watchObservedRunningTime="2025-11-26 00:26:04.205171823 +0000 UTC m=+145.053942263" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.221798 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-bz6r8" event={"ID":"d94c7fca-2c24-4005-9839-0578e1b39283","Type":"ContainerStarted","Data":"c33d730a566d6b6fa01330e47127b32017077bf91309b436721bea77a0e0c833"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.222272 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-bz6r8" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.227637 4766 patch_prober.go:28] interesting pod/downloads-7954f5f757-bz6r8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.227720 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bz6r8" podUID="d94c7fca-2c24-4005-9839-0578e1b39283" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.239697 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:04 crc kubenswrapper[4766]: E1126 00:26:04.241209 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:04.741195529 +0000 UTC m=+145.589965949 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.241749 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-hqj86" event={"ID":"e4d6938a-1f6b-4b90-8c85-4616c4def2e5","Type":"ContainerStarted","Data":"ad32a7197fd193373afa7a36a53334c699a74cad5295cb9b7674cfaa44b97768"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.244878 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-bz6r8" podStartSLOduration=121.24485901 podStartE2EDuration="2m1.24485901s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:04.243904306 +0000 UTC m=+145.092674736" watchObservedRunningTime="2025-11-26 00:26:04.24485901 +0000 UTC m=+145.093629440" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.249758 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mq7cp" event={"ID":"00486d12-48b6-4c9b-b844-78ce031ff80e","Type":"ContainerStarted","Data":"6ccf72d9280970bc3e0e2297442a631cc5d53fd8017144c9f9a1f33e0a59e5f3"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.253547 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-blhhp" event={"ID":"45bb08cb-d545-4568-9a64-780628c87786","Type":"ContainerStarted","Data":"30f9794f596b6da4525055df63ecd79cb091af07f84ac195bedab56e949c2bff"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.259016 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pm5j5" event={"ID":"752819e1-d0ad-4820-979a-bf6ebfc59b4d","Type":"ContainerStarted","Data":"2773d5d7777d431f636b3af03b0dbb947e0d5f365c206a714d99b2e8b2472a35"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.263303 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9" event={"ID":"dea338f1-f870-484b-ac43-7da254eac47a","Type":"ContainerStarted","Data":"00ed048d8a19eff954d1780d38adc209fb6f3f39901431c063811524ef864ae2"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.271685 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tnhkb" event={"ID":"6714d5a2-a6be-4841-8268-66de19d238ec","Type":"ContainerStarted","Data":"3af7e1b0207cc26437eeadd8c0dcc467a09bf53a9139a551c592105d86574d09"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.272570 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tnhkb" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.274291 4766 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-tnhkb container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.274352 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tnhkb" podUID="6714d5a2-a6be-4841-8268-66de19d238ec" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.277500 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vvlgv" event={"ID":"28602273-98a9-4649-80f3-4e8848b16ee3","Type":"ContainerStarted","Data":"00dd91172623bd999bbb2a6f8541be53f6b1fb3ee0d1bc5d47d735b5362010aa"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.277735 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-hqj86" podStartSLOduration=121.277715928 podStartE2EDuration="2m1.277715928s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:04.267447636 +0000 UTC m=+145.116218076" watchObservedRunningTime="2025-11-26 00:26:04.277715928 +0000 UTC m=+145.126486358" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.285480 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dt2nb" event={"ID":"5fe4c6f0-f223-41fe-adaf-58434d4f66d8","Type":"ContainerStarted","Data":"3b2384ba3ff7286bb8f28bdc6b8e87aa5b9ec262596802382d5068546f125cd9"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.287997 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-k2shv" event={"ID":"c3dad121-17b2-4c92-a99f-34439a678fa9","Type":"ContainerStarted","Data":"4cfe6ec89a62f6bcbbbfa9663b6e9278bf8ff6380b5d78380ac418085f8530e4"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.290161 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7kxvg" event={"ID":"02a5e11b-33ad-4c1f-8073-2558ba723cd3","Type":"ContainerStarted","Data":"1785925b2c25f47dffaf45a37058bdca548e92149c37cc7bdd92baec175559ac"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.299612 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-blhhp" podStartSLOduration=121.299593317 podStartE2EDuration="2m1.299593317s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:04.291608641 +0000 UTC m=+145.140379071" watchObservedRunningTime="2025-11-26 00:26:04.299593317 +0000 UTC m=+145.148363737" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.310503 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.310969 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.321161 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pm5j5" podStartSLOduration=120.321142578 podStartE2EDuration="2m0.321142578s" podCreationTimestamp="2025-11-26 00:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:04.319920807 +0000 UTC m=+145.168691247" watchObservedRunningTime="2025-11-26 00:26:04.321142578 +0000 UTC m=+145.169913008" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.329319 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.331500 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bnkxg" event={"ID":"e00b34f6-3577-4658-b819-43ada470fbe7","Type":"ContainerStarted","Data":"5b60f3573a223081813485a629fa3837ce702edef6dea696c366f8a7bb5185fc"} Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.348236 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.351846 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tnhkb" podStartSLOduration=120.351811333 podStartE2EDuration="2m0.351811333s" podCreationTimestamp="2025-11-26 00:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:04.334897886 +0000 UTC m=+145.183668316" watchObservedRunningTime="2025-11-26 00:26:04.351811333 +0000 UTC m=+145.200581763" Nov 26 00:26:04 crc kubenswrapper[4766]: E1126 00:26:04.377954 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:04.877922085 +0000 UTC m=+145.726692515 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.378375 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:04 crc kubenswrapper[4766]: E1126 00:26:04.385540 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:04.885523622 +0000 UTC m=+145.734294052 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.391035 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-bnkxg" podStartSLOduration=121.391012248 podStartE2EDuration="2m1.391012248s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:04.378627043 +0000 UTC m=+145.227397493" watchObservedRunningTime="2025-11-26 00:26:04.391012248 +0000 UTC m=+145.239782678" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.427085 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-pruner-29401920-82g4w" podStartSLOduration=121.427069105 podStartE2EDuration="2m1.427069105s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:04.426348948 +0000 UTC m=+145.275119378" watchObservedRunningTime="2025-11-26 00:26:04.427069105 +0000 UTC m=+145.275839525" Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.482219 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:04 crc kubenswrapper[4766]: E1126 00:26:04.484636 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:04.984610332 +0000 UTC m=+145.833380862 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.585226 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:04 crc kubenswrapper[4766]: E1126 00:26:04.585664 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:05.085626779 +0000 UTC m=+145.934397209 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.686376 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:04 crc kubenswrapper[4766]: E1126 00:26:04.686637 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:05.186593764 +0000 UTC m=+146.035364194 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.686879 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:04 crc kubenswrapper[4766]: E1126 00:26:04.687212 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:05.187196989 +0000 UTC m=+146.035967419 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.787849 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:04 crc kubenswrapper[4766]: E1126 00:26:04.788203 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:05.288189515 +0000 UTC m=+146.136959935 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.888974 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:04 crc kubenswrapper[4766]: E1126 00:26:04.889281 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:05.389271114 +0000 UTC m=+146.238041544 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:04 crc kubenswrapper[4766]: I1126 00:26:04.989862 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:04 crc kubenswrapper[4766]: E1126 00:26:04.990191 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:05.490176358 +0000 UTC m=+146.338946788 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.091161 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:05 crc kubenswrapper[4766]: E1126 00:26:05.091583 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:05.591561204 +0000 UTC m=+146.440331714 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.192320 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:05 crc kubenswrapper[4766]: E1126 00:26:05.192722 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:05.692702584 +0000 UTC m=+146.541473014 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.293804 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:05 crc kubenswrapper[4766]: E1126 00:26:05.294322 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:05.794307635 +0000 UTC m=+146.643078065 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.356414 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" event={"ID":"c91e0334-6829-4a8f-9239-3930e84e65c4","Type":"ContainerStarted","Data":"e5dd72bee270578ea299f3e401e5d84d62f7fe4ef40b724b2badeb051bc874b6"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.356865 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.359151 4766 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-tgrk8 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.26:6443/healthz\": dial tcp 10.217.0.26:6443: connect: connection refused" start-of-body= Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.359194 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" podUID="c91e0334-6829-4a8f-9239-3930e84e65c4" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.26:6443/healthz\": dial tcp 10.217.0.26:6443: connect: connection refused" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.361182 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7kxvg" event={"ID":"02a5e11b-33ad-4c1f-8073-2558ba723cd3","Type":"ContainerStarted","Data":"2a840e2dcbedfb12178146f28e6c2424df88339da684ddcf0b00d5d76025b397"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.372662 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qsp7" event={"ID":"c71045e6-eb17-417d-b9fe-12d3a21f4a56","Type":"ContainerStarted","Data":"9d25bffee140ed665527f591dc1c8058eb6ed5d3d161d9f2a74e45dc6a93e1ce"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.394993 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:05 crc kubenswrapper[4766]: E1126 00:26:05.395351 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:05.895332922 +0000 UTC m=+146.744103352 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.403316 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" podStartSLOduration=122.403295408 podStartE2EDuration="2m2.403295408s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:05.395984998 +0000 UTC m=+146.244755428" watchObservedRunningTime="2025-11-26 00:26:05.403295408 +0000 UTC m=+146.252065838" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.418091 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8sgwh" event={"ID":"9c103959-be49-4b6a-9689-0a58451bc2b7","Type":"ContainerStarted","Data":"25be0a44f874d6489e5d70a8d23aa9a403494913a8c463f9ef837e271a5a108e"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.436736 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-45dcj" event={"ID":"06530996-dd63-4ed5-adb3-790ecc93d11e","Type":"ContainerStarted","Data":"e32d606757a691cee02a993b44525bd4f7c6df48d0539d402e339f99e46b35ab"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.446744 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p5mt6" event={"ID":"c69cf39a-28f6-491f-8563-6063ffef11aa","Type":"ContainerStarted","Data":"48dcd7d3571fcfbec19efab44818f6ef733c49be16bb168251fc5329275b2d84"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.448916 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-45dcj" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.449553 4766 patch_prober.go:28] interesting pod/router-default-5444994796-45dcj container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.449599 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-45dcj" podUID="06530996-dd63-4ed5-adb3-790ecc93d11e" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.451787 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vvlgv" event={"ID":"28602273-98a9-4649-80f3-4e8848b16ee3","Type":"ContainerStarted","Data":"4a026a2a63b3cd1f192b99a532251575e04aa4c61cb772c0cf6b81ba908182ed"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.451832 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vvlgv" event={"ID":"28602273-98a9-4649-80f3-4e8848b16ee3","Type":"ContainerStarted","Data":"a216a091633d8a9e49044ca9fe820ba5416478019f0392c0c8a57a6e01986e7d"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.452524 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vvlgv" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.459929 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dt2nb" event={"ID":"5fe4c6f0-f223-41fe-adaf-58434d4f66d8","Type":"ContainerStarted","Data":"e497d7178da5d612a4712b34779e0be34540f4c27f8c2ecddf2f8acecbbc5f81"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.463870 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qsp7" podStartSLOduration=121.463852619 podStartE2EDuration="2m1.463852619s" podCreationTimestamp="2025-11-26 00:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:05.430158429 +0000 UTC m=+146.278928859" watchObservedRunningTime="2025-11-26 00:26:05.463852619 +0000 UTC m=+146.312623049" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.465364 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8sgwh" podStartSLOduration=122.465354926 podStartE2EDuration="2m2.465354926s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:05.461817699 +0000 UTC m=+146.310588129" watchObservedRunningTime="2025-11-26 00:26:05.465354926 +0000 UTC m=+146.314125356" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.474610 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-k2shv" event={"ID":"c3dad121-17b2-4c92-a99f-34439a678fa9","Type":"ContainerStarted","Data":"fcc1140fca9eca109403606a29b1e06661608f44d01434eccf1d171bc180d7f7"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.487757 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vvlgv" podStartSLOduration=121.487739677 podStartE2EDuration="2m1.487739677s" podCreationTimestamp="2025-11-26 00:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:05.487307056 +0000 UTC m=+146.336077486" watchObservedRunningTime="2025-11-26 00:26:05.487739677 +0000 UTC m=+146.336510107" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.487793 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9" event={"ID":"dea338f1-f870-484b-ac43-7da254eac47a","Type":"ContainerStarted","Data":"098f03a0a95f8d40a5860d0c453b6ff77121c9b9e68b9d956cda0c589be117cb"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.496414 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:05 crc kubenswrapper[4766]: E1126 00:26:05.498365 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:05.998347838 +0000 UTC m=+146.847118268 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.502862 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-x2vlb" event={"ID":"c8260737-caa6-42f1-8854-ad2b97a1ee42","Type":"ContainerStarted","Data":"597ab881ad34a055a06ce0feefaf4593ace70b4f087957574ed88ef05f8222bf"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.517062 4766 generic.go:334] "Generic (PLEG): container finished" podID="28b3189b-79c5-4da2-9ce5-d067f66a47c2" containerID="6daf9d6a0c3efc8123e42865b32c45f795a751d22232b4503847e6ee6b5740fc" exitCode=0 Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.517183 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" event={"ID":"28b3189b-79c5-4da2-9ce5-d067f66a47c2","Type":"ContainerDied","Data":"6daf9d6a0c3efc8123e42865b32c45f795a751d22232b4503847e6ee6b5740fc"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.533701 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-zbsq2" event={"ID":"26e7d7f8-eccb-479a-bd88-69e526049c18","Type":"ContainerStarted","Data":"7839700952f0becb0a058041147ab55f28513dbb2a00f5f3a31002f7a107ec11"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.542851 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p5mt6" podStartSLOduration=121.542833823 podStartE2EDuration="2m1.542833823s" podCreationTimestamp="2025-11-26 00:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:05.510910007 +0000 UTC m=+146.359680437" watchObservedRunningTime="2025-11-26 00:26:05.542833823 +0000 UTC m=+146.391604253" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.551069 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-59zt8" event={"ID":"cc0815f7-09db-4adf-b525-205c8fd8e64e","Type":"ContainerStarted","Data":"ecdef39441ec1f61bbb89dd10bc0e796e16f482d0b272b0e43a349283162b5aa"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.551184 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-59zt8" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.556115 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rf9kz" event={"ID":"645b66a3-584a-4791-8558-64b73b38e96b","Type":"ContainerStarted","Data":"96ecddeda6818e69a3008f84ac159c237e4011b47372cceb7c3c7bcc859baade"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.572077 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dt2nb" podStartSLOduration=121.572035662 podStartE2EDuration="2m1.572035662s" podCreationTimestamp="2025-11-26 00:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:05.543194382 +0000 UTC m=+146.391964812" watchObservedRunningTime="2025-11-26 00:26:05.572035662 +0000 UTC m=+146.420806092" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.584501 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5thz5" event={"ID":"bc2ff29c-9c87-44f9-95f9-7043db18ffe3","Type":"ContainerStarted","Data":"1a6c878f599538c710d884a374f0b115caf943ac9af26d854c58fc177932f334"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.603492 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:05 crc kubenswrapper[4766]: E1126 00:26:05.608290 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:06.108270644 +0000 UTC m=+146.957041074 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.608990 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-59zt8" podStartSLOduration=122.608954371 podStartE2EDuration="2m2.608954371s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:05.60202388 +0000 UTC m=+146.450794310" watchObservedRunningTime="2025-11-26 00:26:05.608954371 +0000 UTC m=+146.457724801" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.610354 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-z22jf" event={"ID":"20bb15a2-ef87-40cf-b450-f2a50a290325","Type":"ContainerStarted","Data":"dd71035550e91a11d9238c5709f2072cecfdc8876cc73dac32b43d764cd898ba"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.610477 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-z22jf" event={"ID":"20bb15a2-ef87-40cf-b450-f2a50a290325","Type":"ContainerStarted","Data":"759f9a167207484ae3d5eae7cee1fa5956c307416ac6a483ff341361b79faf8a"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.611336 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-45dcj" podStartSLOduration=122.611318959 podStartE2EDuration="2m2.611318959s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:05.57357312 +0000 UTC m=+146.422343550" watchObservedRunningTime="2025-11-26 00:26:05.611318959 +0000 UTC m=+146.460089389" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.613291 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8kdj9" event={"ID":"36e16640-f72c-4e52-9c0d-61d37c1f91d8","Type":"ContainerStarted","Data":"a27fa10d91536eb935bb94c0fd026dc9107086ca489cfb1865764d04e56554c2"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.617943 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-z5kzz" event={"ID":"40ba7989-d314-4273-b2e1-aaf05d4b50a4","Type":"ContainerStarted","Data":"e932dea2ca23ee85c295e2b3477d16f287d266ceb30b2479f2df1faf1b0401ce"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.633002 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9" podStartSLOduration=122.632982733 podStartE2EDuration="2m2.632982733s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:05.622279369 +0000 UTC m=+146.471049799" watchObservedRunningTime="2025-11-26 00:26:05.632982733 +0000 UTC m=+146.481753163" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.636522 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bqmvb" event={"ID":"cb710ab4-0d33-4e91-b749-68169fc723c2","Type":"ContainerStarted","Data":"013b28554b49feac0ec3e1c59b87a388294df5d0cdbd5e4b8e5220e98ce10940"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.640431 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" event={"ID":"f1e5e774-51ea-47e8-8b02-1bde1728da35","Type":"ContainerStarted","Data":"11bd5ba6a7c03a5a475d63b231ab4a3a247b539d5b4b9b0fab89e3ee729c88b1"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.640855 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.643188 4766 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2vsl7 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.643388 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" podUID="f1e5e774-51ea-47e8-8b02-1bde1728da35" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.643777 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mq7cp" event={"ID":"00486d12-48b6-4c9b-b844-78ce031ff80e","Type":"ContainerStarted","Data":"dc6994cbb96f7f5bd3393f2f4f95a3764860cd49739fb0f4d92a3bf39a27f69e"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.646252 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tnhkb" event={"ID":"6714d5a2-a6be-4841-8268-66de19d238ec","Type":"ContainerStarted","Data":"c585a14e7bcf4b4bbc129bd72697606110c9118a7f4f3068aa103d2432c185d4"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.647197 4766 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-tnhkb container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.647230 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tnhkb" podUID="6714d5a2-a6be-4841-8268-66de19d238ec" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.650170 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-zbsq2" podStartSLOduration=7.650158175 podStartE2EDuration="7.650158175s" podCreationTimestamp="2025-11-26 00:25:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:05.645950462 +0000 UTC m=+146.494720892" watchObservedRunningTime="2025-11-26 00:26:05.650158175 +0000 UTC m=+146.498928605" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.655370 4766 patch_prober.go:28] interesting pod/downloads-7954f5f757-bz6r8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.655428 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bz6r8" podUID="d94c7fca-2c24-4005-9839-0578e1b39283" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.655495 4766 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-bs5t5 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" start-of-body= Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.655515 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5" podUID="fb2e8401-8c07-4083-8b0f-821cd21b5858" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.655629 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-w62xw" event={"ID":"7196e744-19f7-4295-9d97-20ddfe1c44b8","Type":"ContainerStarted","Data":"91c8c912cefe6db7fd5b5eefa569697f6992591747878a2b384edbf9cdb0ecbe"} Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.656176 4766 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-mjvdt container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.656379 4766 patch_prober.go:28] interesting pod/console-operator-58897d9998-cbzmc container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.656414 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-cbzmc" podUID="734b50e7-a98e-4999-bab8-ae2547f08c0c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.656484 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" podUID="9aaa727c-565e-4819-801d-9f1787118d20" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.663484 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f4gkn" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.667867 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-x2vlb" podStartSLOduration=121.667848781 podStartE2EDuration="2m1.667848781s" podCreationTimestamp="2025-11-26 00:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:05.664098649 +0000 UTC m=+146.512869099" watchObservedRunningTime="2025-11-26 00:26:05.667848781 +0000 UTC m=+146.516619211" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.692851 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xbr6t" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.718745 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:05 crc kubenswrapper[4766]: E1126 00:26:05.720790 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:06.220777424 +0000 UTC m=+147.069547854 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.763822 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rf9kz" podStartSLOduration=121.763806913 podStartE2EDuration="2m1.763806913s" podCreationTimestamp="2025-11-26 00:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:05.730845002 +0000 UTC m=+146.579615432" watchObservedRunningTime="2025-11-26 00:26:05.763806913 +0000 UTC m=+146.612577343" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.764191 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-z22jf" podStartSLOduration=122.764186673 podStartE2EDuration="2m2.764186673s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:05.759922938 +0000 UTC m=+146.608693358" watchObservedRunningTime="2025-11-26 00:26:05.764186673 +0000 UTC m=+146.612957093" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.790547 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mq7cp" podStartSLOduration=121.790528231 podStartE2EDuration="2m1.790528231s" podCreationTimestamp="2025-11-26 00:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:05.78725487 +0000 UTC m=+146.636025300" watchObservedRunningTime="2025-11-26 00:26:05.790528231 +0000 UTC m=+146.639298661" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.810468 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5thz5" podStartSLOduration=121.810446141 podStartE2EDuration="2m1.810446141s" podCreationTimestamp="2025-11-26 00:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:05.807914799 +0000 UTC m=+146.656685239" watchObservedRunningTime="2025-11-26 00:26:05.810446141 +0000 UTC m=+146.659216571" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.819756 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:05 crc kubenswrapper[4766]: E1126 00:26:05.821861 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:06.321840212 +0000 UTC m=+147.170610652 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.850104 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bqmvb" podStartSLOduration=122.850086647 podStartE2EDuration="2m2.850086647s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:05.847620467 +0000 UTC m=+146.696390907" watchObservedRunningTime="2025-11-26 00:26:05.850086647 +0000 UTC m=+146.698857077" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.922503 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:05 crc kubenswrapper[4766]: E1126 00:26:05.922942 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:06.422931111 +0000 UTC m=+147.271701541 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.957738 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-z5kzz" podStartSLOduration=121.957721007 podStartE2EDuration="2m1.957721007s" podCreationTimestamp="2025-11-26 00:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:05.9257681 +0000 UTC m=+146.774538530" watchObservedRunningTime="2025-11-26 00:26:05.957721007 +0000 UTC m=+146.806491437" Nov 26 00:26:05 crc kubenswrapper[4766]: I1126 00:26:05.958130 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" podStartSLOduration=121.958125687 podStartE2EDuration="2m1.958125687s" podCreationTimestamp="2025-11-26 00:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:05.958093206 +0000 UTC m=+146.806863636" watchObservedRunningTime="2025-11-26 00:26:05.958125687 +0000 UTC m=+146.806896107" Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.023543 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:06 crc kubenswrapper[4766]: E1126 00:26:06.024096 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:06.524081561 +0000 UTC m=+147.372851991 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.086059 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-8kdj9" podStartSLOduration=8.086042276 podStartE2EDuration="8.086042276s" podCreationTimestamp="2025-11-26 00:25:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:06.040875434 +0000 UTC m=+146.889645864" watchObservedRunningTime="2025-11-26 00:26:06.086042276 +0000 UTC m=+146.934812706" Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.128066 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:06 crc kubenswrapper[4766]: E1126 00:26:06.128612 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:06.628596184 +0000 UTC m=+147.477366604 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.229356 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:06 crc kubenswrapper[4766]: E1126 00:26:06.229555 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:06.729525698 +0000 UTC m=+147.578296138 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.229749 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:06 crc kubenswrapper[4766]: E1126 00:26:06.230107 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:06.730094682 +0000 UTC m=+147.578865112 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.330994 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:06 crc kubenswrapper[4766]: E1126 00:26:06.331232 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:06.831194401 +0000 UTC m=+147.679964841 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.331484 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:06 crc kubenswrapper[4766]: E1126 00:26:06.331795 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:06.831771685 +0000 UTC m=+147.680552106 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.432702 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:06 crc kubenswrapper[4766]: E1126 00:26:06.432887 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:06.932860484 +0000 UTC m=+147.781630914 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.433301 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:06 crc kubenswrapper[4766]: E1126 00:26:06.433680 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:06.933671454 +0000 UTC m=+147.782441874 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.452798 4766 patch_prober.go:28] interesting pod/router-default-5444994796-45dcj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 00:26:06 crc kubenswrapper[4766]: [-]has-synced failed: reason withheld Nov 26 00:26:06 crc kubenswrapper[4766]: [+]process-running ok Nov 26 00:26:06 crc kubenswrapper[4766]: healthz check failed Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.452854 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-45dcj" podUID="06530996-dd63-4ed5-adb3-790ecc93d11e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.534255 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:06 crc kubenswrapper[4766]: E1126 00:26:06.534509 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:07.034465965 +0000 UTC m=+147.883236395 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.534558 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:06 crc kubenswrapper[4766]: E1126 00:26:06.534887 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:07.034880305 +0000 UTC m=+147.883650725 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.635581 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:06 crc kubenswrapper[4766]: E1126 00:26:06.636403 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:07.136012265 +0000 UTC m=+147.984782695 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.644262 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:06 crc kubenswrapper[4766]: E1126 00:26:06.645680 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:07.145661973 +0000 UTC m=+147.994432403 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.670429 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-k2shv" event={"ID":"c3dad121-17b2-4c92-a99f-34439a678fa9","Type":"ContainerStarted","Data":"d65582728e51826174ad315b0991cfe5bd67c62a0e1df94d5b26ab9741b25d3d"} Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.670500 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-k2shv" Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.684350 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-z5kzz" event={"ID":"40ba7989-d314-4273-b2e1-aaf05d4b50a4","Type":"ContainerStarted","Data":"f3d8cdcf60cf8e732b0be52c004577d27f873951723c8a3326a3880b11db611d"} Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.702579 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-w62xw" podStartSLOduration=122.702565994 podStartE2EDuration="2m2.702565994s" podCreationTimestamp="2025-11-26 00:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:06.086642941 +0000 UTC m=+146.935413381" watchObservedRunningTime="2025-11-26 00:26:06.702565994 +0000 UTC m=+147.551336424" Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.708945 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7kxvg" event={"ID":"02a5e11b-33ad-4c1f-8073-2558ba723cd3","Type":"ContainerStarted","Data":"0179de6772df2b2bf583709a1ded15e8c4cb3e3a1784e2079268d7c0b0e1d939"} Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.738696 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" event={"ID":"28b3189b-79c5-4da2-9ce5-d067f66a47c2","Type":"ContainerStarted","Data":"3413c953141cc44546972c2dba3a6a1efb7d6012a828646caaac26b08b019dcf"} Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.738737 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" event={"ID":"28b3189b-79c5-4da2-9ce5-d067f66a47c2","Type":"ContainerStarted","Data":"edfa733c3a5c47cfcb867df7b446a0e2adb8b239d5673a3c747196e82fe35a08"} Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.745819 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.746100 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.746155 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.746179 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.746200 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:26:06 crc kubenswrapper[4766]: E1126 00:26:06.747080 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:07.247059659 +0000 UTC m=+148.095830089 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.750780 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.756788 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" event={"ID":"abf46edd-34b9-42e7-94ca-a3066047cb53","Type":"ContainerStarted","Data":"58dab614f04d0bb33b2f4bb9b2395997bcc7a65a0c7ad1ae180f70540709cf17"} Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.757253 4766 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2vsl7 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.757321 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" podUID="f1e5e774-51ea-47e8-8b02-1bde1728da35" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.766547 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7kxvg" podStartSLOduration=123.766528348 podStartE2EDuration="2m3.766528348s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:06.764827696 +0000 UTC m=+147.613598156" watchObservedRunningTime="2025-11-26 00:26:06.766528348 +0000 UTC m=+147.615298778" Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.766769 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-k2shv" podStartSLOduration=8.766765784 podStartE2EDuration="8.766765784s" podCreationTimestamp="2025-11-26 00:25:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:06.702332418 +0000 UTC m=+147.551102868" watchObservedRunningTime="2025-11-26 00:26:06.766765784 +0000 UTC m=+147.615536214" Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.816349 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-cbzmc" Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.818964 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tnhkb" Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.824837 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bs5t5" Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.864383 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.865843 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:06 crc kubenswrapper[4766]: E1126 00:26:06.874126 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:07.374110937 +0000 UTC m=+148.222881367 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.875796 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.883257 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.887747 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.921257 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" podStartSLOduration=123.921237017 podStartE2EDuration="2m3.921237017s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:06.819308238 +0000 UTC m=+147.668078668" watchObservedRunningTime="2025-11-26 00:26:06.921237017 +0000 UTC m=+147.770007447" Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.968165 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:06 crc kubenswrapper[4766]: E1126 00:26:06.968344 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:07.468312516 +0000 UTC m=+148.317082946 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:06 crc kubenswrapper[4766]: I1126 00:26:06.968507 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:06 crc kubenswrapper[4766]: E1126 00:26:06.969721 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:07.46971083 +0000 UTC m=+148.318481350 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.051162 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.060608 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.070498 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:07 crc kubenswrapper[4766]: E1126 00:26:07.070868 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:07.57085128 +0000 UTC m=+148.419621710 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.071743 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.171797 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:07 crc kubenswrapper[4766]: E1126 00:26:07.172121 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:07.672107563 +0000 UTC m=+148.520877993 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.181718 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.272273 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:07 crc kubenswrapper[4766]: E1126 00:26:07.272801 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:07.772787721 +0000 UTC m=+148.621558151 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.373516 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:07 crc kubenswrapper[4766]: E1126 00:26:07.373943 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:07.873929161 +0000 UTC m=+148.722699591 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.390164 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bjz9v"] Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.407787 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bjz9v" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.414035 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.420509 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bjz9v"] Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.483932 4766 patch_prober.go:28] interesting pod/router-default-5444994796-45dcj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 00:26:07 crc kubenswrapper[4766]: [-]has-synced failed: reason withheld Nov 26 00:26:07 crc kubenswrapper[4766]: [+]process-running ok Nov 26 00:26:07 crc kubenswrapper[4766]: healthz check failed Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.483993 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-45dcj" podUID="06530996-dd63-4ed5-adb3-790ecc93d11e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.484788 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.485040 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0d4c85c-2a27-4a2f-903e-8f7473351104-catalog-content\") pod \"certified-operators-bjz9v\" (UID: \"e0d4c85c-2a27-4a2f-903e-8f7473351104\") " pod="openshift-marketplace/certified-operators-bjz9v" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.485120 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0d4c85c-2a27-4a2f-903e-8f7473351104-utilities\") pod \"certified-operators-bjz9v\" (UID: \"e0d4c85c-2a27-4a2f-903e-8f7473351104\") " pod="openshift-marketplace/certified-operators-bjz9v" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.485152 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgnbg\" (UniqueName: \"kubernetes.io/projected/e0d4c85c-2a27-4a2f-903e-8f7473351104-kube-api-access-lgnbg\") pod \"certified-operators-bjz9v\" (UID: \"e0d4c85c-2a27-4a2f-903e-8f7473351104\") " pod="openshift-marketplace/certified-operators-bjz9v" Nov 26 00:26:07 crc kubenswrapper[4766]: E1126 00:26:07.485280 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:07.985265861 +0000 UTC m=+148.834036291 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.578208 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jnqdt"] Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.579388 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jnqdt" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.587638 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.588437 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgnbg\" (UniqueName: \"kubernetes.io/projected/e0d4c85c-2a27-4a2f-903e-8f7473351104-kube-api-access-lgnbg\") pod \"certified-operators-bjz9v\" (UID: \"e0d4c85c-2a27-4a2f-903e-8f7473351104\") " pod="openshift-marketplace/certified-operators-bjz9v" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.588478 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.588514 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0d4c85c-2a27-4a2f-903e-8f7473351104-catalog-content\") pod \"certified-operators-bjz9v\" (UID: \"e0d4c85c-2a27-4a2f-903e-8f7473351104\") " pod="openshift-marketplace/certified-operators-bjz9v" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.588573 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0d4c85c-2a27-4a2f-903e-8f7473351104-utilities\") pod \"certified-operators-bjz9v\" (UID: \"e0d4c85c-2a27-4a2f-903e-8f7473351104\") " pod="openshift-marketplace/certified-operators-bjz9v" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.589266 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0d4c85c-2a27-4a2f-903e-8f7473351104-utilities\") pod \"certified-operators-bjz9v\" (UID: \"e0d4c85c-2a27-4a2f-903e-8f7473351104\") " pod="openshift-marketplace/certified-operators-bjz9v" Nov 26 00:26:07 crc kubenswrapper[4766]: E1126 00:26:07.589747 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:08.089735393 +0000 UTC m=+148.938505823 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.590022 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0d4c85c-2a27-4a2f-903e-8f7473351104-catalog-content\") pod \"certified-operators-bjz9v\" (UID: \"e0d4c85c-2a27-4a2f-903e-8f7473351104\") " pod="openshift-marketplace/certified-operators-bjz9v" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.611141 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jnqdt"] Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.626977 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgnbg\" (UniqueName: \"kubernetes.io/projected/e0d4c85c-2a27-4a2f-903e-8f7473351104-kube-api-access-lgnbg\") pod \"certified-operators-bjz9v\" (UID: \"e0d4c85c-2a27-4a2f-903e-8f7473351104\") " pod="openshift-marketplace/certified-operators-bjz9v" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.692373 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.692878 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90830d76-8bef-4c13-b127-23fff3e32770-utilities\") pod \"community-operators-jnqdt\" (UID: \"90830d76-8bef-4c13-b127-23fff3e32770\") " pod="openshift-marketplace/community-operators-jnqdt" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.692948 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90830d76-8bef-4c13-b127-23fff3e32770-catalog-content\") pod \"community-operators-jnqdt\" (UID: \"90830d76-8bef-4c13-b127-23fff3e32770\") " pod="openshift-marketplace/community-operators-jnqdt" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.693008 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bwtw\" (UniqueName: \"kubernetes.io/projected/90830d76-8bef-4c13-b127-23fff3e32770-kube-api-access-9bwtw\") pod \"community-operators-jnqdt\" (UID: \"90830d76-8bef-4c13-b127-23fff3e32770\") " pod="openshift-marketplace/community-operators-jnqdt" Nov 26 00:26:07 crc kubenswrapper[4766]: E1126 00:26:07.693108 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:08.193094007 +0000 UTC m=+149.041864437 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.764884 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5hq26"] Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.766024 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5hq26" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.770897 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bjz9v" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.796887 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90830d76-8bef-4c13-b127-23fff3e32770-utilities\") pod \"community-operators-jnqdt\" (UID: \"90830d76-8bef-4c13-b127-23fff3e32770\") " pod="openshift-marketplace/community-operators-jnqdt" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.796939 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90830d76-8bef-4c13-b127-23fff3e32770-catalog-content\") pod \"community-operators-jnqdt\" (UID: \"90830d76-8bef-4c13-b127-23fff3e32770\") " pod="openshift-marketplace/community-operators-jnqdt" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.796984 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bwtw\" (UniqueName: \"kubernetes.io/projected/90830d76-8bef-4c13-b127-23fff3e32770-kube-api-access-9bwtw\") pod \"community-operators-jnqdt\" (UID: \"90830d76-8bef-4c13-b127-23fff3e32770\") " pod="openshift-marketplace/community-operators-jnqdt" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.797005 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:07 crc kubenswrapper[4766]: E1126 00:26:07.797236 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:08.297226131 +0000 UTC m=+149.145996561 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.797665 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90830d76-8bef-4c13-b127-23fff3e32770-utilities\") pod \"community-operators-jnqdt\" (UID: \"90830d76-8bef-4c13-b127-23fff3e32770\") " pod="openshift-marketplace/community-operators-jnqdt" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.797900 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90830d76-8bef-4c13-b127-23fff3e32770-catalog-content\") pod \"community-operators-jnqdt\" (UID: \"90830d76-8bef-4c13-b127-23fff3e32770\") " pod="openshift-marketplace/community-operators-jnqdt" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.798279 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5hq26"] Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.805887 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"46e744c1a46b9f8b96182c90a2ea28ed679a324a62aea0989c10686178493d9e"} Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.813803 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" event={"ID":"abf46edd-34b9-42e7-94ca-a3066047cb53","Type":"ContainerStarted","Data":"15ae2c9ebeb419b7fbb641a555852c06b94cbadabf02badb2877c48040cb4727"} Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.818254 4766 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2vsl7 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.818302 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" podUID="f1e5e774-51ea-47e8-8b02-1bde1728da35" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.852425 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bwtw\" (UniqueName: \"kubernetes.io/projected/90830d76-8bef-4c13-b127-23fff3e32770-kube-api-access-9bwtw\") pod \"community-operators-jnqdt\" (UID: \"90830d76-8bef-4c13-b127-23fff3e32770\") " pod="openshift-marketplace/community-operators-jnqdt" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.901472 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.902151 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2-catalog-content\") pod \"certified-operators-5hq26\" (UID: \"cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2\") " pod="openshift-marketplace/certified-operators-5hq26" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.902187 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2-utilities\") pod \"certified-operators-5hq26\" (UID: \"cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2\") " pod="openshift-marketplace/certified-operators-5hq26" Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.902271 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9bwg\" (UniqueName: \"kubernetes.io/projected/cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2-kube-api-access-j9bwg\") pod \"certified-operators-5hq26\" (UID: \"cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2\") " pod="openshift-marketplace/certified-operators-5hq26" Nov 26 00:26:07 crc kubenswrapper[4766]: E1126 00:26:07.903107 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:08.403090847 +0000 UTC m=+149.251861277 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:07 crc kubenswrapper[4766]: I1126 00:26:07.945804 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jnqdt" Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:07.999234 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gxwnj"] Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.000104 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxwnj" Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.003877 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2-catalog-content\") pod \"certified-operators-5hq26\" (UID: \"cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2\") " pod="openshift-marketplace/certified-operators-5hq26" Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.003920 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2-utilities\") pod \"certified-operators-5hq26\" (UID: \"cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2\") " pod="openshift-marketplace/certified-operators-5hq26" Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.003961 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.004005 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9bwg\" (UniqueName: \"kubernetes.io/projected/cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2-kube-api-access-j9bwg\") pod \"certified-operators-5hq26\" (UID: \"cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2\") " pod="openshift-marketplace/certified-operators-5hq26" Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.004716 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2-catalog-content\") pod \"certified-operators-5hq26\" (UID: \"cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2\") " pod="openshift-marketplace/certified-operators-5hq26" Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.004953 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2-utilities\") pod \"certified-operators-5hq26\" (UID: \"cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2\") " pod="openshift-marketplace/certified-operators-5hq26" Nov 26 00:26:08 crc kubenswrapper[4766]: E1126 00:26:08.005245 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:08.505231182 +0000 UTC m=+149.354001612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.037088 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9bwg\" (UniqueName: \"kubernetes.io/projected/cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2-kube-api-access-j9bwg\") pod \"certified-operators-5hq26\" (UID: \"cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2\") " pod="openshift-marketplace/certified-operators-5hq26" Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.104803 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.105354 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvmjt\" (UniqueName: \"kubernetes.io/projected/c9808e13-0b77-4493-820a-80a80111d0a0-kube-api-access-vvmjt\") pod \"community-operators-gxwnj\" (UID: \"c9808e13-0b77-4493-820a-80a80111d0a0\") " pod="openshift-marketplace/community-operators-gxwnj" Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.105388 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9808e13-0b77-4493-820a-80a80111d0a0-catalog-content\") pod \"community-operators-gxwnj\" (UID: \"c9808e13-0b77-4493-820a-80a80111d0a0\") " pod="openshift-marketplace/community-operators-gxwnj" Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.105492 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9808e13-0b77-4493-820a-80a80111d0a0-utilities\") pod \"community-operators-gxwnj\" (UID: \"c9808e13-0b77-4493-820a-80a80111d0a0\") " pod="openshift-marketplace/community-operators-gxwnj" Nov 26 00:26:08 crc kubenswrapper[4766]: E1126 00:26:08.105625 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:08.605608053 +0000 UTC m=+149.454378483 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.122787 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5hq26" Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.162476 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gxwnj"] Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.207563 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvmjt\" (UniqueName: \"kubernetes.io/projected/c9808e13-0b77-4493-820a-80a80111d0a0-kube-api-access-vvmjt\") pod \"community-operators-gxwnj\" (UID: \"c9808e13-0b77-4493-820a-80a80111d0a0\") " pod="openshift-marketplace/community-operators-gxwnj" Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.207613 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9808e13-0b77-4493-820a-80a80111d0a0-catalog-content\") pod \"community-operators-gxwnj\" (UID: \"c9808e13-0b77-4493-820a-80a80111d0a0\") " pod="openshift-marketplace/community-operators-gxwnj" Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.207670 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.207754 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9808e13-0b77-4493-820a-80a80111d0a0-utilities\") pod \"community-operators-gxwnj\" (UID: \"c9808e13-0b77-4493-820a-80a80111d0a0\") " pod="openshift-marketplace/community-operators-gxwnj" Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.211855 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9808e13-0b77-4493-820a-80a80111d0a0-utilities\") pod \"community-operators-gxwnj\" (UID: \"c9808e13-0b77-4493-820a-80a80111d0a0\") " pod="openshift-marketplace/community-operators-gxwnj" Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.212071 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9808e13-0b77-4493-820a-80a80111d0a0-catalog-content\") pod \"community-operators-gxwnj\" (UID: \"c9808e13-0b77-4493-820a-80a80111d0a0\") " pod="openshift-marketplace/community-operators-gxwnj" Nov 26 00:26:08 crc kubenswrapper[4766]: E1126 00:26:08.212239 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:08.712221327 +0000 UTC m=+149.560991827 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.267458 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvmjt\" (UniqueName: \"kubernetes.io/projected/c9808e13-0b77-4493-820a-80a80111d0a0-kube-api-access-vvmjt\") pod \"community-operators-gxwnj\" (UID: \"c9808e13-0b77-4493-820a-80a80111d0a0\") " pod="openshift-marketplace/community-operators-gxwnj" Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.308311 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:08 crc kubenswrapper[4766]: E1126 00:26:08.308610 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:08.80859602 +0000 UTC m=+149.657366450 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.409427 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:08 crc kubenswrapper[4766]: E1126 00:26:08.409790 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:08.909775801 +0000 UTC m=+149.758546231 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.443813 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxwnj" Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.460595 4766 patch_prober.go:28] interesting pod/router-default-5444994796-45dcj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 00:26:08 crc kubenswrapper[4766]: [-]has-synced failed: reason withheld Nov 26 00:26:08 crc kubenswrapper[4766]: [+]process-running ok Nov 26 00:26:08 crc kubenswrapper[4766]: healthz check failed Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.460675 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-45dcj" podUID="06530996-dd63-4ed5-adb3-790ecc93d11e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.512309 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:08 crc kubenswrapper[4766]: E1126 00:26:08.512878 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:09.012862259 +0000 UTC m=+149.861632689 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.615226 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:08 crc kubenswrapper[4766]: E1126 00:26:08.615687 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:09.115667749 +0000 UTC m=+149.964438189 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.716148 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:08 crc kubenswrapper[4766]: E1126 00:26:08.716500 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:09.216487042 +0000 UTC m=+150.065257472 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.777574 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bjz9v"] Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.817715 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:08 crc kubenswrapper[4766]: E1126 00:26:08.818137 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:09.318122604 +0000 UTC m=+150.166893034 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.869911 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"276e6a06d3ad32bbef2848190d8e72d6eeac22dbeaef421ded785534e05abae7"} Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.885203 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jnqdt"] Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.893900 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"0daf0200a34b0cc2891f1130a66ce1bdcafa86ae102d72207db7dad2e6ca33a5"} Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.895003 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.906188 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2e511b1a5d43a4bfd1d1a4b2c26272ac8f8915bb201f433f1f16d664753a75aa"} Nov 26 00:26:08 crc kubenswrapper[4766]: W1126 00:26:08.918161 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90830d76_8bef_4c13_b127_23fff3e32770.slice/crio-8807f756df08ca1317bcb38ca901f5cc3389d1ef3280a62a8384a2c4ac3e7ba1 WatchSource:0}: Error finding container 8807f756df08ca1317bcb38ca901f5cc3389d1ef3280a62a8384a2c4ac3e7ba1: Status 404 returned error can't find the container with id 8807f756df08ca1317bcb38ca901f5cc3389d1ef3280a62a8384a2c4ac3e7ba1 Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.919328 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:08 crc kubenswrapper[4766]: E1126 00:26:08.919708 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:09.419686854 +0000 UTC m=+150.268457284 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:08 crc kubenswrapper[4766]: I1126 00:26:08.975850 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" event={"ID":"abf46edd-34b9-42e7-94ca-a3066047cb53","Type":"ContainerStarted","Data":"2e7d37c1ca38b1572610ec56a0f6f694b625eeacd10aa50cb2b874331e6b2235"} Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.020369 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:09 crc kubenswrapper[4766]: E1126 00:26:09.021067 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:09.521051459 +0000 UTC m=+150.369821889 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.046992 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gxwnj"] Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.121023 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:09 crc kubenswrapper[4766]: E1126 00:26:09.121379 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:09.621363489 +0000 UTC m=+150.470133909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.175586 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5hq26"] Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.222429 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:09 crc kubenswrapper[4766]: E1126 00:26:09.222866 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:09.722849457 +0000 UTC m=+150.571619887 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.323865 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:09 crc kubenswrapper[4766]: E1126 00:26:09.324166 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:09.824152431 +0000 UTC m=+150.672922861 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.359060 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fhlnl"] Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.360314 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fhlnl" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.361930 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.373828 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fhlnl"] Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.377011 4766 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.425169 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.425225 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1529f474-493a-485a-80b7-ee8ec1c79ccc-utilities\") pod \"redhat-marketplace-fhlnl\" (UID: \"1529f474-493a-485a-80b7-ee8ec1c79ccc\") " pod="openshift-marketplace/redhat-marketplace-fhlnl" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.425295 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1529f474-493a-485a-80b7-ee8ec1c79ccc-catalog-content\") pod \"redhat-marketplace-fhlnl\" (UID: \"1529f474-493a-485a-80b7-ee8ec1c79ccc\") " pod="openshift-marketplace/redhat-marketplace-fhlnl" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.425342 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xktn8\" (UniqueName: \"kubernetes.io/projected/1529f474-493a-485a-80b7-ee8ec1c79ccc-kube-api-access-xktn8\") pod \"redhat-marketplace-fhlnl\" (UID: \"1529f474-493a-485a-80b7-ee8ec1c79ccc\") " pod="openshift-marketplace/redhat-marketplace-fhlnl" Nov 26 00:26:09 crc kubenswrapper[4766]: E1126 00:26:09.425533 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:09.925517346 +0000 UTC m=+150.774287776 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.452136 4766 patch_prober.go:28] interesting pod/router-default-5444994796-45dcj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 00:26:09 crc kubenswrapper[4766]: [-]has-synced failed: reason withheld Nov 26 00:26:09 crc kubenswrapper[4766]: [+]process-running ok Nov 26 00:26:09 crc kubenswrapper[4766]: healthz check failed Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.452195 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-45dcj" podUID="06530996-dd63-4ed5-adb3-790ecc93d11e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.526320 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:09 crc kubenswrapper[4766]: E1126 00:26:09.526524 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:10.026491852 +0000 UTC m=+150.875262292 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.526579 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1529f474-493a-485a-80b7-ee8ec1c79ccc-utilities\") pod \"redhat-marketplace-fhlnl\" (UID: \"1529f474-493a-485a-80b7-ee8ec1c79ccc\") " pod="openshift-marketplace/redhat-marketplace-fhlnl" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.526672 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1529f474-493a-485a-80b7-ee8ec1c79ccc-catalog-content\") pod \"redhat-marketplace-fhlnl\" (UID: \"1529f474-493a-485a-80b7-ee8ec1c79ccc\") " pod="openshift-marketplace/redhat-marketplace-fhlnl" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.526703 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xktn8\" (UniqueName: \"kubernetes.io/projected/1529f474-493a-485a-80b7-ee8ec1c79ccc-kube-api-access-xktn8\") pod \"redhat-marketplace-fhlnl\" (UID: \"1529f474-493a-485a-80b7-ee8ec1c79ccc\") " pod="openshift-marketplace/redhat-marketplace-fhlnl" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.526775 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:09 crc kubenswrapper[4766]: E1126 00:26:09.527102 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:10.027086507 +0000 UTC m=+150.875856937 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.527170 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1529f474-493a-485a-80b7-ee8ec1c79ccc-utilities\") pod \"redhat-marketplace-fhlnl\" (UID: \"1529f474-493a-485a-80b7-ee8ec1c79ccc\") " pod="openshift-marketplace/redhat-marketplace-fhlnl" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.527253 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1529f474-493a-485a-80b7-ee8ec1c79ccc-catalog-content\") pod \"redhat-marketplace-fhlnl\" (UID: \"1529f474-493a-485a-80b7-ee8ec1c79ccc\") " pod="openshift-marketplace/redhat-marketplace-fhlnl" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.576836 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xktn8\" (UniqueName: \"kubernetes.io/projected/1529f474-493a-485a-80b7-ee8ec1c79ccc-kube-api-access-xktn8\") pod \"redhat-marketplace-fhlnl\" (UID: \"1529f474-493a-485a-80b7-ee8ec1c79ccc\") " pod="openshift-marketplace/redhat-marketplace-fhlnl" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.629254 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:09 crc kubenswrapper[4766]: E1126 00:26:09.629877 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:10.129862377 +0000 UTC m=+150.978632807 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.731122 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:09 crc kubenswrapper[4766]: E1126 00:26:09.731475 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:10.231457918 +0000 UTC m=+151.080228338 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.756221 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tqzxf"] Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.757172 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tqzxf" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.774094 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tqzxf"] Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.804231 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fhlnl" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.835429 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.835927 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/519cd16f-3b69-4365-866d-47f01ecd0c74-utilities\") pod \"redhat-marketplace-tqzxf\" (UID: \"519cd16f-3b69-4365-866d-47f01ecd0c74\") " pod="openshift-marketplace/redhat-marketplace-tqzxf" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.835980 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/519cd16f-3b69-4365-866d-47f01ecd0c74-catalog-content\") pod \"redhat-marketplace-tqzxf\" (UID: \"519cd16f-3b69-4365-866d-47f01ecd0c74\") " pod="openshift-marketplace/redhat-marketplace-tqzxf" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.836040 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvfrt\" (UniqueName: \"kubernetes.io/projected/519cd16f-3b69-4365-866d-47f01ecd0c74-kube-api-access-wvfrt\") pod \"redhat-marketplace-tqzxf\" (UID: \"519cd16f-3b69-4365-866d-47f01ecd0c74\") " pod="openshift-marketplace/redhat-marketplace-tqzxf" Nov 26 00:26:09 crc kubenswrapper[4766]: E1126 00:26:09.836192 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 00:26:10.336173946 +0000 UTC m=+151.184944376 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.936827 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/519cd16f-3b69-4365-866d-47f01ecd0c74-utilities\") pod \"redhat-marketplace-tqzxf\" (UID: \"519cd16f-3b69-4365-866d-47f01ecd0c74\") " pod="openshift-marketplace/redhat-marketplace-tqzxf" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.936885 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/519cd16f-3b69-4365-866d-47f01ecd0c74-catalog-content\") pod \"redhat-marketplace-tqzxf\" (UID: \"519cd16f-3b69-4365-866d-47f01ecd0c74\") " pod="openshift-marketplace/redhat-marketplace-tqzxf" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.936909 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.936948 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvfrt\" (UniqueName: \"kubernetes.io/projected/519cd16f-3b69-4365-866d-47f01ecd0c74-kube-api-access-wvfrt\") pod \"redhat-marketplace-tqzxf\" (UID: \"519cd16f-3b69-4365-866d-47f01ecd0c74\") " pod="openshift-marketplace/redhat-marketplace-tqzxf" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.937670 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/519cd16f-3b69-4365-866d-47f01ecd0c74-utilities\") pod \"redhat-marketplace-tqzxf\" (UID: \"519cd16f-3b69-4365-866d-47f01ecd0c74\") " pod="openshift-marketplace/redhat-marketplace-tqzxf" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.937883 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/519cd16f-3b69-4365-866d-47f01ecd0c74-catalog-content\") pod \"redhat-marketplace-tqzxf\" (UID: \"519cd16f-3b69-4365-866d-47f01ecd0c74\") " pod="openshift-marketplace/redhat-marketplace-tqzxf" Nov 26 00:26:09 crc kubenswrapper[4766]: E1126 00:26:09.938165 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 00:26:10.438154046 +0000 UTC m=+151.286924466 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-m8qgq" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.968476 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.969097 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.975347 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.975500 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.977802 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-59zt8" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.984830 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvfrt\" (UniqueName: \"kubernetes.io/projected/519cd16f-3b69-4365-866d-47f01ecd0c74-kube-api-access-wvfrt\") pod \"redhat-marketplace-tqzxf\" (UID: \"519cd16f-3b69-4365-866d-47f01ecd0c74\") " pod="openshift-marketplace/redhat-marketplace-tqzxf" Nov 26 00:26:09 crc kubenswrapper[4766]: I1126 00:26:09.994585 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.000205 4766 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-26T00:26:09.377037383Z","Handler":null,"Name":""} Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.013391 4766 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.013440 4766 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.019543 4766 generic.go:334] "Generic (PLEG): container finished" podID="c9808e13-0b77-4493-820a-80a80111d0a0" containerID="9957c308d5b601bc8c4bd635b7d9417fb0aeeaa0b8b5962d5f5f9aecba4d1058" exitCode=0 Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.019710 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxwnj" event={"ID":"c9808e13-0b77-4493-820a-80a80111d0a0","Type":"ContainerDied","Data":"9957c308d5b601bc8c4bd635b7d9417fb0aeeaa0b8b5962d5f5f9aecba4d1058"} Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.022349 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxwnj" event={"ID":"c9808e13-0b77-4493-820a-80a80111d0a0","Type":"ContainerStarted","Data":"e58376dd6256554d5dd8338928444be6b32d4ab8ff5a9caaa99ec62d32b22677"} Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.024074 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.027018 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"b1eae3eb6e67281ff0ce5db7fe650f130dfbc5c45636c3eac61bc41279f6f1a1"} Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.029778 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" event={"ID":"abf46edd-34b9-42e7-94ca-a3066047cb53","Type":"ContainerStarted","Data":"0903610a3b4e399a3f05c01d5db4992e35ba8c92fd36601ccaa83935312fd1c2"} Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.036280 4766 generic.go:334] "Generic (PLEG): container finished" podID="e0d4c85c-2a27-4a2f-903e-8f7473351104" containerID="9bcbabef2bd02fa775ac02940096c80cb7ef049913d92fa23bbfdd938502f110" exitCode=0 Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.036349 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bjz9v" event={"ID":"e0d4c85c-2a27-4a2f-903e-8f7473351104","Type":"ContainerDied","Data":"9bcbabef2bd02fa775ac02940096c80cb7ef049913d92fa23bbfdd938502f110"} Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.036375 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bjz9v" event={"ID":"e0d4c85c-2a27-4a2f-903e-8f7473351104","Type":"ContainerStarted","Data":"20d598dd11a975425eeaf8b05e42fc9c384e1d1a16ae8d144664055124a3cb6e"} Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.037438 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.037642 4766 generic.go:334] "Generic (PLEG): container finished" podID="90830d76-8bef-4c13-b127-23fff3e32770" containerID="f2da372e0c8d91ca02e5ff2d3b6e62a853dad47179dc42c14d64a58dbe69170a" exitCode=0 Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.037714 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnqdt" event={"ID":"90830d76-8bef-4c13-b127-23fff3e32770","Type":"ContainerDied","Data":"f2da372e0c8d91ca02e5ff2d3b6e62a853dad47179dc42c14d64a58dbe69170a"} Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.037736 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnqdt" event={"ID":"90830d76-8bef-4c13-b127-23fff3e32770","Type":"ContainerStarted","Data":"8807f756df08ca1317bcb38ca901f5cc3389d1ef3280a62a8384a2c4ac3e7ba1"} Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.037799 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5dba56bc-b11d-4f0a-aea4-205c48a40765-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5dba56bc-b11d-4f0a-aea4-205c48a40765\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.037874 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5dba56bc-b11d-4f0a-aea4-205c48a40765-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5dba56bc-b11d-4f0a-aea4-205c48a40765\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.039730 4766 generic.go:334] "Generic (PLEG): container finished" podID="cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2" containerID="da26f486c36b0ad61f384ad1c45e78be590cd00ef690601b7ef7ee2a45548810" exitCode=0 Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.039817 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hq26" event={"ID":"cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2","Type":"ContainerDied","Data":"da26f486c36b0ad61f384ad1c45e78be590cd00ef690601b7ef7ee2a45548810"} Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.039862 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hq26" event={"ID":"cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2","Type":"ContainerStarted","Data":"7b4c8dac3ccba57af9d731811581ed84553ce4e516437a74ba0d170825968250"} Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.049184 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d7ff4217e2c3c9c87c570a794b599a0e6179056243f7dfb42892f1747f6720db"} Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.055284 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.065040 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" podStartSLOduration=12.06502769 podStartE2EDuration="12.06502769s" podCreationTimestamp="2025-11-26 00:25:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:10.064385734 +0000 UTC m=+150.913156174" watchObservedRunningTime="2025-11-26 00:26:10.06502769 +0000 UTC m=+150.913798120" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.071911 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tqzxf" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.145089 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5dba56bc-b11d-4f0a-aea4-205c48a40765-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5dba56bc-b11d-4f0a-aea4-205c48a40765\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.145195 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5dba56bc-b11d-4f0a-aea4-205c48a40765-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5dba56bc-b11d-4f0a-aea4-205c48a40765\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.145233 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.147578 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5dba56bc-b11d-4f0a-aea4-205c48a40765-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5dba56bc-b11d-4f0a-aea4-205c48a40765\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.167602 4766 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.167671 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:10 crc kubenswrapper[4766]: W1126 00:26:10.204822 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1529f474_493a_485a_80b7_ee8ec1c79ccc.slice/crio-ea2f0265f358576113f615350961c4fe6a0457d01dbb85040661ef54f33fe153 WatchSource:0}: Error finding container ea2f0265f358576113f615350961c4fe6a0457d01dbb85040661ef54f33fe153: Status 404 returned error can't find the container with id ea2f0265f358576113f615350961c4fe6a0457d01dbb85040661ef54f33fe153 Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.205505 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5dba56bc-b11d-4f0a-aea4-205c48a40765-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5dba56bc-b11d-4f0a-aea4-205c48a40765\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.206340 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fhlnl"] Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.241635 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-m8qgq\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.314147 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.336620 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tqzxf"] Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.338481 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.451773 4766 patch_prober.go:28] interesting pod/router-default-5444994796-45dcj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 00:26:10 crc kubenswrapper[4766]: [-]has-synced failed: reason withheld Nov 26 00:26:10 crc kubenswrapper[4766]: [+]process-running ok Nov 26 00:26:10 crc kubenswrapper[4766]: healthz check failed Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.452066 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-45dcj" podUID="06530996-dd63-4ed5-adb3-790ecc93d11e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.540469 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.568847 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-s9x5n"] Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.569990 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s9x5n" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.571681 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.571991 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s9x5n"] Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.662365 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2z5vl\" (UniqueName: \"kubernetes.io/projected/5f6c20b5-b168-4529-b84a-bef5c19254d5-kube-api-access-2z5vl\") pod \"redhat-operators-s9x5n\" (UID: \"5f6c20b5-b168-4529-b84a-bef5c19254d5\") " pod="openshift-marketplace/redhat-operators-s9x5n" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.662440 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f6c20b5-b168-4529-b84a-bef5c19254d5-utilities\") pod \"redhat-operators-s9x5n\" (UID: \"5f6c20b5-b168-4529-b84a-bef5c19254d5\") " pod="openshift-marketplace/redhat-operators-s9x5n" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.662629 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f6c20b5-b168-4529-b84a-bef5c19254d5-catalog-content\") pod \"redhat-operators-s9x5n\" (UID: \"5f6c20b5-b168-4529-b84a-bef5c19254d5\") " pod="openshift-marketplace/redhat-operators-s9x5n" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.714934 4766 patch_prober.go:28] interesting pod/downloads-7954f5f757-bz6r8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.715468 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bz6r8" podUID="d94c7fca-2c24-4005-9839-0578e1b39283" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.718676 4766 patch_prober.go:28] interesting pod/downloads-7954f5f757-bz6r8 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.718728 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-bz6r8" podUID="d94c7fca-2c24-4005-9839-0578e1b39283" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.764396 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f6c20b5-b168-4529-b84a-bef5c19254d5-utilities\") pod \"redhat-operators-s9x5n\" (UID: \"5f6c20b5-b168-4529-b84a-bef5c19254d5\") " pod="openshift-marketplace/redhat-operators-s9x5n" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.764472 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f6c20b5-b168-4529-b84a-bef5c19254d5-catalog-content\") pod \"redhat-operators-s9x5n\" (UID: \"5f6c20b5-b168-4529-b84a-bef5c19254d5\") " pod="openshift-marketplace/redhat-operators-s9x5n" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.764555 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2z5vl\" (UniqueName: \"kubernetes.io/projected/5f6c20b5-b168-4529-b84a-bef5c19254d5-kube-api-access-2z5vl\") pod \"redhat-operators-s9x5n\" (UID: \"5f6c20b5-b168-4529-b84a-bef5c19254d5\") " pod="openshift-marketplace/redhat-operators-s9x5n" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.764924 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f6c20b5-b168-4529-b84a-bef5c19254d5-utilities\") pod \"redhat-operators-s9x5n\" (UID: \"5f6c20b5-b168-4529-b84a-bef5c19254d5\") " pod="openshift-marketplace/redhat-operators-s9x5n" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.765309 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f6c20b5-b168-4529-b84a-bef5c19254d5-catalog-content\") pod \"redhat-operators-s9x5n\" (UID: \"5f6c20b5-b168-4529-b84a-bef5c19254d5\") " pod="openshift-marketplace/redhat-operators-s9x5n" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.783274 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2z5vl\" (UniqueName: \"kubernetes.io/projected/5f6c20b5-b168-4529-b84a-bef5c19254d5-kube-api-access-2z5vl\") pod \"redhat-operators-s9x5n\" (UID: \"5f6c20b5-b168-4529-b84a-bef5c19254d5\") " pod="openshift-marketplace/redhat-operators-s9x5n" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.799108 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-m8qgq"] Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.902379 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s9x5n" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.960675 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.960715 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.970380 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-45g8j"] Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.971506 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-45g8j" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.972048 4766 patch_prober.go:28] interesting pod/console-f9d7485db-hqj86 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.972076 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-hqj86" podUID="e4d6938a-1f6b-4b90-8c85-4616c4def2e5" containerName="console" probeResult="failure" output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.974210 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.974244 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.976293 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-45g8j"] Nov 26 00:26:10 crc kubenswrapper[4766]: I1126 00:26:10.987553 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.059150 4766 generic.go:334] "Generic (PLEG): container finished" podID="519cd16f-3b69-4365-866d-47f01ecd0c74" containerID="cdc529ca6da9251692495bb3cb6bd42b87b8e1cfc2fa2f5834ef0d2b50f2760b" exitCode=0 Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.059207 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tqzxf" event={"ID":"519cd16f-3b69-4365-866d-47f01ecd0c74","Type":"ContainerDied","Data":"cdc529ca6da9251692495bb3cb6bd42b87b8e1cfc2fa2f5834ef0d2b50f2760b"} Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.059231 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tqzxf" event={"ID":"519cd16f-3b69-4365-866d-47f01ecd0c74","Type":"ContainerStarted","Data":"ea32ed1709e4b06227212ccb1ab32d85c44dab0b5cecc022cdb8a224d2e58120"} Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.062894 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5dba56bc-b11d-4f0a-aea4-205c48a40765","Type":"ContainerStarted","Data":"ec367e3c330ef234aaf0e957ed4a0833da05b504ea0034c97e35aa5b7e82c138"} Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.067904 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp92b\" (UniqueName: \"kubernetes.io/projected/a8031d42-e714-4bf0-9976-92a118257476-kube-api-access-tp92b\") pod \"redhat-operators-45g8j\" (UID: \"a8031d42-e714-4bf0-9976-92a118257476\") " pod="openshift-marketplace/redhat-operators-45g8j" Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.067972 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8031d42-e714-4bf0-9976-92a118257476-catalog-content\") pod \"redhat-operators-45g8j\" (UID: \"a8031d42-e714-4bf0-9976-92a118257476\") " pod="openshift-marketplace/redhat-operators-45g8j" Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.068047 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8031d42-e714-4bf0-9976-92a118257476-utilities\") pod \"redhat-operators-45g8j\" (UID: \"a8031d42-e714-4bf0-9976-92a118257476\") " pod="openshift-marketplace/redhat-operators-45g8j" Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.073942 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" event={"ID":"1e392e29-fdb8-450d-ac24-ba91e802139a","Type":"ContainerStarted","Data":"58f4d4bc35d4f8a388a6bfb7a6bedc2e77fe9e2a9c077d21e3a31da75826cb7b"} Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.104191 4766 generic.go:334] "Generic (PLEG): container finished" podID="dea338f1-f870-484b-ac43-7da254eac47a" containerID="098f03a0a95f8d40a5860d0c453b6ff77121c9b9e68b9d956cda0c589be117cb" exitCode=0 Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.104320 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9" event={"ID":"dea338f1-f870-484b-ac43-7da254eac47a","Type":"ContainerDied","Data":"098f03a0a95f8d40a5860d0c453b6ff77121c9b9e68b9d956cda0c589be117cb"} Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.107998 4766 generic.go:334] "Generic (PLEG): container finished" podID="1529f474-493a-485a-80b7-ee8ec1c79ccc" containerID="e6a5e794fa2c82f4ac11133d96f64e1cb2b14592dd97dfc020e9863dea5b7ea9" exitCode=0 Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.108383 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fhlnl" event={"ID":"1529f474-493a-485a-80b7-ee8ec1c79ccc","Type":"ContainerDied","Data":"e6a5e794fa2c82f4ac11133d96f64e1cb2b14592dd97dfc020e9863dea5b7ea9"} Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.108589 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fhlnl" event={"ID":"1529f474-493a-485a-80b7-ee8ec1c79ccc","Type":"ContainerStarted","Data":"ea2f0265f358576113f615350961c4fe6a0457d01dbb85040661ef54f33fe153"} Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.115601 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-8zjx6" Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.123468 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.122642645 podStartE2EDuration="2.122642645s" podCreationTimestamp="2025-11-26 00:26:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:11.099836724 +0000 UTC m=+151.948607174" watchObservedRunningTime="2025-11-26 00:26:11.122642645 +0000 UTC m=+151.971413085" Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.170023 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp92b\" (UniqueName: \"kubernetes.io/projected/a8031d42-e714-4bf0-9976-92a118257476-kube-api-access-tp92b\") pod \"redhat-operators-45g8j\" (UID: \"a8031d42-e714-4bf0-9976-92a118257476\") " pod="openshift-marketplace/redhat-operators-45g8j" Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.170166 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8031d42-e714-4bf0-9976-92a118257476-catalog-content\") pod \"redhat-operators-45g8j\" (UID: \"a8031d42-e714-4bf0-9976-92a118257476\") " pod="openshift-marketplace/redhat-operators-45g8j" Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.170240 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8031d42-e714-4bf0-9976-92a118257476-utilities\") pod \"redhat-operators-45g8j\" (UID: \"a8031d42-e714-4bf0-9976-92a118257476\") " pod="openshift-marketplace/redhat-operators-45g8j" Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.171851 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8031d42-e714-4bf0-9976-92a118257476-utilities\") pod \"redhat-operators-45g8j\" (UID: \"a8031d42-e714-4bf0-9976-92a118257476\") " pod="openshift-marketplace/redhat-operators-45g8j" Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.172451 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8031d42-e714-4bf0-9976-92a118257476-catalog-content\") pod \"redhat-operators-45g8j\" (UID: \"a8031d42-e714-4bf0-9976-92a118257476\") " pod="openshift-marketplace/redhat-operators-45g8j" Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.202616 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp92b\" (UniqueName: \"kubernetes.io/projected/a8031d42-e714-4bf0-9976-92a118257476-kube-api-access-tp92b\") pod \"redhat-operators-45g8j\" (UID: \"a8031d42-e714-4bf0-9976-92a118257476\") " pod="openshift-marketplace/redhat-operators-45g8j" Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.333702 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-45g8j" Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.447816 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-45dcj" Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.451078 4766 patch_prober.go:28] interesting pod/router-default-5444994796-45dcj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 00:26:11 crc kubenswrapper[4766]: [-]has-synced failed: reason withheld Nov 26 00:26:11 crc kubenswrapper[4766]: [+]process-running ok Nov 26 00:26:11 crc kubenswrapper[4766]: healthz check failed Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.451155 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-45dcj" podUID="06530996-dd63-4ed5-adb3-790ecc93d11e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.458620 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s9x5n"] Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.469751 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.479491 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.479564 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:26:11 crc kubenswrapper[4766]: W1126 00:26:11.565508 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f6c20b5_b168_4529_b84a_bef5c19254d5.slice/crio-3308cfeb2a0fdbbbf0283eff8f9ea60ac64f9e9e83c0b5391da6560400a0b636 WatchSource:0}: Error finding container 3308cfeb2a0fdbbbf0283eff8f9ea60ac64f9e9e83c0b5391da6560400a0b636: Status 404 returned error can't find the container with id 3308cfeb2a0fdbbbf0283eff8f9ea60ac64f9e9e83c0b5391da6560400a0b636 Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.840848 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 26 00:26:11 crc kubenswrapper[4766]: I1126 00:26:11.899276 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-45g8j"] Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.118385 4766 generic.go:334] "Generic (PLEG): container finished" podID="5dba56bc-b11d-4f0a-aea4-205c48a40765" containerID="d11a065a8e70e9d059aa88db726322adec29c791160e3b79489d36f88020c978" exitCode=0 Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.118445 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5dba56bc-b11d-4f0a-aea4-205c48a40765","Type":"ContainerDied","Data":"d11a065a8e70e9d059aa88db726322adec29c791160e3b79489d36f88020c978"} Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.122537 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" event={"ID":"1e392e29-fdb8-450d-ac24-ba91e802139a","Type":"ContainerStarted","Data":"8bacca62923ee72b2716bbdd37654c255ec6777757d32bdf7e2550d337052e07"} Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.122693 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.143663 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-45g8j" event={"ID":"a8031d42-e714-4bf0-9976-92a118257476","Type":"ContainerStarted","Data":"3731a3f5214a5a7c10cf92e2c05c4232197e7116fd192c64324be6a4b0c73818"} Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.155348 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" podStartSLOduration=129.155319928 podStartE2EDuration="2m9.155319928s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:12.153731189 +0000 UTC m=+153.002501639" watchObservedRunningTime="2025-11-26 00:26:12.155319928 +0000 UTC m=+153.004090348" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.161258 4766 generic.go:334] "Generic (PLEG): container finished" podID="5f6c20b5-b168-4529-b84a-bef5c19254d5" containerID="3e83bc756ab461748f9094ecab67a4fc677a1dcc76993c47e831c2f3fe9f0f43" exitCode=0 Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.161352 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9x5n" event={"ID":"5f6c20b5-b168-4529-b84a-bef5c19254d5","Type":"ContainerDied","Data":"3e83bc756ab461748f9094ecab67a4fc677a1dcc76993c47e831c2f3fe9f0f43"} Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.161410 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9x5n" event={"ID":"5f6c20b5-b168-4529-b84a-bef5c19254d5","Type":"ContainerStarted","Data":"3308cfeb2a0fdbbbf0283eff8f9ea60ac64f9e9e83c0b5391da6560400a0b636"} Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.453169 4766 patch_prober.go:28] interesting pod/router-default-5444994796-45dcj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 00:26:12 crc kubenswrapper[4766]: [-]has-synced failed: reason withheld Nov 26 00:26:12 crc kubenswrapper[4766]: [+]process-running ok Nov 26 00:26:12 crc kubenswrapper[4766]: healthz check failed Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.453226 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-45dcj" podUID="06530996-dd63-4ed5-adb3-790ecc93d11e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.505937 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.554144 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 26 00:26:12 crc kubenswrapper[4766]: E1126 00:26:12.554456 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dea338f1-f870-484b-ac43-7da254eac47a" containerName="collect-profiles" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.554472 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="dea338f1-f870-484b-ac43-7da254eac47a" containerName="collect-profiles" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.554574 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="dea338f1-f870-484b-ac43-7da254eac47a" containerName="collect-profiles" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.555338 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.558191 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.558588 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.558773 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.592776 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dea338f1-f870-484b-ac43-7da254eac47a-config-volume\") pod \"dea338f1-f870-484b-ac43-7da254eac47a\" (UID: \"dea338f1-f870-484b-ac43-7da254eac47a\") " Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.592939 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dea338f1-f870-484b-ac43-7da254eac47a-secret-volume\") pod \"dea338f1-f870-484b-ac43-7da254eac47a\" (UID: \"dea338f1-f870-484b-ac43-7da254eac47a\") " Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.593095 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4vzv\" (UniqueName: \"kubernetes.io/projected/dea338f1-f870-484b-ac43-7da254eac47a-kube-api-access-k4vzv\") pod \"dea338f1-f870-484b-ac43-7da254eac47a\" (UID: \"dea338f1-f870-484b-ac43-7da254eac47a\") " Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.594816 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dea338f1-f870-484b-ac43-7da254eac47a-config-volume" (OuterVolumeSpecName: "config-volume") pod "dea338f1-f870-484b-ac43-7da254eac47a" (UID: "dea338f1-f870-484b-ac43-7da254eac47a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.599309 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea338f1-f870-484b-ac43-7da254eac47a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "dea338f1-f870-484b-ac43-7da254eac47a" (UID: "dea338f1-f870-484b-ac43-7da254eac47a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.599324 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dea338f1-f870-484b-ac43-7da254eac47a-kube-api-access-k4vzv" (OuterVolumeSpecName: "kube-api-access-k4vzv") pod "dea338f1-f870-484b-ac43-7da254eac47a" (UID: "dea338f1-f870-484b-ac43-7da254eac47a"). InnerVolumeSpecName "kube-api-access-k4vzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.694889 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/93c919b9-1485-42fd-a0bd-0608a608f926-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"93c919b9-1485-42fd-a0bd-0608a608f926\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.694977 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/93c919b9-1485-42fd-a0bd-0608a608f926-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"93c919b9-1485-42fd-a0bd-0608a608f926\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.695075 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4vzv\" (UniqueName: \"kubernetes.io/projected/dea338f1-f870-484b-ac43-7da254eac47a-kube-api-access-k4vzv\") on node \"crc\" DevicePath \"\"" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.695099 4766 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dea338f1-f870-484b-ac43-7da254eac47a-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.695109 4766 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dea338f1-f870-484b-ac43-7da254eac47a-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.796450 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/93c919b9-1485-42fd-a0bd-0608a608f926-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"93c919b9-1485-42fd-a0bd-0608a608f926\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.796513 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/93c919b9-1485-42fd-a0bd-0608a608f926-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"93c919b9-1485-42fd-a0bd-0608a608f926\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.796680 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/93c919b9-1485-42fd-a0bd-0608a608f926-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"93c919b9-1485-42fd-a0bd-0608a608f926\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.816853 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/93c919b9-1485-42fd-a0bd-0608a608f926-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"93c919b9-1485-42fd-a0bd-0608a608f926\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 00:26:12 crc kubenswrapper[4766]: I1126 00:26:12.890957 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 00:26:13 crc kubenswrapper[4766]: I1126 00:26:13.168404 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9" event={"ID":"dea338f1-f870-484b-ac43-7da254eac47a","Type":"ContainerDied","Data":"00ed048d8a19eff954d1780d38adc209fb6f3f39901431c063811524ef864ae2"} Nov 26 00:26:13 crc kubenswrapper[4766]: I1126 00:26:13.168477 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00ed048d8a19eff954d1780d38adc209fb6f3f39901431c063811524ef864ae2" Nov 26 00:26:13 crc kubenswrapper[4766]: I1126 00:26:13.168414 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9" Nov 26 00:26:13 crc kubenswrapper[4766]: I1126 00:26:13.171703 4766 generic.go:334] "Generic (PLEG): container finished" podID="a8031d42-e714-4bf0-9976-92a118257476" containerID="5531d8a3a113f10f65f7c24a63d1615c0db05b9d3c92bfdaf3bb0222bd1fa71e" exitCode=0 Nov 26 00:26:13 crc kubenswrapper[4766]: I1126 00:26:13.171855 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-45g8j" event={"ID":"a8031d42-e714-4bf0-9976-92a118257476","Type":"ContainerDied","Data":"5531d8a3a113f10f65f7c24a63d1615c0db05b9d3c92bfdaf3bb0222bd1fa71e"} Nov 26 00:26:13 crc kubenswrapper[4766]: I1126 00:26:13.224561 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 26 00:26:13 crc kubenswrapper[4766]: I1126 00:26:13.364698 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 00:26:13 crc kubenswrapper[4766]: I1126 00:26:13.405933 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5dba56bc-b11d-4f0a-aea4-205c48a40765-kubelet-dir\") pod \"5dba56bc-b11d-4f0a-aea4-205c48a40765\" (UID: \"5dba56bc-b11d-4f0a-aea4-205c48a40765\") " Nov 26 00:26:13 crc kubenswrapper[4766]: I1126 00:26:13.406088 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5dba56bc-b11d-4f0a-aea4-205c48a40765-kube-api-access\") pod \"5dba56bc-b11d-4f0a-aea4-205c48a40765\" (UID: \"5dba56bc-b11d-4f0a-aea4-205c48a40765\") " Nov 26 00:26:13 crc kubenswrapper[4766]: I1126 00:26:13.406748 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dba56bc-b11d-4f0a-aea4-205c48a40765-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5dba56bc-b11d-4f0a-aea4-205c48a40765" (UID: "5dba56bc-b11d-4f0a-aea4-205c48a40765"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:26:13 crc kubenswrapper[4766]: I1126 00:26:13.411326 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dba56bc-b11d-4f0a-aea4-205c48a40765-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5dba56bc-b11d-4f0a-aea4-205c48a40765" (UID: "5dba56bc-b11d-4f0a-aea4-205c48a40765"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:26:13 crc kubenswrapper[4766]: I1126 00:26:13.451349 4766 patch_prober.go:28] interesting pod/router-default-5444994796-45dcj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 00:26:13 crc kubenswrapper[4766]: [-]has-synced failed: reason withheld Nov 26 00:26:13 crc kubenswrapper[4766]: [+]process-running ok Nov 26 00:26:13 crc kubenswrapper[4766]: healthz check failed Nov 26 00:26:13 crc kubenswrapper[4766]: I1126 00:26:13.451397 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-45dcj" podUID="06530996-dd63-4ed5-adb3-790ecc93d11e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 00:26:13 crc kubenswrapper[4766]: I1126 00:26:13.507181 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5dba56bc-b11d-4f0a-aea4-205c48a40765-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 00:26:13 crc kubenswrapper[4766]: I1126 00:26:13.507214 4766 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5dba56bc-b11d-4f0a-aea4-205c48a40765-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 26 00:26:14 crc kubenswrapper[4766]: I1126 00:26:14.232303 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 00:26:14 crc kubenswrapper[4766]: I1126 00:26:14.232303 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5dba56bc-b11d-4f0a-aea4-205c48a40765","Type":"ContainerDied","Data":"ec367e3c330ef234aaf0e957ed4a0833da05b504ea0034c97e35aa5b7e82c138"} Nov 26 00:26:14 crc kubenswrapper[4766]: I1126 00:26:14.232350 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec367e3c330ef234aaf0e957ed4a0833da05b504ea0034c97e35aa5b7e82c138" Nov 26 00:26:14 crc kubenswrapper[4766]: I1126 00:26:14.235008 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"93c919b9-1485-42fd-a0bd-0608a608f926","Type":"ContainerStarted","Data":"18561a09698fd74f521d657d1b54f29602f28fa47eb55121622d483937a82991"} Nov 26 00:26:14 crc kubenswrapper[4766]: I1126 00:26:14.235037 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"93c919b9-1485-42fd-a0bd-0608a608f926","Type":"ContainerStarted","Data":"e846a992bfed8625e20ff86c34f9c32e58964874061f4567c764e8851acec97a"} Nov 26 00:26:14 crc kubenswrapper[4766]: I1126 00:26:14.251176 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.251162233 podStartE2EDuration="2.251162233s" podCreationTimestamp="2025-11-26 00:26:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:14.248147939 +0000 UTC m=+155.096918369" watchObservedRunningTime="2025-11-26 00:26:14.251162233 +0000 UTC m=+155.099932663" Nov 26 00:26:14 crc kubenswrapper[4766]: I1126 00:26:14.450198 4766 patch_prober.go:28] interesting pod/router-default-5444994796-45dcj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 00:26:14 crc kubenswrapper[4766]: [-]has-synced failed: reason withheld Nov 26 00:26:14 crc kubenswrapper[4766]: [+]process-running ok Nov 26 00:26:14 crc kubenswrapper[4766]: healthz check failed Nov 26 00:26:14 crc kubenswrapper[4766]: I1126 00:26:14.450256 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-45dcj" podUID="06530996-dd63-4ed5-adb3-790ecc93d11e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 00:26:15 crc kubenswrapper[4766]: I1126 00:26:15.248246 4766 generic.go:334] "Generic (PLEG): container finished" podID="93c919b9-1485-42fd-a0bd-0608a608f926" containerID="18561a09698fd74f521d657d1b54f29602f28fa47eb55121622d483937a82991" exitCode=0 Nov 26 00:26:15 crc kubenswrapper[4766]: I1126 00:26:15.248617 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"93c919b9-1485-42fd-a0bd-0608a608f926","Type":"ContainerDied","Data":"18561a09698fd74f521d657d1b54f29602f28fa47eb55121622d483937a82991"} Nov 26 00:26:15 crc kubenswrapper[4766]: I1126 00:26:15.452433 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-45dcj" Nov 26 00:26:15 crc kubenswrapper[4766]: I1126 00:26:15.456791 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-45dcj" Nov 26 00:26:16 crc kubenswrapper[4766]: I1126 00:26:16.264307 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-k2shv" Nov 26 00:26:20 crc kubenswrapper[4766]: I1126 00:26:20.715432 4766 patch_prober.go:28] interesting pod/downloads-7954f5f757-bz6r8 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Nov 26 00:26:20 crc kubenswrapper[4766]: I1126 00:26:20.716014 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-bz6r8" podUID="d94c7fca-2c24-4005-9839-0578e1b39283" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Nov 26 00:26:20 crc kubenswrapper[4766]: I1126 00:26:20.715450 4766 patch_prober.go:28] interesting pod/downloads-7954f5f757-bz6r8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Nov 26 00:26:20 crc kubenswrapper[4766]: I1126 00:26:20.716105 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bz6r8" podUID="d94c7fca-2c24-4005-9839-0578e1b39283" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Nov 26 00:26:20 crc kubenswrapper[4766]: I1126 00:26:20.962683 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:20 crc kubenswrapper[4766]: I1126 00:26:20.968887 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:26:22 crc kubenswrapper[4766]: I1126 00:26:22.715228 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 00:26:22 crc kubenswrapper[4766]: I1126 00:26:22.766481 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/93c919b9-1485-42fd-a0bd-0608a608f926-kube-api-access\") pod \"93c919b9-1485-42fd-a0bd-0608a608f926\" (UID: \"93c919b9-1485-42fd-a0bd-0608a608f926\") " Nov 26 00:26:22 crc kubenswrapper[4766]: I1126 00:26:22.766559 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/93c919b9-1485-42fd-a0bd-0608a608f926-kubelet-dir\") pod \"93c919b9-1485-42fd-a0bd-0608a608f926\" (UID: \"93c919b9-1485-42fd-a0bd-0608a608f926\") " Nov 26 00:26:22 crc kubenswrapper[4766]: I1126 00:26:22.767088 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93c919b9-1485-42fd-a0bd-0608a608f926-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "93c919b9-1485-42fd-a0bd-0608a608f926" (UID: "93c919b9-1485-42fd-a0bd-0608a608f926"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:26:22 crc kubenswrapper[4766]: I1126 00:26:22.772905 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93c919b9-1485-42fd-a0bd-0608a608f926-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "93c919b9-1485-42fd-a0bd-0608a608f926" (UID: "93c919b9-1485-42fd-a0bd-0608a608f926"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:26:22 crc kubenswrapper[4766]: I1126 00:26:22.868085 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/93c919b9-1485-42fd-a0bd-0608a608f926-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 00:26:22 crc kubenswrapper[4766]: I1126 00:26:22.868125 4766 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/93c919b9-1485-42fd-a0bd-0608a608f926-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 26 00:26:23 crc kubenswrapper[4766]: I1126 00:26:23.355568 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"93c919b9-1485-42fd-a0bd-0608a608f926","Type":"ContainerDied","Data":"e846a992bfed8625e20ff86c34f9c32e58964874061f4567c764e8851acec97a"} Nov 26 00:26:23 crc kubenswrapper[4766]: I1126 00:26:23.355632 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e846a992bfed8625e20ff86c34f9c32e58964874061f4567c764e8851acec97a" Nov 26 00:26:23 crc kubenswrapper[4766]: I1126 00:26:23.355640 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 00:26:26 crc kubenswrapper[4766]: I1126 00:26:26.321127 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs\") pod \"network-metrics-daemon-5z9ll\" (UID: \"b6600efa-4858-41aa-a890-44ce262b63dd\") " pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:26:26 crc kubenswrapper[4766]: I1126 00:26:26.329576 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6600efa-4858-41aa-a890-44ce262b63dd-metrics-certs\") pod \"network-metrics-daemon-5z9ll\" (UID: \"b6600efa-4858-41aa-a890-44ce262b63dd\") " pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:26:26 crc kubenswrapper[4766]: I1126 00:26:26.579116 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-5z9ll" Nov 26 00:26:26 crc kubenswrapper[4766]: I1126 00:26:26.975263 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-5z9ll"] Nov 26 00:26:30 crc kubenswrapper[4766]: I1126 00:26:30.350821 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:26:30 crc kubenswrapper[4766]: I1126 00:26:30.722562 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-bz6r8" Nov 26 00:26:34 crc kubenswrapper[4766]: I1126 00:26:34.434771 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-5z9ll" event={"ID":"b6600efa-4858-41aa-a890-44ce262b63dd","Type":"ContainerStarted","Data":"49aeb9630fd1bd97dcd361249ffc2d8bcce287430a4710b18cc77573ff51eb17"} Nov 26 00:26:38 crc kubenswrapper[4766]: I1126 00:26:38.458148 4766 generic.go:334] "Generic (PLEG): container finished" podID="284847a7-94c0-4f03-a54d-6d22fe77353f" containerID="c6b8df07f4c9fc0d3940bf06cd60a70cb5b80db504d71e9a09de043504fb206e" exitCode=0 Nov 26 00:26:38 crc kubenswrapper[4766]: I1126 00:26:38.458276 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29401920-82g4w" event={"ID":"284847a7-94c0-4f03-a54d-6d22fe77353f","Type":"ContainerDied","Data":"c6b8df07f4c9fc0d3940bf06cd60a70cb5b80db504d71e9a09de043504fb206e"} Nov 26 00:26:40 crc kubenswrapper[4766]: I1126 00:26:40.472637 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29401920-82g4w" event={"ID":"284847a7-94c0-4f03-a54d-6d22fe77353f","Type":"ContainerDied","Data":"33242ffa5a061764e610c7855f70ed79b90fb80b3a9d9300ca24769ab2d50bfb"} Nov 26 00:26:40 crc kubenswrapper[4766]: I1126 00:26:40.473225 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33242ffa5a061764e610c7855f70ed79b90fb80b3a9d9300ca24769ab2d50bfb" Nov 26 00:26:40 crc kubenswrapper[4766]: I1126 00:26:40.498794 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29401920-82g4w" Nov 26 00:26:40 crc kubenswrapper[4766]: I1126 00:26:40.524441 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/284847a7-94c0-4f03-a54d-6d22fe77353f-serviceca\") pod \"284847a7-94c0-4f03-a54d-6d22fe77353f\" (UID: \"284847a7-94c0-4f03-a54d-6d22fe77353f\") " Nov 26 00:26:40 crc kubenswrapper[4766]: I1126 00:26:40.524554 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gml6x\" (UniqueName: \"kubernetes.io/projected/284847a7-94c0-4f03-a54d-6d22fe77353f-kube-api-access-gml6x\") pod \"284847a7-94c0-4f03-a54d-6d22fe77353f\" (UID: \"284847a7-94c0-4f03-a54d-6d22fe77353f\") " Nov 26 00:26:40 crc kubenswrapper[4766]: I1126 00:26:40.525146 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/284847a7-94c0-4f03-a54d-6d22fe77353f-serviceca" (OuterVolumeSpecName: "serviceca") pod "284847a7-94c0-4f03-a54d-6d22fe77353f" (UID: "284847a7-94c0-4f03-a54d-6d22fe77353f"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:26:40 crc kubenswrapper[4766]: I1126 00:26:40.532917 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/284847a7-94c0-4f03-a54d-6d22fe77353f-kube-api-access-gml6x" (OuterVolumeSpecName: "kube-api-access-gml6x") pod "284847a7-94c0-4f03-a54d-6d22fe77353f" (UID: "284847a7-94c0-4f03-a54d-6d22fe77353f"). InnerVolumeSpecName "kube-api-access-gml6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:26:40 crc kubenswrapper[4766]: I1126 00:26:40.626402 4766 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/284847a7-94c0-4f03-a54d-6d22fe77353f-serviceca\") on node \"crc\" DevicePath \"\"" Nov 26 00:26:40 crc kubenswrapper[4766]: I1126 00:26:40.626448 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gml6x\" (UniqueName: \"kubernetes.io/projected/284847a7-94c0-4f03-a54d-6d22fe77353f-kube-api-access-gml6x\") on node \"crc\" DevicePath \"\"" Nov 26 00:26:41 crc kubenswrapper[4766]: I1126 00:26:41.179100 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vvlgv" Nov 26 00:26:41 crc kubenswrapper[4766]: I1126 00:26:41.476852 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29401920-82g4w" Nov 26 00:26:41 crc kubenswrapper[4766]: I1126 00:26:41.479395 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:26:41 crc kubenswrapper[4766]: I1126 00:26:41.479437 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:26:42 crc kubenswrapper[4766]: E1126 00:26:42.327605 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 26 00:26:42 crc kubenswrapper[4766]: E1126 00:26:42.327798 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9bwtw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-jnqdt_openshift-marketplace(90830d76-8bef-4c13-b127-23fff3e32770): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 00:26:42 crc kubenswrapper[4766]: E1126 00:26:42.329579 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-jnqdt" podUID="90830d76-8bef-4c13-b127-23fff3e32770" Nov 26 00:26:46 crc kubenswrapper[4766]: E1126 00:26:46.212288 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-jnqdt" podUID="90830d76-8bef-4c13-b127-23fff3e32770" Nov 26 00:26:46 crc kubenswrapper[4766]: E1126 00:26:46.260029 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 26 00:26:46 crc kubenswrapper[4766]: E1126 00:26:46.260196 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vvmjt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-gxwnj_openshift-marketplace(c9808e13-0b77-4493-820a-80a80111d0a0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 00:26:46 crc kubenswrapper[4766]: E1126 00:26:46.262098 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-gxwnj" podUID="c9808e13-0b77-4493-820a-80a80111d0a0" Nov 26 00:26:46 crc kubenswrapper[4766]: E1126 00:26:46.626574 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 26 00:26:46 crc kubenswrapper[4766]: E1126 00:26:46.626732 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lgnbg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-bjz9v_openshift-marketplace(e0d4c85c-2a27-4a2f-903e-8f7473351104): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 00:26:46 crc kubenswrapper[4766]: E1126 00:26:46.628021 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-bjz9v" podUID="e0d4c85c-2a27-4a2f-903e-8f7473351104" Nov 26 00:26:47 crc kubenswrapper[4766]: I1126 00:26:47.087909 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 00:26:53 crc kubenswrapper[4766]: E1126 00:26:53.097524 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-gxwnj" podUID="c9808e13-0b77-4493-820a-80a80111d0a0" Nov 26 00:26:53 crc kubenswrapper[4766]: E1126 00:26:53.098245 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bjz9v" podUID="e0d4c85c-2a27-4a2f-903e-8f7473351104" Nov 26 00:26:53 crc kubenswrapper[4766]: E1126 00:26:53.984211 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 26 00:26:53 crc kubenswrapper[4766]: E1126 00:26:53.984401 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tp92b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-45g8j_openshift-marketplace(a8031d42-e714-4bf0-9976-92a118257476): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 00:26:53 crc kubenswrapper[4766]: E1126 00:26:53.985617 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-45g8j" podUID="a8031d42-e714-4bf0-9976-92a118257476" Nov 26 00:26:54 crc kubenswrapper[4766]: E1126 00:26:54.122308 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 26 00:26:54 crc kubenswrapper[4766]: E1126 00:26:54.122489 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2z5vl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-s9x5n_openshift-marketplace(5f6c20b5-b168-4529-b84a-bef5c19254d5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 00:26:54 crc kubenswrapper[4766]: E1126 00:26:54.123695 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-s9x5n" podUID="5f6c20b5-b168-4529-b84a-bef5c19254d5" Nov 26 00:26:55 crc kubenswrapper[4766]: E1126 00:26:55.138544 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-s9x5n" podUID="5f6c20b5-b168-4529-b84a-bef5c19254d5" Nov 26 00:26:55 crc kubenswrapper[4766]: E1126 00:26:55.138993 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-45g8j" podUID="a8031d42-e714-4bf0-9976-92a118257476" Nov 26 00:26:55 crc kubenswrapper[4766]: E1126 00:26:55.296099 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 26 00:26:55 crc kubenswrapper[4766]: E1126 00:26:55.296496 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xktn8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-fhlnl_openshift-marketplace(1529f474-493a-485a-80b7-ee8ec1c79ccc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 00:26:55 crc kubenswrapper[4766]: E1126 00:26:55.297702 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-fhlnl" podUID="1529f474-493a-485a-80b7-ee8ec1c79ccc" Nov 26 00:26:55 crc kubenswrapper[4766]: E1126 00:26:55.476102 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 26 00:26:55 crc kubenswrapper[4766]: E1126 00:26:55.476258 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j9bwg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-5hq26_openshift-marketplace(cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 00:26:55 crc kubenswrapper[4766]: E1126 00:26:55.477478 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-5hq26" podUID="cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2" Nov 26 00:26:55 crc kubenswrapper[4766]: E1126 00:26:55.552530 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-fhlnl" podUID="1529f474-493a-485a-80b7-ee8ec1c79ccc" Nov 26 00:26:55 crc kubenswrapper[4766]: E1126 00:26:55.554486 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5hq26" podUID="cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2" Nov 26 00:26:55 crc kubenswrapper[4766]: E1126 00:26:55.709995 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 26 00:26:55 crc kubenswrapper[4766]: E1126 00:26:55.710165 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wvfrt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-tqzxf_openshift-marketplace(519cd16f-3b69-4365-866d-47f01ecd0c74): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 00:26:55 crc kubenswrapper[4766]: E1126 00:26:55.712104 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-tqzxf" podUID="519cd16f-3b69-4365-866d-47f01ecd0c74" Nov 26 00:26:56 crc kubenswrapper[4766]: I1126 00:26:56.559923 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-5z9ll" event={"ID":"b6600efa-4858-41aa-a890-44ce262b63dd","Type":"ContainerStarted","Data":"77e7245bf11b664471c9510aec74b8d36ebddae8d388aff4634c6db9c0bdbe52"} Nov 26 00:26:56 crc kubenswrapper[4766]: E1126 00:26:56.562435 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-tqzxf" podUID="519cd16f-3b69-4365-866d-47f01ecd0c74" Nov 26 00:26:57 crc kubenswrapper[4766]: I1126 00:26:57.568123 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-5z9ll" event={"ID":"b6600efa-4858-41aa-a890-44ce262b63dd","Type":"ContainerStarted","Data":"e4ea6a77d4c466f7cba808d67f9f413265b386298867a35ff1702b752c0fd5a2"} Nov 26 00:26:57 crc kubenswrapper[4766]: I1126 00:26:57.593981 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-5z9ll" podStartSLOduration=174.593954572 podStartE2EDuration="2m54.593954572s" podCreationTimestamp="2025-11-26 00:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:26:57.585412834 +0000 UTC m=+198.434183304" watchObservedRunningTime="2025-11-26 00:26:57.593954572 +0000 UTC m=+198.442725012" Nov 26 00:26:58 crc kubenswrapper[4766]: I1126 00:26:58.574915 4766 generic.go:334] "Generic (PLEG): container finished" podID="90830d76-8bef-4c13-b127-23fff3e32770" containerID="5592e87ce814e03fa260d84001e369aa18aac8f3a8f478b6c309b81b7db5ee5f" exitCode=0 Nov 26 00:26:58 crc kubenswrapper[4766]: I1126 00:26:58.574988 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnqdt" event={"ID":"90830d76-8bef-4c13-b127-23fff3e32770","Type":"ContainerDied","Data":"5592e87ce814e03fa260d84001e369aa18aac8f3a8f478b6c309b81b7db5ee5f"} Nov 26 00:26:59 crc kubenswrapper[4766]: I1126 00:26:59.581715 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnqdt" event={"ID":"90830d76-8bef-4c13-b127-23fff3e32770","Type":"ContainerStarted","Data":"88933fddc0f7bbceaeeaf1000987d3306168ed1c04a17ee51b95a84b87a57e90"} Nov 26 00:26:59 crc kubenswrapper[4766]: I1126 00:26:59.599446 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jnqdt" podStartSLOduration=3.330807441 podStartE2EDuration="52.599426668s" podCreationTimestamp="2025-11-26 00:26:07 +0000 UTC" firstStartedPulling="2025-11-26 00:26:10.038756603 +0000 UTC m=+150.887527033" lastFinishedPulling="2025-11-26 00:26:59.30737582 +0000 UTC m=+200.156146260" observedRunningTime="2025-11-26 00:26:59.597788367 +0000 UTC m=+200.446558827" watchObservedRunningTime="2025-11-26 00:26:59.599426668 +0000 UTC m=+200.448197098" Nov 26 00:27:06 crc kubenswrapper[4766]: I1126 00:27:06.621939 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9x5n" event={"ID":"5f6c20b5-b168-4529-b84a-bef5c19254d5","Type":"ContainerStarted","Data":"d3daea790428b806b4c3fb6edd90c6968882cacfdc2cbffc88c3418f31ab79bc"} Nov 26 00:27:07 crc kubenswrapper[4766]: I1126 00:27:07.628294 4766 generic.go:334] "Generic (PLEG): container finished" podID="5f6c20b5-b168-4529-b84a-bef5c19254d5" containerID="d3daea790428b806b4c3fb6edd90c6968882cacfdc2cbffc88c3418f31ab79bc" exitCode=0 Nov 26 00:27:07 crc kubenswrapper[4766]: I1126 00:27:07.628361 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9x5n" event={"ID":"5f6c20b5-b168-4529-b84a-bef5c19254d5","Type":"ContainerDied","Data":"d3daea790428b806b4c3fb6edd90c6968882cacfdc2cbffc88c3418f31ab79bc"} Nov 26 00:27:07 crc kubenswrapper[4766]: I1126 00:27:07.946399 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jnqdt" Nov 26 00:27:07 crc kubenswrapper[4766]: I1126 00:27:07.946595 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jnqdt" Nov 26 00:27:08 crc kubenswrapper[4766]: I1126 00:27:08.113869 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jnqdt" Nov 26 00:27:08 crc kubenswrapper[4766]: I1126 00:27:08.634686 4766 generic.go:334] "Generic (PLEG): container finished" podID="cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2" containerID="cfa545f753933cfde8402e62bc238067b73040b42c1ba6861d277c2ab175906b" exitCode=0 Nov 26 00:27:08 crc kubenswrapper[4766]: I1126 00:27:08.634737 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hq26" event={"ID":"cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2","Type":"ContainerDied","Data":"cfa545f753933cfde8402e62bc238067b73040b42c1ba6861d277c2ab175906b"} Nov 26 00:27:08 crc kubenswrapper[4766]: I1126 00:27:08.639752 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxwnj" event={"ID":"c9808e13-0b77-4493-820a-80a80111d0a0","Type":"ContainerStarted","Data":"db88f06b41fc55c52153c6dc469566fcb7d690952679227835cebc655c198d1a"} Nov 26 00:27:08 crc kubenswrapper[4766]: I1126 00:27:08.642719 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9x5n" event={"ID":"5f6c20b5-b168-4529-b84a-bef5c19254d5","Type":"ContainerStarted","Data":"589b3e5703712a442130255630f9f46e0f819c363e3eb4fd4223fae322d7fdc5"} Nov 26 00:27:08 crc kubenswrapper[4766]: I1126 00:27:08.668363 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-s9x5n" podStartSLOduration=2.730692243 podStartE2EDuration="58.668334045s" podCreationTimestamp="2025-11-26 00:26:10 +0000 UTC" firstStartedPulling="2025-11-26 00:26:12.164698829 +0000 UTC m=+153.013469259" lastFinishedPulling="2025-11-26 00:27:08.102340631 +0000 UTC m=+208.951111061" observedRunningTime="2025-11-26 00:27:08.667047372 +0000 UTC m=+209.515817802" watchObservedRunningTime="2025-11-26 00:27:08.668334045 +0000 UTC m=+209.517104475" Nov 26 00:27:08 crc kubenswrapper[4766]: I1126 00:27:08.686594 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jnqdt" Nov 26 00:27:09 crc kubenswrapper[4766]: I1126 00:27:09.653500 4766 generic.go:334] "Generic (PLEG): container finished" podID="c9808e13-0b77-4493-820a-80a80111d0a0" containerID="db88f06b41fc55c52153c6dc469566fcb7d690952679227835cebc655c198d1a" exitCode=0 Nov 26 00:27:09 crc kubenswrapper[4766]: I1126 00:27:09.653567 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxwnj" event={"ID":"c9808e13-0b77-4493-820a-80a80111d0a0","Type":"ContainerDied","Data":"db88f06b41fc55c52153c6dc469566fcb7d690952679227835cebc655c198d1a"} Nov 26 00:27:09 crc kubenswrapper[4766]: I1126 00:27:09.656410 4766 generic.go:334] "Generic (PLEG): container finished" podID="a8031d42-e714-4bf0-9976-92a118257476" containerID="86b1bad4e51b33d503df14bd78c94023af97dc3e3e74f2521b39ca568196be22" exitCode=0 Nov 26 00:27:09 crc kubenswrapper[4766]: I1126 00:27:09.656467 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-45g8j" event={"ID":"a8031d42-e714-4bf0-9976-92a118257476","Type":"ContainerDied","Data":"86b1bad4e51b33d503df14bd78c94023af97dc3e3e74f2521b39ca568196be22"} Nov 26 00:27:09 crc kubenswrapper[4766]: I1126 00:27:09.661540 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hq26" event={"ID":"cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2","Type":"ContainerStarted","Data":"cc95df5b82620cb48f2b8c099b43cbc580d65b8ead0fad0d7113134b93ce2a8b"} Nov 26 00:27:09 crc kubenswrapper[4766]: I1126 00:27:09.664792 4766 generic.go:334] "Generic (PLEG): container finished" podID="1529f474-493a-485a-80b7-ee8ec1c79ccc" containerID="053afe7852af7f429134c0b1751a485763eb8c26ff50a42037d8eb4238d3b56d" exitCode=0 Nov 26 00:27:09 crc kubenswrapper[4766]: I1126 00:27:09.665081 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fhlnl" event={"ID":"1529f474-493a-485a-80b7-ee8ec1c79ccc","Type":"ContainerDied","Data":"053afe7852af7f429134c0b1751a485763eb8c26ff50a42037d8eb4238d3b56d"} Nov 26 00:27:09 crc kubenswrapper[4766]: I1126 00:27:09.712936 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5hq26" podStartSLOduration=3.678430004 podStartE2EDuration="1m2.712915745s" podCreationTimestamp="2025-11-26 00:26:07 +0000 UTC" firstStartedPulling="2025-11-26 00:26:10.040842264 +0000 UTC m=+150.889612694" lastFinishedPulling="2025-11-26 00:27:09.075328015 +0000 UTC m=+209.924098435" observedRunningTime="2025-11-26 00:27:09.69350927 +0000 UTC m=+210.542279700" watchObservedRunningTime="2025-11-26 00:27:09.712915745 +0000 UTC m=+210.561686185" Nov 26 00:27:10 crc kubenswrapper[4766]: I1126 00:27:10.903243 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-s9x5n" Nov 26 00:27:10 crc kubenswrapper[4766]: I1126 00:27:10.903760 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-s9x5n" Nov 26 00:27:11 crc kubenswrapper[4766]: I1126 00:27:11.480563 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:27:11 crc kubenswrapper[4766]: I1126 00:27:11.481311 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:27:11 crc kubenswrapper[4766]: I1126 00:27:11.481366 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:27:11 crc kubenswrapper[4766]: I1126 00:27:11.482103 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 00:27:11 crc kubenswrapper[4766]: I1126 00:27:11.482235 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf" gracePeriod=600 Nov 26 00:27:11 crc kubenswrapper[4766]: I1126 00:27:11.680416 4766 generic.go:334] "Generic (PLEG): container finished" podID="519cd16f-3b69-4365-866d-47f01ecd0c74" containerID="e103055425cfc553a7ba8f54d73774e760d7496ff55334dfb0f2b9b9f68b3352" exitCode=0 Nov 26 00:27:11 crc kubenswrapper[4766]: I1126 00:27:11.680488 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tqzxf" event={"ID":"519cd16f-3b69-4365-866d-47f01ecd0c74","Type":"ContainerDied","Data":"e103055425cfc553a7ba8f54d73774e760d7496ff55334dfb0f2b9b9f68b3352"} Nov 26 00:27:11 crc kubenswrapper[4766]: I1126 00:27:11.687993 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bjz9v" event={"ID":"e0d4c85c-2a27-4a2f-903e-8f7473351104","Type":"ContainerStarted","Data":"834a1c0046413df441fd1d31eb8b082de3727f22cabb3144a9f0029dc6fe4984"} Nov 26 00:27:11 crc kubenswrapper[4766]: I1126 00:27:11.689908 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf" exitCode=0 Nov 26 00:27:11 crc kubenswrapper[4766]: I1126 00:27:11.689952 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf"} Nov 26 00:27:11 crc kubenswrapper[4766]: I1126 00:27:11.691759 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fhlnl" event={"ID":"1529f474-493a-485a-80b7-ee8ec1c79ccc","Type":"ContainerStarted","Data":"7bbaf76b46cc33ba0b12824530776175690a99499a1ca6cdce26885e22214afe"} Nov 26 00:27:11 crc kubenswrapper[4766]: I1126 00:27:11.730616 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fhlnl" podStartSLOduration=2.548437167 podStartE2EDuration="1m2.730592424s" podCreationTimestamp="2025-11-26 00:26:09 +0000 UTC" firstStartedPulling="2025-11-26 00:26:11.12039609 +0000 UTC m=+151.969166520" lastFinishedPulling="2025-11-26 00:27:11.302551347 +0000 UTC m=+212.151321777" observedRunningTime="2025-11-26 00:27:11.724283363 +0000 UTC m=+212.573053793" watchObservedRunningTime="2025-11-26 00:27:11.730592424 +0000 UTC m=+212.579362864" Nov 26 00:27:11 crc kubenswrapper[4766]: I1126 00:27:11.956571 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-s9x5n" podUID="5f6c20b5-b168-4529-b84a-bef5c19254d5" containerName="registry-server" probeResult="failure" output=< Nov 26 00:27:11 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 00:27:11 crc kubenswrapper[4766]: > Nov 26 00:27:12 crc kubenswrapper[4766]: I1126 00:27:12.697669 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxwnj" event={"ID":"c9808e13-0b77-4493-820a-80a80111d0a0","Type":"ContainerStarted","Data":"6f6536718602539efcf5904c3e38d9a742ea042afdaa08c209e03fbfec04e08f"} Nov 26 00:27:12 crc kubenswrapper[4766]: I1126 00:27:12.699280 4766 generic.go:334] "Generic (PLEG): container finished" podID="e0d4c85c-2a27-4a2f-903e-8f7473351104" containerID="834a1c0046413df441fd1d31eb8b082de3727f22cabb3144a9f0029dc6fe4984" exitCode=0 Nov 26 00:27:12 crc kubenswrapper[4766]: I1126 00:27:12.699346 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bjz9v" event={"ID":"e0d4c85c-2a27-4a2f-903e-8f7473351104","Type":"ContainerDied","Data":"834a1c0046413df441fd1d31eb8b082de3727f22cabb3144a9f0029dc6fe4984"} Nov 26 00:27:12 crc kubenswrapper[4766]: I1126 00:27:12.702107 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-45g8j" event={"ID":"a8031d42-e714-4bf0-9976-92a118257476","Type":"ContainerStarted","Data":"ee15729c23ecffe1160cced39eb71af4481782f53f7c036dc06718cc0b3f22c7"} Nov 26 00:27:12 crc kubenswrapper[4766]: I1126 00:27:12.705568 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"8e60726cc9ad1a2202d54a17584e9c974ee0288e42d42f37ee8591c1fcee8e0e"} Nov 26 00:27:12 crc kubenswrapper[4766]: I1126 00:27:12.708551 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tqzxf" event={"ID":"519cd16f-3b69-4365-866d-47f01ecd0c74","Type":"ContainerStarted","Data":"ade52e3c1c34cb5c1cef3967f146829c39a9b70d0794b3e04f70eca12714a566"} Nov 26 00:27:12 crc kubenswrapper[4766]: I1126 00:27:12.748937 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gxwnj" podStartSLOduration=4.341194693 podStartE2EDuration="1m5.748916464s" podCreationTimestamp="2025-11-26 00:26:07 +0000 UTC" firstStartedPulling="2025-11-26 00:26:10.023790055 +0000 UTC m=+150.872560485" lastFinishedPulling="2025-11-26 00:27:11.431511816 +0000 UTC m=+212.280282256" observedRunningTime="2025-11-26 00:27:12.733408868 +0000 UTC m=+213.582179298" watchObservedRunningTime="2025-11-26 00:27:12.748916464 +0000 UTC m=+213.597686894" Nov 26 00:27:12 crc kubenswrapper[4766]: I1126 00:27:12.782357 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tqzxf" podStartSLOduration=2.369375101 podStartE2EDuration="1m3.782340856s" podCreationTimestamp="2025-11-26 00:26:09 +0000 UTC" firstStartedPulling="2025-11-26 00:26:11.062387432 +0000 UTC m=+151.911157862" lastFinishedPulling="2025-11-26 00:27:12.475353187 +0000 UTC m=+213.324123617" observedRunningTime="2025-11-26 00:27:12.779588166 +0000 UTC m=+213.628358606" watchObservedRunningTime="2025-11-26 00:27:12.782340856 +0000 UTC m=+213.631111286" Nov 26 00:27:12 crc kubenswrapper[4766]: I1126 00:27:12.800817 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-45g8j" podStartSLOduration=4.407951265 podStartE2EDuration="1m2.800803187s" podCreationTimestamp="2025-11-26 00:26:10 +0000 UTC" firstStartedPulling="2025-11-26 00:26:13.174049047 +0000 UTC m=+154.022819477" lastFinishedPulling="2025-11-26 00:27:11.566900969 +0000 UTC m=+212.415671399" observedRunningTime="2025-11-26 00:27:12.798285243 +0000 UTC m=+213.647055663" watchObservedRunningTime="2025-11-26 00:27:12.800803187 +0000 UTC m=+213.649573607" Nov 26 00:27:13 crc kubenswrapper[4766]: I1126 00:27:13.718280 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bjz9v" event={"ID":"e0d4c85c-2a27-4a2f-903e-8f7473351104","Type":"ContainerStarted","Data":"5a5cd04a55e765bfe4abeea8753044a245db57f489369162d3e283d02e8cc241"} Nov 26 00:27:17 crc kubenswrapper[4766]: I1126 00:27:17.771404 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bjz9v" Nov 26 00:27:17 crc kubenswrapper[4766]: I1126 00:27:17.772152 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bjz9v" Nov 26 00:27:17 crc kubenswrapper[4766]: I1126 00:27:17.811787 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bjz9v" Nov 26 00:27:17 crc kubenswrapper[4766]: I1126 00:27:17.831075 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bjz9v" podStartSLOduration=7.701949761 podStartE2EDuration="1m10.831059376s" podCreationTimestamp="2025-11-26 00:26:07 +0000 UTC" firstStartedPulling="2025-11-26 00:26:10.037989184 +0000 UTC m=+150.886759614" lastFinishedPulling="2025-11-26 00:27:13.167098799 +0000 UTC m=+214.015869229" observedRunningTime="2025-11-26 00:27:13.742067533 +0000 UTC m=+214.590837963" watchObservedRunningTime="2025-11-26 00:27:17.831059376 +0000 UTC m=+218.679829806" Nov 26 00:27:18 crc kubenswrapper[4766]: I1126 00:27:18.123805 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5hq26" Nov 26 00:27:18 crc kubenswrapper[4766]: I1126 00:27:18.123880 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5hq26" Nov 26 00:27:18 crc kubenswrapper[4766]: I1126 00:27:18.166635 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5hq26" Nov 26 00:27:18 crc kubenswrapper[4766]: I1126 00:27:18.445008 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gxwnj" Nov 26 00:27:18 crc kubenswrapper[4766]: I1126 00:27:18.445084 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gxwnj" Nov 26 00:27:18 crc kubenswrapper[4766]: I1126 00:27:18.487402 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gxwnj" Nov 26 00:27:18 crc kubenswrapper[4766]: I1126 00:27:18.781671 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bjz9v" Nov 26 00:27:18 crc kubenswrapper[4766]: I1126 00:27:18.782505 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5hq26" Nov 26 00:27:18 crc kubenswrapper[4766]: I1126 00:27:18.786566 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gxwnj" Nov 26 00:27:19 crc kubenswrapper[4766]: I1126 00:27:19.457048 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5hq26"] Nov 26 00:27:19 crc kubenswrapper[4766]: I1126 00:27:19.805437 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fhlnl" Nov 26 00:27:19 crc kubenswrapper[4766]: I1126 00:27:19.805511 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fhlnl" Nov 26 00:27:19 crc kubenswrapper[4766]: I1126 00:27:19.845999 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fhlnl" Nov 26 00:27:20 crc kubenswrapper[4766]: I1126 00:27:20.072277 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tqzxf" Nov 26 00:27:20 crc kubenswrapper[4766]: I1126 00:27:20.072327 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tqzxf" Nov 26 00:27:20 crc kubenswrapper[4766]: I1126 00:27:20.118233 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tqzxf" Nov 26 00:27:20 crc kubenswrapper[4766]: I1126 00:27:20.757253 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5hq26" podUID="cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2" containerName="registry-server" containerID="cri-o://cc95df5b82620cb48f2b8c099b43cbc580d65b8ead0fad0d7113134b93ce2a8b" gracePeriod=2 Nov 26 00:27:20 crc kubenswrapper[4766]: I1126 00:27:20.797753 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tqzxf" Nov 26 00:27:20 crc kubenswrapper[4766]: I1126 00:27:20.804252 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fhlnl" Nov 26 00:27:20 crc kubenswrapper[4766]: I1126 00:27:20.857159 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gxwnj"] Nov 26 00:27:20 crc kubenswrapper[4766]: I1126 00:27:20.857414 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gxwnj" podUID="c9808e13-0b77-4493-820a-80a80111d0a0" containerName="registry-server" containerID="cri-o://6f6536718602539efcf5904c3e38d9a742ea042afdaa08c209e03fbfec04e08f" gracePeriod=2 Nov 26 00:27:20 crc kubenswrapper[4766]: I1126 00:27:20.945979 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-s9x5n" Nov 26 00:27:20 crc kubenswrapper[4766]: I1126 00:27:20.984175 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-s9x5n" Nov 26 00:27:21 crc kubenswrapper[4766]: I1126 00:27:21.334126 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-45g8j" Nov 26 00:27:21 crc kubenswrapper[4766]: I1126 00:27:21.334176 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-45g8j" Nov 26 00:27:21 crc kubenswrapper[4766]: I1126 00:27:21.389346 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-45g8j" Nov 26 00:27:21 crc kubenswrapper[4766]: E1126 00:27:21.583939 4766 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9808e13_0b77_4493_820a_80a80111d0a0.slice/crio-6f6536718602539efcf5904c3e38d9a742ea042afdaa08c209e03fbfec04e08f.scope\": RecentStats: unable to find data in memory cache]" Nov 26 00:27:21 crc kubenswrapper[4766]: I1126 00:27:21.804205 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-45g8j" Nov 26 00:27:23 crc kubenswrapper[4766]: I1126 00:27:23.255354 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tqzxf"] Nov 26 00:27:23 crc kubenswrapper[4766]: I1126 00:27:23.255636 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tqzxf" podUID="519cd16f-3b69-4365-866d-47f01ecd0c74" containerName="registry-server" containerID="cri-o://ade52e3c1c34cb5c1cef3967f146829c39a9b70d0794b3e04f70eca12714a566" gracePeriod=2 Nov 26 00:27:23 crc kubenswrapper[4766]: I1126 00:27:23.772564 4766 generic.go:334] "Generic (PLEG): container finished" podID="c9808e13-0b77-4493-820a-80a80111d0a0" containerID="6f6536718602539efcf5904c3e38d9a742ea042afdaa08c209e03fbfec04e08f" exitCode=0 Nov 26 00:27:23 crc kubenswrapper[4766]: I1126 00:27:23.772642 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxwnj" event={"ID":"c9808e13-0b77-4493-820a-80a80111d0a0","Type":"ContainerDied","Data":"6f6536718602539efcf5904c3e38d9a742ea042afdaa08c209e03fbfec04e08f"} Nov 26 00:27:25 crc kubenswrapper[4766]: I1126 00:27:25.657103 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-45g8j"] Nov 26 00:27:25 crc kubenswrapper[4766]: I1126 00:27:25.658182 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-45g8j" podUID="a8031d42-e714-4bf0-9976-92a118257476" containerName="registry-server" containerID="cri-o://ee15729c23ecffe1160cced39eb71af4481782f53f7c036dc06718cc0b3f22c7" gracePeriod=2 Nov 26 00:27:25 crc kubenswrapper[4766]: I1126 00:27:25.783235 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5hq26_cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2/registry-server/0.log" Nov 26 00:27:25 crc kubenswrapper[4766]: I1126 00:27:25.784109 4766 generic.go:334] "Generic (PLEG): container finished" podID="cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2" containerID="cc95df5b82620cb48f2b8c099b43cbc580d65b8ead0fad0d7113134b93ce2a8b" exitCode=137 Nov 26 00:27:25 crc kubenswrapper[4766]: I1126 00:27:25.784189 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hq26" event={"ID":"cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2","Type":"ContainerDied","Data":"cc95df5b82620cb48f2b8c099b43cbc580d65b8ead0fad0d7113134b93ce2a8b"} Nov 26 00:27:25 crc kubenswrapper[4766]: I1126 00:27:25.786105 4766 generic.go:334] "Generic (PLEG): container finished" podID="519cd16f-3b69-4365-866d-47f01ecd0c74" containerID="ade52e3c1c34cb5c1cef3967f146829c39a9b70d0794b3e04f70eca12714a566" exitCode=0 Nov 26 00:27:25 crc kubenswrapper[4766]: I1126 00:27:25.786141 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tqzxf" event={"ID":"519cd16f-3b69-4365-866d-47f01ecd0c74","Type":"ContainerDied","Data":"ade52e3c1c34cb5c1cef3967f146829c39a9b70d0794b3e04f70eca12714a566"} Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.138666 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxwnj" Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.246986 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9808e13-0b77-4493-820a-80a80111d0a0-catalog-content\") pod \"c9808e13-0b77-4493-820a-80a80111d0a0\" (UID: \"c9808e13-0b77-4493-820a-80a80111d0a0\") " Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.247078 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvmjt\" (UniqueName: \"kubernetes.io/projected/c9808e13-0b77-4493-820a-80a80111d0a0-kube-api-access-vvmjt\") pod \"c9808e13-0b77-4493-820a-80a80111d0a0\" (UID: \"c9808e13-0b77-4493-820a-80a80111d0a0\") " Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.247148 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9808e13-0b77-4493-820a-80a80111d0a0-utilities\") pod \"c9808e13-0b77-4493-820a-80a80111d0a0\" (UID: \"c9808e13-0b77-4493-820a-80a80111d0a0\") " Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.247886 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9808e13-0b77-4493-820a-80a80111d0a0-utilities" (OuterVolumeSpecName: "utilities") pod "c9808e13-0b77-4493-820a-80a80111d0a0" (UID: "c9808e13-0b77-4493-820a-80a80111d0a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.272930 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9808e13-0b77-4493-820a-80a80111d0a0-kube-api-access-vvmjt" (OuterVolumeSpecName: "kube-api-access-vvmjt") pod "c9808e13-0b77-4493-820a-80a80111d0a0" (UID: "c9808e13-0b77-4493-820a-80a80111d0a0"). InnerVolumeSpecName "kube-api-access-vvmjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.349013 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9808e13-0b77-4493-820a-80a80111d0a0-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.349055 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvmjt\" (UniqueName: \"kubernetes.io/projected/c9808e13-0b77-4493-820a-80a80111d0a0-kube-api-access-vvmjt\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.791772 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gxwnj" event={"ID":"c9808e13-0b77-4493-820a-80a80111d0a0","Type":"ContainerDied","Data":"e58376dd6256554d5dd8338928444be6b32d4ab8ff5a9caaa99ec62d32b22677"} Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.791831 4766 scope.go:117] "RemoveContainer" containerID="6f6536718602539efcf5904c3e38d9a742ea042afdaa08c209e03fbfec04e08f" Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.791970 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gxwnj" Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.800138 4766 generic.go:334] "Generic (PLEG): container finished" podID="a8031d42-e714-4bf0-9976-92a118257476" containerID="ee15729c23ecffe1160cced39eb71af4481782f53f7c036dc06718cc0b3f22c7" exitCode=0 Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.800225 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-45g8j" event={"ID":"a8031d42-e714-4bf0-9976-92a118257476","Type":"ContainerDied","Data":"ee15729c23ecffe1160cced39eb71af4481782f53f7c036dc06718cc0b3f22c7"} Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.801630 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5hq26_cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2/registry-server/0.log" Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.802422 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hq26" event={"ID":"cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2","Type":"ContainerDied","Data":"7b4c8dac3ccba57af9d731811581ed84553ce4e516437a74ba0d170825968250"} Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.802465 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b4c8dac3ccba57af9d731811581ed84553ce4e516437a74ba0d170825968250" Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.806301 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5hq26_cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2/registry-server/0.log" Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.807210 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5hq26" Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.813908 4766 scope.go:117] "RemoveContainer" containerID="db88f06b41fc55c52153c6dc469566fcb7d690952679227835cebc655c198d1a" Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.833727 4766 scope.go:117] "RemoveContainer" containerID="9957c308d5b601bc8c4bd635b7d9417fb0aeeaa0b8b5962d5f5f9aecba4d1058" Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.900743 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9808e13-0b77-4493-820a-80a80111d0a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c9808e13-0b77-4493-820a-80a80111d0a0" (UID: "c9808e13-0b77-4493-820a-80a80111d0a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.955559 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2-utilities\") pod \"cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2\" (UID: \"cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2\") " Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.955840 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9bwg\" (UniqueName: \"kubernetes.io/projected/cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2-kube-api-access-j9bwg\") pod \"cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2\" (UID: \"cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2\") " Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.956000 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2-catalog-content\") pod \"cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2\" (UID: \"cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2\") " Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.956299 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2-utilities" (OuterVolumeSpecName: "utilities") pod "cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2" (UID: "cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.956512 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9808e13-0b77-4493-820a-80a80111d0a0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.956602 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:26 crc kubenswrapper[4766]: I1126 00:27:26.958872 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2-kube-api-access-j9bwg" (OuterVolumeSpecName: "kube-api-access-j9bwg") pod "cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2" (UID: "cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2"). InnerVolumeSpecName "kube-api-access-j9bwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.007342 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2" (UID: "cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.057920 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.057947 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9bwg\" (UniqueName: \"kubernetes.io/projected/cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2-kube-api-access-j9bwg\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.121162 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gxwnj"] Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.128504 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gxwnj"] Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.667751 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tqzxf" Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.811527 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tqzxf" event={"ID":"519cd16f-3b69-4365-866d-47f01ecd0c74","Type":"ContainerDied","Data":"ea32ed1709e4b06227212ccb1ab32d85c44dab0b5cecc022cdb8a224d2e58120"} Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.811564 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tqzxf" Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.811587 4766 scope.go:117] "RemoveContainer" containerID="ade52e3c1c34cb5c1cef3967f146829c39a9b70d0794b3e04f70eca12714a566" Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.812311 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5hq26" Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.823660 4766 scope.go:117] "RemoveContainer" containerID="e103055425cfc553a7ba8f54d73774e760d7496ff55334dfb0f2b9b9f68b3352" Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.840295 4766 scope.go:117] "RemoveContainer" containerID="cdc529ca6da9251692495bb3cb6bd42b87b8e1cfc2fa2f5834ef0d2b50f2760b" Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.842260 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9808e13-0b77-4493-820a-80a80111d0a0" path="/var/lib/kubelet/pods/c9808e13-0b77-4493-820a-80a80111d0a0/volumes" Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.852641 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5hq26"] Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.852706 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5hq26"] Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.865994 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvfrt\" (UniqueName: \"kubernetes.io/projected/519cd16f-3b69-4365-866d-47f01ecd0c74-kube-api-access-wvfrt\") pod \"519cd16f-3b69-4365-866d-47f01ecd0c74\" (UID: \"519cd16f-3b69-4365-866d-47f01ecd0c74\") " Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.866453 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/519cd16f-3b69-4365-866d-47f01ecd0c74-catalog-content\") pod \"519cd16f-3b69-4365-866d-47f01ecd0c74\" (UID: \"519cd16f-3b69-4365-866d-47f01ecd0c74\") " Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.866496 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/519cd16f-3b69-4365-866d-47f01ecd0c74-utilities\") pod \"519cd16f-3b69-4365-866d-47f01ecd0c74\" (UID: \"519cd16f-3b69-4365-866d-47f01ecd0c74\") " Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.868070 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/519cd16f-3b69-4365-866d-47f01ecd0c74-utilities" (OuterVolumeSpecName: "utilities") pod "519cd16f-3b69-4365-866d-47f01ecd0c74" (UID: "519cd16f-3b69-4365-866d-47f01ecd0c74"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.869020 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/519cd16f-3b69-4365-866d-47f01ecd0c74-kube-api-access-wvfrt" (OuterVolumeSpecName: "kube-api-access-wvfrt") pod "519cd16f-3b69-4365-866d-47f01ecd0c74" (UID: "519cd16f-3b69-4365-866d-47f01ecd0c74"). InnerVolumeSpecName "kube-api-access-wvfrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.882377 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/519cd16f-3b69-4365-866d-47f01ecd0c74-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "519cd16f-3b69-4365-866d-47f01ecd0c74" (UID: "519cd16f-3b69-4365-866d-47f01ecd0c74"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.971291 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/519cd16f-3b69-4365-866d-47f01ecd0c74-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.971324 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/519cd16f-3b69-4365-866d-47f01ecd0c74-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:27 crc kubenswrapper[4766]: I1126 00:27:27.971336 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvfrt\" (UniqueName: \"kubernetes.io/projected/519cd16f-3b69-4365-866d-47f01ecd0c74-kube-api-access-wvfrt\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:28 crc kubenswrapper[4766]: I1126 00:27:28.145272 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tqzxf"] Nov 26 00:27:28 crc kubenswrapper[4766]: I1126 00:27:28.149840 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tqzxf"] Nov 26 00:27:28 crc kubenswrapper[4766]: I1126 00:27:28.327913 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-45g8j" Nov 26 00:27:28 crc kubenswrapper[4766]: I1126 00:27:28.478244 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8031d42-e714-4bf0-9976-92a118257476-catalog-content\") pod \"a8031d42-e714-4bf0-9976-92a118257476\" (UID: \"a8031d42-e714-4bf0-9976-92a118257476\") " Nov 26 00:27:28 crc kubenswrapper[4766]: I1126 00:27:28.478370 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8031d42-e714-4bf0-9976-92a118257476-utilities\") pod \"a8031d42-e714-4bf0-9976-92a118257476\" (UID: \"a8031d42-e714-4bf0-9976-92a118257476\") " Nov 26 00:27:28 crc kubenswrapper[4766]: I1126 00:27:28.478400 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tp92b\" (UniqueName: \"kubernetes.io/projected/a8031d42-e714-4bf0-9976-92a118257476-kube-api-access-tp92b\") pod \"a8031d42-e714-4bf0-9976-92a118257476\" (UID: \"a8031d42-e714-4bf0-9976-92a118257476\") " Nov 26 00:27:28 crc kubenswrapper[4766]: I1126 00:27:28.479601 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8031d42-e714-4bf0-9976-92a118257476-utilities" (OuterVolumeSpecName: "utilities") pod "a8031d42-e714-4bf0-9976-92a118257476" (UID: "a8031d42-e714-4bf0-9976-92a118257476"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:27:28 crc kubenswrapper[4766]: I1126 00:27:28.483384 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8031d42-e714-4bf0-9976-92a118257476-kube-api-access-tp92b" (OuterVolumeSpecName: "kube-api-access-tp92b") pod "a8031d42-e714-4bf0-9976-92a118257476" (UID: "a8031d42-e714-4bf0-9976-92a118257476"). InnerVolumeSpecName "kube-api-access-tp92b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:27:28 crc kubenswrapper[4766]: I1126 00:27:28.580567 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8031d42-e714-4bf0-9976-92a118257476-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:28 crc kubenswrapper[4766]: I1126 00:27:28.580629 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tp92b\" (UniqueName: \"kubernetes.io/projected/a8031d42-e714-4bf0-9976-92a118257476-kube-api-access-tp92b\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:28 crc kubenswrapper[4766]: I1126 00:27:28.591444 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8031d42-e714-4bf0-9976-92a118257476-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a8031d42-e714-4bf0-9976-92a118257476" (UID: "a8031d42-e714-4bf0-9976-92a118257476"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:27:28 crc kubenswrapper[4766]: I1126 00:27:28.681907 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8031d42-e714-4bf0-9976-92a118257476-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:28 crc kubenswrapper[4766]: I1126 00:27:28.818222 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-45g8j" event={"ID":"a8031d42-e714-4bf0-9976-92a118257476","Type":"ContainerDied","Data":"3731a3f5214a5a7c10cf92e2c05c4232197e7116fd192c64324be6a4b0c73818"} Nov 26 00:27:28 crc kubenswrapper[4766]: I1126 00:27:28.818251 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-45g8j" Nov 26 00:27:28 crc kubenswrapper[4766]: I1126 00:27:28.818843 4766 scope.go:117] "RemoveContainer" containerID="ee15729c23ecffe1160cced39eb71af4481782f53f7c036dc06718cc0b3f22c7" Nov 26 00:27:28 crc kubenswrapper[4766]: I1126 00:27:28.832790 4766 scope.go:117] "RemoveContainer" containerID="86b1bad4e51b33d503df14bd78c94023af97dc3e3e74f2521b39ca568196be22" Nov 26 00:27:28 crc kubenswrapper[4766]: I1126 00:27:28.848071 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-45g8j"] Nov 26 00:27:28 crc kubenswrapper[4766]: I1126 00:27:28.849867 4766 scope.go:117] "RemoveContainer" containerID="5531d8a3a113f10f65f7c24a63d1615c0db05b9d3c92bfdaf3bb0222bd1fa71e" Nov 26 00:27:28 crc kubenswrapper[4766]: I1126 00:27:28.852188 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-45g8j"] Nov 26 00:27:29 crc kubenswrapper[4766]: I1126 00:27:29.526936 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-tgrk8"] Nov 26 00:27:29 crc kubenswrapper[4766]: I1126 00:27:29.832693 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="519cd16f-3b69-4365-866d-47f01ecd0c74" path="/var/lib/kubelet/pods/519cd16f-3b69-4365-866d-47f01ecd0c74/volumes" Nov 26 00:27:29 crc kubenswrapper[4766]: I1126 00:27:29.834060 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8031d42-e714-4bf0-9976-92a118257476" path="/var/lib/kubelet/pods/a8031d42-e714-4bf0-9976-92a118257476/volumes" Nov 26 00:27:29 crc kubenswrapper[4766]: I1126 00:27:29.835027 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2" path="/var/lib/kubelet/pods/cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2/volumes" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.550929 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" podUID="c91e0334-6829-4a8f-9239-3930e84e65c4" containerName="oauth-openshift" containerID="cri-o://e5dd72bee270578ea299f3e401e5d84d62f7fe4ef40b724b2badeb051bc874b6" gracePeriod=15 Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.876340 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.911895 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5f95b94cb-g6rdn"] Nov 26 00:27:54 crc kubenswrapper[4766]: E1126 00:27:54.912166 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="519cd16f-3b69-4365-866d-47f01ecd0c74" containerName="registry-server" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912189 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="519cd16f-3b69-4365-866d-47f01ecd0c74" containerName="registry-server" Nov 26 00:27:54 crc kubenswrapper[4766]: E1126 00:27:54.912208 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dba56bc-b11d-4f0a-aea4-205c48a40765" containerName="pruner" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912217 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dba56bc-b11d-4f0a-aea4-205c48a40765" containerName="pruner" Nov 26 00:27:54 crc kubenswrapper[4766]: E1126 00:27:54.912230 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2" containerName="registry-server" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912239 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2" containerName="registry-server" Nov 26 00:27:54 crc kubenswrapper[4766]: E1126 00:27:54.912251 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9808e13-0b77-4493-820a-80a80111d0a0" containerName="registry-server" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912260 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9808e13-0b77-4493-820a-80a80111d0a0" containerName="registry-server" Nov 26 00:27:54 crc kubenswrapper[4766]: E1126 00:27:54.912275 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="284847a7-94c0-4f03-a54d-6d22fe77353f" containerName="image-pruner" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912283 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="284847a7-94c0-4f03-a54d-6d22fe77353f" containerName="image-pruner" Nov 26 00:27:54 crc kubenswrapper[4766]: E1126 00:27:54.912298 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8031d42-e714-4bf0-9976-92a118257476" containerName="registry-server" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912310 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8031d42-e714-4bf0-9976-92a118257476" containerName="registry-server" Nov 26 00:27:54 crc kubenswrapper[4766]: E1126 00:27:54.912323 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9808e13-0b77-4493-820a-80a80111d0a0" containerName="extract-content" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912334 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9808e13-0b77-4493-820a-80a80111d0a0" containerName="extract-content" Nov 26 00:27:54 crc kubenswrapper[4766]: E1126 00:27:54.912347 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c91e0334-6829-4a8f-9239-3930e84e65c4" containerName="oauth-openshift" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912357 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c91e0334-6829-4a8f-9239-3930e84e65c4" containerName="oauth-openshift" Nov 26 00:27:54 crc kubenswrapper[4766]: E1126 00:27:54.912372 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8031d42-e714-4bf0-9976-92a118257476" containerName="extract-content" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912382 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8031d42-e714-4bf0-9976-92a118257476" containerName="extract-content" Nov 26 00:27:54 crc kubenswrapper[4766]: E1126 00:27:54.912395 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="519cd16f-3b69-4365-866d-47f01ecd0c74" containerName="extract-utilities" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912403 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="519cd16f-3b69-4365-866d-47f01ecd0c74" containerName="extract-utilities" Nov 26 00:27:54 crc kubenswrapper[4766]: E1126 00:27:54.912416 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8031d42-e714-4bf0-9976-92a118257476" containerName="extract-utilities" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912425 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8031d42-e714-4bf0-9976-92a118257476" containerName="extract-utilities" Nov 26 00:27:54 crc kubenswrapper[4766]: E1126 00:27:54.912433 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9808e13-0b77-4493-820a-80a80111d0a0" containerName="extract-utilities" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912441 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9808e13-0b77-4493-820a-80a80111d0a0" containerName="extract-utilities" Nov 26 00:27:54 crc kubenswrapper[4766]: E1126 00:27:54.912454 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93c919b9-1485-42fd-a0bd-0608a608f926" containerName="pruner" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912462 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="93c919b9-1485-42fd-a0bd-0608a608f926" containerName="pruner" Nov 26 00:27:54 crc kubenswrapper[4766]: E1126 00:27:54.912474 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2" containerName="extract-utilities" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912482 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2" containerName="extract-utilities" Nov 26 00:27:54 crc kubenswrapper[4766]: E1126 00:27:54.912492 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="519cd16f-3b69-4365-866d-47f01ecd0c74" containerName="extract-content" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912500 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="519cd16f-3b69-4365-866d-47f01ecd0c74" containerName="extract-content" Nov 26 00:27:54 crc kubenswrapper[4766]: E1126 00:27:54.912511 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2" containerName="extract-content" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912519 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2" containerName="extract-content" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912635 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dba56bc-b11d-4f0a-aea4-205c48a40765" containerName="pruner" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912692 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="519cd16f-3b69-4365-866d-47f01ecd0c74" containerName="registry-server" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912708 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd349f2f-fc2d-4cbd-bbb3-d1f028bde5d2" containerName="registry-server" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912718 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8031d42-e714-4bf0-9976-92a118257476" containerName="registry-server" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912726 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9808e13-0b77-4493-820a-80a80111d0a0" containerName="registry-server" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912737 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="284847a7-94c0-4f03-a54d-6d22fe77353f" containerName="image-pruner" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912747 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="c91e0334-6829-4a8f-9239-3930e84e65c4" containerName="oauth-openshift" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.912758 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="93c919b9-1485-42fd-a0bd-0608a608f926" containerName="pruner" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.913243 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.915418 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5f95b94cb-g6rdn"] Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.951382 4766 generic.go:334] "Generic (PLEG): container finished" podID="c91e0334-6829-4a8f-9239-3930e84e65c4" containerID="e5dd72bee270578ea299f3e401e5d84d62f7fe4ef40b724b2badeb051bc874b6" exitCode=0 Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.951425 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" event={"ID":"c91e0334-6829-4a8f-9239-3930e84e65c4","Type":"ContainerDied","Data":"e5dd72bee270578ea299f3e401e5d84d62f7fe4ef40b724b2badeb051bc874b6"} Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.951451 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" event={"ID":"c91e0334-6829-4a8f-9239-3930e84e65c4","Type":"ContainerDied","Data":"34e8e4dbbc1668d50f4d50aa8d3603a15dbeb2018309e8cb329660751898cc06"} Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.951467 4766 scope.go:117] "RemoveContainer" containerID="e5dd72bee270578ea299f3e401e5d84d62f7fe4ef40b724b2badeb051bc874b6" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.951570 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-tgrk8" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.975520 4766 scope.go:117] "RemoveContainer" containerID="e5dd72bee270578ea299f3e401e5d84d62f7fe4ef40b724b2badeb051bc874b6" Nov 26 00:27:54 crc kubenswrapper[4766]: E1126 00:27:54.976037 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5dd72bee270578ea299f3e401e5d84d62f7fe4ef40b724b2badeb051bc874b6\": container with ID starting with e5dd72bee270578ea299f3e401e5d84d62f7fe4ef40b724b2badeb051bc874b6 not found: ID does not exist" containerID="e5dd72bee270578ea299f3e401e5d84d62f7fe4ef40b724b2badeb051bc874b6" Nov 26 00:27:54 crc kubenswrapper[4766]: I1126 00:27:54.976076 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5dd72bee270578ea299f3e401e5d84d62f7fe4ef40b724b2badeb051bc874b6"} err="failed to get container status \"e5dd72bee270578ea299f3e401e5d84d62f7fe4ef40b724b2badeb051bc874b6\": rpc error: code = NotFound desc = could not find container \"e5dd72bee270578ea299f3e401e5d84d62f7fe4ef40b724b2badeb051bc874b6\": container with ID starting with e5dd72bee270578ea299f3e401e5d84d62f7fe4ef40b724b2badeb051bc874b6 not found: ID does not exist" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.005443 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-template-provider-selection\") pod \"c91e0334-6829-4a8f-9239-3930e84e65c4\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.005483 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-audit-policies\") pod \"c91e0334-6829-4a8f-9239-3930e84e65c4\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.005506 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-serving-cert\") pod \"c91e0334-6829-4a8f-9239-3930e84e65c4\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.005527 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-template-login\") pod \"c91e0334-6829-4a8f-9239-3930e84e65c4\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.005554 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-service-ca\") pod \"c91e0334-6829-4a8f-9239-3930e84e65c4\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.005590 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-trusted-ca-bundle\") pod \"c91e0334-6829-4a8f-9239-3930e84e65c4\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.005745 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-session\") pod \"c91e0334-6829-4a8f-9239-3930e84e65c4\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.005776 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-idp-0-file-data\") pod \"c91e0334-6829-4a8f-9239-3930e84e65c4\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.005796 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-ocp-branding-template\") pod \"c91e0334-6829-4a8f-9239-3930e84e65c4\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.005821 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-router-certs\") pod \"c91e0334-6829-4a8f-9239-3930e84e65c4\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.005842 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-template-error\") pod \"c91e0334-6829-4a8f-9239-3930e84e65c4\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.005857 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c91e0334-6829-4a8f-9239-3930e84e65c4-audit-dir\") pod \"c91e0334-6829-4a8f-9239-3930e84e65c4\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.005872 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbztm\" (UniqueName: \"kubernetes.io/projected/c91e0334-6829-4a8f-9239-3930e84e65c4-kube-api-access-sbztm\") pod \"c91e0334-6829-4a8f-9239-3930e84e65c4\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.005890 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-cliconfig\") pod \"c91e0334-6829-4a8f-9239-3930e84e65c4\" (UID: \"c91e0334-6829-4a8f-9239-3930e84e65c4\") " Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.006511 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "c91e0334-6829-4a8f-9239-3930e84e65c4" (UID: "c91e0334-6829-4a8f-9239-3930e84e65c4"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.006744 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c91e0334-6829-4a8f-9239-3930e84e65c4-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "c91e0334-6829-4a8f-9239-3930e84e65c4" (UID: "c91e0334-6829-4a8f-9239-3930e84e65c4"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.006718 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "c91e0334-6829-4a8f-9239-3930e84e65c4" (UID: "c91e0334-6829-4a8f-9239-3930e84e65c4"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.006947 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "c91e0334-6829-4a8f-9239-3930e84e65c4" (UID: "c91e0334-6829-4a8f-9239-3930e84e65c4"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.007318 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "c91e0334-6829-4a8f-9239-3930e84e65c4" (UID: "c91e0334-6829-4a8f-9239-3930e84e65c4"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.011104 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "c91e0334-6829-4a8f-9239-3930e84e65c4" (UID: "c91e0334-6829-4a8f-9239-3930e84e65c4"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.012007 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "c91e0334-6829-4a8f-9239-3930e84e65c4" (UID: "c91e0334-6829-4a8f-9239-3930e84e65c4"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.012022 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c91e0334-6829-4a8f-9239-3930e84e65c4-kube-api-access-sbztm" (OuterVolumeSpecName: "kube-api-access-sbztm") pod "c91e0334-6829-4a8f-9239-3930e84e65c4" (UID: "c91e0334-6829-4a8f-9239-3930e84e65c4"). InnerVolumeSpecName "kube-api-access-sbztm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.012809 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "c91e0334-6829-4a8f-9239-3930e84e65c4" (UID: "c91e0334-6829-4a8f-9239-3930e84e65c4"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.012923 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "c91e0334-6829-4a8f-9239-3930e84e65c4" (UID: "c91e0334-6829-4a8f-9239-3930e84e65c4"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.013150 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "c91e0334-6829-4a8f-9239-3930e84e65c4" (UID: "c91e0334-6829-4a8f-9239-3930e84e65c4"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.013193 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "c91e0334-6829-4a8f-9239-3930e84e65c4" (UID: "c91e0334-6829-4a8f-9239-3930e84e65c4"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.013440 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "c91e0334-6829-4a8f-9239-3930e84e65c4" (UID: "c91e0334-6829-4a8f-9239-3930e84e65c4"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.022977 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "c91e0334-6829-4a8f-9239-3930e84e65c4" (UID: "c91e0334-6829-4a8f-9239-3930e84e65c4"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.107119 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-user-template-login\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.107229 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c2131e10-47d2-4d48-a776-c6363d32972a-audit-policies\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.107284 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.107341 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqdkn\" (UniqueName: \"kubernetes.io/projected/c2131e10-47d2-4d48-a776-c6363d32972a-kube-api-access-kqdkn\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.107381 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.107478 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.107513 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c2131e10-47d2-4d48-a776-c6363d32972a-audit-dir\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.107554 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-user-template-error\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.107681 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.107736 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-system-service-ca\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.107832 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.107885 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-system-router-certs\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.107927 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.108084 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-system-session\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.108147 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.108166 4766 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.108181 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.108193 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.108205 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.108219 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.108232 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.108245 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.108259 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.108272 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.108285 4766 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c91e0334-6829-4a8f-9239-3930e84e65c4-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.108297 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.108311 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbztm\" (UniqueName: \"kubernetes.io/projected/c91e0334-6829-4a8f-9239-3930e84e65c4-kube-api-access-sbztm\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.108322 4766 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c91e0334-6829-4a8f-9239-3930e84e65c4-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.209499 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-system-router-certs\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.209569 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.209609 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-system-session\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.209712 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-user-template-login\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.209769 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c2131e10-47d2-4d48-a776-c6363d32972a-audit-policies\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.209870 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.209957 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqdkn\" (UniqueName: \"kubernetes.io/projected/c2131e10-47d2-4d48-a776-c6363d32972a-kube-api-access-kqdkn\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.210081 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.210153 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.210204 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c2131e10-47d2-4d48-a776-c6363d32972a-audit-dir\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.210266 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-user-template-error\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.210332 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.210390 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-system-service-ca\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.210405 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c2131e10-47d2-4d48-a776-c6363d32972a-audit-dir\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.210478 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.211946 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.212034 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c2131e10-47d2-4d48-a776-c6363d32972a-audit-policies\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.212700 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-system-service-ca\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.213127 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.214305 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-system-session\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.214367 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-user-template-login\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.215047 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-system-router-certs\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.215330 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.215372 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.215785 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.216341 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-user-template-error\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.216346 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2131e10-47d2-4d48-a776-c6363d32972a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.230507 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqdkn\" (UniqueName: \"kubernetes.io/projected/c2131e10-47d2-4d48-a776-c6363d32972a-kube-api-access-kqdkn\") pod \"oauth-openshift-5f95b94cb-g6rdn\" (UID: \"c2131e10-47d2-4d48-a776-c6363d32972a\") " pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.290685 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-tgrk8"] Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.296720 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-tgrk8"] Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.529598 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.761608 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5f95b94cb-g6rdn"] Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.833084 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c91e0334-6829-4a8f-9239-3930e84e65c4" path="/var/lib/kubelet/pods/c91e0334-6829-4a8f-9239-3930e84e65c4/volumes" Nov 26 00:27:55 crc kubenswrapper[4766]: I1126 00:27:55.959053 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" event={"ID":"c2131e10-47d2-4d48-a776-c6363d32972a","Type":"ContainerStarted","Data":"300448352a022bff75b0d9d51e77ee836bd71aa9877f2e3a5172b8ace9d1d3e8"} Nov 26 00:27:56 crc kubenswrapper[4766]: I1126 00:27:56.965814 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" event={"ID":"c2131e10-47d2-4d48-a776-c6363d32972a","Type":"ContainerStarted","Data":"e720ac813e4b19c07762f63874d1a51e31363b1858e61ab681b062a404a26752"} Nov 26 00:27:56 crc kubenswrapper[4766]: I1126 00:27:56.966090 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:56 crc kubenswrapper[4766]: I1126 00:27:56.975797 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" Nov 26 00:27:56 crc kubenswrapper[4766]: I1126 00:27:56.987589 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5f95b94cb-g6rdn" podStartSLOduration=27.987568699 podStartE2EDuration="27.987568699s" podCreationTimestamp="2025-11-26 00:27:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:27:56.983180837 +0000 UTC m=+257.831951287" watchObservedRunningTime="2025-11-26 00:27:56.987568699 +0000 UTC m=+257.836339129" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.338864 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bjz9v"] Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.339577 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bjz9v" podUID="e0d4c85c-2a27-4a2f-903e-8f7473351104" containerName="registry-server" containerID="cri-o://5a5cd04a55e765bfe4abeea8753044a245db57f489369162d3e283d02e8cc241" gracePeriod=30 Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.343150 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jnqdt"] Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.343381 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jnqdt" podUID="90830d76-8bef-4c13-b127-23fff3e32770" containerName="registry-server" containerID="cri-o://88933fddc0f7bbceaeeaf1000987d3306168ed1c04a17ee51b95a84b87a57e90" gracePeriod=30 Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.348636 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2vsl7"] Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.349143 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" podUID="f1e5e774-51ea-47e8-8b02-1bde1728da35" containerName="marketplace-operator" containerID="cri-o://11bd5ba6a7c03a5a475d63b231ab4a3a247b539d5b4b9b0fab89e3ee729c88b1" gracePeriod=30 Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.370349 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fhlnl"] Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.370635 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fhlnl" podUID="1529f474-493a-485a-80b7-ee8ec1c79ccc" containerName="registry-server" containerID="cri-o://7bbaf76b46cc33ba0b12824530776175690a99499a1ca6cdce26885e22214afe" gracePeriod=30 Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.373723 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2xsk5"] Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.376877 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2xsk5" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.379605 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2xsk5"] Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.391283 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s9x5n"] Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.391525 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-s9x5n" podUID="5f6c20b5-b168-4529-b84a-bef5c19254d5" containerName="registry-server" containerID="cri-o://589b3e5703712a442130255630f9f46e0f819c363e3eb4fd4223fae322d7fdc5" gracePeriod=30 Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.448299 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cad040b4-5d74-4cd0-b726-93d9c6cb73e2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2xsk5\" (UID: \"cad040b4-5d74-4cd0-b726-93d9c6cb73e2\") " pod="openshift-marketplace/marketplace-operator-79b997595-2xsk5" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.448418 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cad040b4-5d74-4cd0-b726-93d9c6cb73e2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2xsk5\" (UID: \"cad040b4-5d74-4cd0-b726-93d9c6cb73e2\") " pod="openshift-marketplace/marketplace-operator-79b997595-2xsk5" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.448477 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjlns\" (UniqueName: \"kubernetes.io/projected/cad040b4-5d74-4cd0-b726-93d9c6cb73e2-kube-api-access-mjlns\") pod \"marketplace-operator-79b997595-2xsk5\" (UID: \"cad040b4-5d74-4cd0-b726-93d9c6cb73e2\") " pod="openshift-marketplace/marketplace-operator-79b997595-2xsk5" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.550406 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjlns\" (UniqueName: \"kubernetes.io/projected/cad040b4-5d74-4cd0-b726-93d9c6cb73e2-kube-api-access-mjlns\") pod \"marketplace-operator-79b997595-2xsk5\" (UID: \"cad040b4-5d74-4cd0-b726-93d9c6cb73e2\") " pod="openshift-marketplace/marketplace-operator-79b997595-2xsk5" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.550537 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cad040b4-5d74-4cd0-b726-93d9c6cb73e2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2xsk5\" (UID: \"cad040b4-5d74-4cd0-b726-93d9c6cb73e2\") " pod="openshift-marketplace/marketplace-operator-79b997595-2xsk5" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.550675 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cad040b4-5d74-4cd0-b726-93d9c6cb73e2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2xsk5\" (UID: \"cad040b4-5d74-4cd0-b726-93d9c6cb73e2\") " pod="openshift-marketplace/marketplace-operator-79b997595-2xsk5" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.552131 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cad040b4-5d74-4cd0-b726-93d9c6cb73e2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2xsk5\" (UID: \"cad040b4-5d74-4cd0-b726-93d9c6cb73e2\") " pod="openshift-marketplace/marketplace-operator-79b997595-2xsk5" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.557795 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cad040b4-5d74-4cd0-b726-93d9c6cb73e2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2xsk5\" (UID: \"cad040b4-5d74-4cd0-b726-93d9c6cb73e2\") " pod="openshift-marketplace/marketplace-operator-79b997595-2xsk5" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.573621 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjlns\" (UniqueName: \"kubernetes.io/projected/cad040b4-5d74-4cd0-b726-93d9c6cb73e2-kube-api-access-mjlns\") pod \"marketplace-operator-79b997595-2xsk5\" (UID: \"cad040b4-5d74-4cd0-b726-93d9c6cb73e2\") " pod="openshift-marketplace/marketplace-operator-79b997595-2xsk5" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.738721 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2xsk5" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.760093 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bjz9v" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.789152 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jnqdt" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.789790 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fhlnl" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.815412 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s9x5n" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.825139 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.958974 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1529f474-493a-485a-80b7-ee8ec1c79ccc-catalog-content\") pod \"1529f474-493a-485a-80b7-ee8ec1c79ccc\" (UID: \"1529f474-493a-485a-80b7-ee8ec1c79ccc\") " Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.959046 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bds6d\" (UniqueName: \"kubernetes.io/projected/f1e5e774-51ea-47e8-8b02-1bde1728da35-kube-api-access-bds6d\") pod \"f1e5e774-51ea-47e8-8b02-1bde1728da35\" (UID: \"f1e5e774-51ea-47e8-8b02-1bde1728da35\") " Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.959073 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1529f474-493a-485a-80b7-ee8ec1c79ccc-utilities\") pod \"1529f474-493a-485a-80b7-ee8ec1c79ccc\" (UID: \"1529f474-493a-485a-80b7-ee8ec1c79ccc\") " Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.959099 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bwtw\" (UniqueName: \"kubernetes.io/projected/90830d76-8bef-4c13-b127-23fff3e32770-kube-api-access-9bwtw\") pod \"90830d76-8bef-4c13-b127-23fff3e32770\" (UID: \"90830d76-8bef-4c13-b127-23fff3e32770\") " Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.959118 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0d4c85c-2a27-4a2f-903e-8f7473351104-catalog-content\") pod \"e0d4c85c-2a27-4a2f-903e-8f7473351104\" (UID: \"e0d4c85c-2a27-4a2f-903e-8f7473351104\") " Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.959132 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0d4c85c-2a27-4a2f-903e-8f7473351104-utilities\") pod \"e0d4c85c-2a27-4a2f-903e-8f7473351104\" (UID: \"e0d4c85c-2a27-4a2f-903e-8f7473351104\") " Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.959157 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90830d76-8bef-4c13-b127-23fff3e32770-utilities\") pod \"90830d76-8bef-4c13-b127-23fff3e32770\" (UID: \"90830d76-8bef-4c13-b127-23fff3e32770\") " Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.959943 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2xsk5"] Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.960012 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90830d76-8bef-4c13-b127-23fff3e32770-utilities" (OuterVolumeSpecName: "utilities") pod "90830d76-8bef-4c13-b127-23fff3e32770" (UID: "90830d76-8bef-4c13-b127-23fff3e32770"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.959187 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f1e5e774-51ea-47e8-8b02-1bde1728da35-marketplace-operator-metrics\") pod \"f1e5e774-51ea-47e8-8b02-1bde1728da35\" (UID: \"f1e5e774-51ea-47e8-8b02-1bde1728da35\") " Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.960037 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1529f474-493a-485a-80b7-ee8ec1c79ccc-utilities" (OuterVolumeSpecName: "utilities") pod "1529f474-493a-485a-80b7-ee8ec1c79ccc" (UID: "1529f474-493a-485a-80b7-ee8ec1c79ccc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.960069 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f1e5e774-51ea-47e8-8b02-1bde1728da35-marketplace-trusted-ca\") pod \"f1e5e774-51ea-47e8-8b02-1bde1728da35\" (UID: \"f1e5e774-51ea-47e8-8b02-1bde1728da35\") " Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.960064 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0d4c85c-2a27-4a2f-903e-8f7473351104-utilities" (OuterVolumeSpecName: "utilities") pod "e0d4c85c-2a27-4a2f-903e-8f7473351104" (UID: "e0d4c85c-2a27-4a2f-903e-8f7473351104"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.960123 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgnbg\" (UniqueName: \"kubernetes.io/projected/e0d4c85c-2a27-4a2f-903e-8f7473351104-kube-api-access-lgnbg\") pod \"e0d4c85c-2a27-4a2f-903e-8f7473351104\" (UID: \"e0d4c85c-2a27-4a2f-903e-8f7473351104\") " Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.960142 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f6c20b5-b168-4529-b84a-bef5c19254d5-utilities\") pod \"5f6c20b5-b168-4529-b84a-bef5c19254d5\" (UID: \"5f6c20b5-b168-4529-b84a-bef5c19254d5\") " Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.960176 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xktn8\" (UniqueName: \"kubernetes.io/projected/1529f474-493a-485a-80b7-ee8ec1c79ccc-kube-api-access-xktn8\") pod \"1529f474-493a-485a-80b7-ee8ec1c79ccc\" (UID: \"1529f474-493a-485a-80b7-ee8ec1c79ccc\") " Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.960203 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2z5vl\" (UniqueName: \"kubernetes.io/projected/5f6c20b5-b168-4529-b84a-bef5c19254d5-kube-api-access-2z5vl\") pod \"5f6c20b5-b168-4529-b84a-bef5c19254d5\" (UID: \"5f6c20b5-b168-4529-b84a-bef5c19254d5\") " Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.960223 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f6c20b5-b168-4529-b84a-bef5c19254d5-catalog-content\") pod \"5f6c20b5-b168-4529-b84a-bef5c19254d5\" (UID: \"5f6c20b5-b168-4529-b84a-bef5c19254d5\") " Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.960243 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90830d76-8bef-4c13-b127-23fff3e32770-catalog-content\") pod \"90830d76-8bef-4c13-b127-23fff3e32770\" (UID: \"90830d76-8bef-4c13-b127-23fff3e32770\") " Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.960937 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90830d76-8bef-4c13-b127-23fff3e32770-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.960959 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1529f474-493a-485a-80b7-ee8ec1c79ccc-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.960967 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0d4c85c-2a27-4a2f-903e-8f7473351104-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.960960 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1e5e774-51ea-47e8-8b02-1bde1728da35-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "f1e5e774-51ea-47e8-8b02-1bde1728da35" (UID: "f1e5e774-51ea-47e8-8b02-1bde1728da35"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.962029 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f6c20b5-b168-4529-b84a-bef5c19254d5-utilities" (OuterVolumeSpecName: "utilities") pod "5f6c20b5-b168-4529-b84a-bef5c19254d5" (UID: "5f6c20b5-b168-4529-b84a-bef5c19254d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.968104 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1e5e774-51ea-47e8-8b02-1bde1728da35-kube-api-access-bds6d" (OuterVolumeSpecName: "kube-api-access-bds6d") pod "f1e5e774-51ea-47e8-8b02-1bde1728da35" (UID: "f1e5e774-51ea-47e8-8b02-1bde1728da35"). InnerVolumeSpecName "kube-api-access-bds6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.977490 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0d4c85c-2a27-4a2f-903e-8f7473351104-kube-api-access-lgnbg" (OuterVolumeSpecName: "kube-api-access-lgnbg") pod "e0d4c85c-2a27-4a2f-903e-8f7473351104" (UID: "e0d4c85c-2a27-4a2f-903e-8f7473351104"). InnerVolumeSpecName "kube-api-access-lgnbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.977785 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1e5e774-51ea-47e8-8b02-1bde1728da35-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "f1e5e774-51ea-47e8-8b02-1bde1728da35" (UID: "f1e5e774-51ea-47e8-8b02-1bde1728da35"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.977561 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f6c20b5-b168-4529-b84a-bef5c19254d5-kube-api-access-2z5vl" (OuterVolumeSpecName: "kube-api-access-2z5vl") pod "5f6c20b5-b168-4529-b84a-bef5c19254d5" (UID: "5f6c20b5-b168-4529-b84a-bef5c19254d5"). InnerVolumeSpecName "kube-api-access-2z5vl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.978003 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90830d76-8bef-4c13-b127-23fff3e32770-kube-api-access-9bwtw" (OuterVolumeSpecName: "kube-api-access-9bwtw") pod "90830d76-8bef-4c13-b127-23fff3e32770" (UID: "90830d76-8bef-4c13-b127-23fff3e32770"). InnerVolumeSpecName "kube-api-access-9bwtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:28:16 crc kubenswrapper[4766]: I1126 00:28:16.984292 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1529f474-493a-485a-80b7-ee8ec1c79ccc-kube-api-access-xktn8" (OuterVolumeSpecName: "kube-api-access-xktn8") pod "1529f474-493a-485a-80b7-ee8ec1c79ccc" (UID: "1529f474-493a-485a-80b7-ee8ec1c79ccc"). InnerVolumeSpecName "kube-api-access-xktn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.004478 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1529f474-493a-485a-80b7-ee8ec1c79ccc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1529f474-493a-485a-80b7-ee8ec1c79ccc" (UID: "1529f474-493a-485a-80b7-ee8ec1c79ccc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.017425 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0d4c85c-2a27-4a2f-903e-8f7473351104-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0d4c85c-2a27-4a2f-903e-8f7473351104" (UID: "e0d4c85c-2a27-4a2f-903e-8f7473351104"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.031942 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90830d76-8bef-4c13-b127-23fff3e32770-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "90830d76-8bef-4c13-b127-23fff3e32770" (UID: "90830d76-8bef-4c13-b127-23fff3e32770"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.062356 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgnbg\" (UniqueName: \"kubernetes.io/projected/e0d4c85c-2a27-4a2f-903e-8f7473351104-kube-api-access-lgnbg\") on node \"crc\" DevicePath \"\"" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.062389 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f6c20b5-b168-4529-b84a-bef5c19254d5-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.062402 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xktn8\" (UniqueName: \"kubernetes.io/projected/1529f474-493a-485a-80b7-ee8ec1c79ccc-kube-api-access-xktn8\") on node \"crc\" DevicePath \"\"" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.062414 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2z5vl\" (UniqueName: \"kubernetes.io/projected/5f6c20b5-b168-4529-b84a-bef5c19254d5-kube-api-access-2z5vl\") on node \"crc\" DevicePath \"\"" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.062428 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90830d76-8bef-4c13-b127-23fff3e32770-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.062440 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1529f474-493a-485a-80b7-ee8ec1c79ccc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.062453 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bds6d\" (UniqueName: \"kubernetes.io/projected/f1e5e774-51ea-47e8-8b02-1bde1728da35-kube-api-access-bds6d\") on node \"crc\" DevicePath \"\"" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.062466 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0d4c85c-2a27-4a2f-903e-8f7473351104-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.062477 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bwtw\" (UniqueName: \"kubernetes.io/projected/90830d76-8bef-4c13-b127-23fff3e32770-kube-api-access-9bwtw\") on node \"crc\" DevicePath \"\"" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.062489 4766 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f1e5e774-51ea-47e8-8b02-1bde1728da35-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.062500 4766 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f1e5e774-51ea-47e8-8b02-1bde1728da35-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.077875 4766 generic.go:334] "Generic (PLEG): container finished" podID="f1e5e774-51ea-47e8-8b02-1bde1728da35" containerID="11bd5ba6a7c03a5a475d63b231ab4a3a247b539d5b4b9b0fab89e3ee729c88b1" exitCode=0 Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.077943 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.077962 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" event={"ID":"f1e5e774-51ea-47e8-8b02-1bde1728da35","Type":"ContainerDied","Data":"11bd5ba6a7c03a5a475d63b231ab4a3a247b539d5b4b9b0fab89e3ee729c88b1"} Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.078855 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2vsl7" event={"ID":"f1e5e774-51ea-47e8-8b02-1bde1728da35","Type":"ContainerDied","Data":"a4a489bbb6a0d52c03da5c3b0e4e949f975dd481d0e735b14dfd4acfd638ceac"} Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.078876 4766 scope.go:117] "RemoveContainer" containerID="11bd5ba6a7c03a5a475d63b231ab4a3a247b539d5b4b9b0fab89e3ee729c88b1" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.081223 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2xsk5" event={"ID":"cad040b4-5d74-4cd0-b726-93d9c6cb73e2","Type":"ContainerStarted","Data":"68f860a3e56cf5f51872a8a8c2294c12e0e2e3de8fe3c34350307a7546c6db7c"} Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.084207 4766 generic.go:334] "Generic (PLEG): container finished" podID="e0d4c85c-2a27-4a2f-903e-8f7473351104" containerID="5a5cd04a55e765bfe4abeea8753044a245db57f489369162d3e283d02e8cc241" exitCode=0 Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.084275 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bjz9v" event={"ID":"e0d4c85c-2a27-4a2f-903e-8f7473351104","Type":"ContainerDied","Data":"5a5cd04a55e765bfe4abeea8753044a245db57f489369162d3e283d02e8cc241"} Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.084303 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bjz9v" event={"ID":"e0d4c85c-2a27-4a2f-903e-8f7473351104","Type":"ContainerDied","Data":"20d598dd11a975425eeaf8b05e42fc9c384e1d1a16ae8d144664055124a3cb6e"} Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.084379 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bjz9v" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.088958 4766 generic.go:334] "Generic (PLEG): container finished" podID="90830d76-8bef-4c13-b127-23fff3e32770" containerID="88933fddc0f7bbceaeeaf1000987d3306168ed1c04a17ee51b95a84b87a57e90" exitCode=0 Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.089046 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jnqdt" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.089076 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnqdt" event={"ID":"90830d76-8bef-4c13-b127-23fff3e32770","Type":"ContainerDied","Data":"88933fddc0f7bbceaeeaf1000987d3306168ed1c04a17ee51b95a84b87a57e90"} Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.089112 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnqdt" event={"ID":"90830d76-8bef-4c13-b127-23fff3e32770","Type":"ContainerDied","Data":"8807f756df08ca1317bcb38ca901f5cc3389d1ef3280a62a8384a2c4ac3e7ba1"} Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.091523 4766 generic.go:334] "Generic (PLEG): container finished" podID="5f6c20b5-b168-4529-b84a-bef5c19254d5" containerID="589b3e5703712a442130255630f9f46e0f819c363e3eb4fd4223fae322d7fdc5" exitCode=0 Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.091699 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9x5n" event={"ID":"5f6c20b5-b168-4529-b84a-bef5c19254d5","Type":"ContainerDied","Data":"589b3e5703712a442130255630f9f46e0f819c363e3eb4fd4223fae322d7fdc5"} Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.091734 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9x5n" event={"ID":"5f6c20b5-b168-4529-b84a-bef5c19254d5","Type":"ContainerDied","Data":"3308cfeb2a0fdbbbf0283eff8f9ea60ac64f9e9e83c0b5391da6560400a0b636"} Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.091783 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s9x5n" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.094295 4766 generic.go:334] "Generic (PLEG): container finished" podID="1529f474-493a-485a-80b7-ee8ec1c79ccc" containerID="7bbaf76b46cc33ba0b12824530776175690a99499a1ca6cdce26885e22214afe" exitCode=0 Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.094322 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fhlnl" event={"ID":"1529f474-493a-485a-80b7-ee8ec1c79ccc","Type":"ContainerDied","Data":"7bbaf76b46cc33ba0b12824530776175690a99499a1ca6cdce26885e22214afe"} Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.094343 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fhlnl" event={"ID":"1529f474-493a-485a-80b7-ee8ec1c79ccc","Type":"ContainerDied","Data":"ea2f0265f358576113f615350961c4fe6a0457d01dbb85040661ef54f33fe153"} Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.094396 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fhlnl" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.099091 4766 scope.go:117] "RemoveContainer" containerID="11bd5ba6a7c03a5a475d63b231ab4a3a247b539d5b4b9b0fab89e3ee729c88b1" Nov 26 00:28:17 crc kubenswrapper[4766]: E1126 00:28:17.100454 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11bd5ba6a7c03a5a475d63b231ab4a3a247b539d5b4b9b0fab89e3ee729c88b1\": container with ID starting with 11bd5ba6a7c03a5a475d63b231ab4a3a247b539d5b4b9b0fab89e3ee729c88b1 not found: ID does not exist" containerID="11bd5ba6a7c03a5a475d63b231ab4a3a247b539d5b4b9b0fab89e3ee729c88b1" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.100584 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11bd5ba6a7c03a5a475d63b231ab4a3a247b539d5b4b9b0fab89e3ee729c88b1"} err="failed to get container status \"11bd5ba6a7c03a5a475d63b231ab4a3a247b539d5b4b9b0fab89e3ee729c88b1\": rpc error: code = NotFound desc = could not find container \"11bd5ba6a7c03a5a475d63b231ab4a3a247b539d5b4b9b0fab89e3ee729c88b1\": container with ID starting with 11bd5ba6a7c03a5a475d63b231ab4a3a247b539d5b4b9b0fab89e3ee729c88b1 not found: ID does not exist" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.100828 4766 scope.go:117] "RemoveContainer" containerID="5a5cd04a55e765bfe4abeea8753044a245db57f489369162d3e283d02e8cc241" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.111293 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f6c20b5-b168-4529-b84a-bef5c19254d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f6c20b5-b168-4529-b84a-bef5c19254d5" (UID: "5f6c20b5-b168-4529-b84a-bef5c19254d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.119834 4766 scope.go:117] "RemoveContainer" containerID="834a1c0046413df441fd1d31eb8b082de3727f22cabb3144a9f0029dc6fe4984" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.124948 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2vsl7"] Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.127682 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2vsl7"] Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.138865 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jnqdt"] Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.142260 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jnqdt"] Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.148958 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fhlnl"] Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.149178 4766 scope.go:117] "RemoveContainer" containerID="9bcbabef2bd02fa775ac02940096c80cb7ef049913d92fa23bbfdd938502f110" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.154368 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fhlnl"] Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.163443 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f6c20b5-b168-4529-b84a-bef5c19254d5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.164230 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bjz9v"] Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.168144 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bjz9v"] Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.169070 4766 scope.go:117] "RemoveContainer" containerID="5a5cd04a55e765bfe4abeea8753044a245db57f489369162d3e283d02e8cc241" Nov 26 00:28:17 crc kubenswrapper[4766]: E1126 00:28:17.169913 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a5cd04a55e765bfe4abeea8753044a245db57f489369162d3e283d02e8cc241\": container with ID starting with 5a5cd04a55e765bfe4abeea8753044a245db57f489369162d3e283d02e8cc241 not found: ID does not exist" containerID="5a5cd04a55e765bfe4abeea8753044a245db57f489369162d3e283d02e8cc241" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.169970 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a5cd04a55e765bfe4abeea8753044a245db57f489369162d3e283d02e8cc241"} err="failed to get container status \"5a5cd04a55e765bfe4abeea8753044a245db57f489369162d3e283d02e8cc241\": rpc error: code = NotFound desc = could not find container \"5a5cd04a55e765bfe4abeea8753044a245db57f489369162d3e283d02e8cc241\": container with ID starting with 5a5cd04a55e765bfe4abeea8753044a245db57f489369162d3e283d02e8cc241 not found: ID does not exist" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.170012 4766 scope.go:117] "RemoveContainer" containerID="834a1c0046413df441fd1d31eb8b082de3727f22cabb3144a9f0029dc6fe4984" Nov 26 00:28:17 crc kubenswrapper[4766]: E1126 00:28:17.170453 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"834a1c0046413df441fd1d31eb8b082de3727f22cabb3144a9f0029dc6fe4984\": container with ID starting with 834a1c0046413df441fd1d31eb8b082de3727f22cabb3144a9f0029dc6fe4984 not found: ID does not exist" containerID="834a1c0046413df441fd1d31eb8b082de3727f22cabb3144a9f0029dc6fe4984" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.170501 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"834a1c0046413df441fd1d31eb8b082de3727f22cabb3144a9f0029dc6fe4984"} err="failed to get container status \"834a1c0046413df441fd1d31eb8b082de3727f22cabb3144a9f0029dc6fe4984\": rpc error: code = NotFound desc = could not find container \"834a1c0046413df441fd1d31eb8b082de3727f22cabb3144a9f0029dc6fe4984\": container with ID starting with 834a1c0046413df441fd1d31eb8b082de3727f22cabb3144a9f0029dc6fe4984 not found: ID does not exist" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.170534 4766 scope.go:117] "RemoveContainer" containerID="9bcbabef2bd02fa775ac02940096c80cb7ef049913d92fa23bbfdd938502f110" Nov 26 00:28:17 crc kubenswrapper[4766]: E1126 00:28:17.170906 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bcbabef2bd02fa775ac02940096c80cb7ef049913d92fa23bbfdd938502f110\": container with ID starting with 9bcbabef2bd02fa775ac02940096c80cb7ef049913d92fa23bbfdd938502f110 not found: ID does not exist" containerID="9bcbabef2bd02fa775ac02940096c80cb7ef049913d92fa23bbfdd938502f110" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.170947 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bcbabef2bd02fa775ac02940096c80cb7ef049913d92fa23bbfdd938502f110"} err="failed to get container status \"9bcbabef2bd02fa775ac02940096c80cb7ef049913d92fa23bbfdd938502f110\": rpc error: code = NotFound desc = could not find container \"9bcbabef2bd02fa775ac02940096c80cb7ef049913d92fa23bbfdd938502f110\": container with ID starting with 9bcbabef2bd02fa775ac02940096c80cb7ef049913d92fa23bbfdd938502f110 not found: ID does not exist" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.170980 4766 scope.go:117] "RemoveContainer" containerID="88933fddc0f7bbceaeeaf1000987d3306168ed1c04a17ee51b95a84b87a57e90" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.185018 4766 scope.go:117] "RemoveContainer" containerID="5592e87ce814e03fa260d84001e369aa18aac8f3a8f478b6c309b81b7db5ee5f" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.202434 4766 scope.go:117] "RemoveContainer" containerID="f2da372e0c8d91ca02e5ff2d3b6e62a853dad47179dc42c14d64a58dbe69170a" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.212754 4766 scope.go:117] "RemoveContainer" containerID="88933fddc0f7bbceaeeaf1000987d3306168ed1c04a17ee51b95a84b87a57e90" Nov 26 00:28:17 crc kubenswrapper[4766]: E1126 00:28:17.213245 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88933fddc0f7bbceaeeaf1000987d3306168ed1c04a17ee51b95a84b87a57e90\": container with ID starting with 88933fddc0f7bbceaeeaf1000987d3306168ed1c04a17ee51b95a84b87a57e90 not found: ID does not exist" containerID="88933fddc0f7bbceaeeaf1000987d3306168ed1c04a17ee51b95a84b87a57e90" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.213299 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88933fddc0f7bbceaeeaf1000987d3306168ed1c04a17ee51b95a84b87a57e90"} err="failed to get container status \"88933fddc0f7bbceaeeaf1000987d3306168ed1c04a17ee51b95a84b87a57e90\": rpc error: code = NotFound desc = could not find container \"88933fddc0f7bbceaeeaf1000987d3306168ed1c04a17ee51b95a84b87a57e90\": container with ID starting with 88933fddc0f7bbceaeeaf1000987d3306168ed1c04a17ee51b95a84b87a57e90 not found: ID does not exist" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.213335 4766 scope.go:117] "RemoveContainer" containerID="5592e87ce814e03fa260d84001e369aa18aac8f3a8f478b6c309b81b7db5ee5f" Nov 26 00:28:17 crc kubenswrapper[4766]: E1126 00:28:17.213716 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5592e87ce814e03fa260d84001e369aa18aac8f3a8f478b6c309b81b7db5ee5f\": container with ID starting with 5592e87ce814e03fa260d84001e369aa18aac8f3a8f478b6c309b81b7db5ee5f not found: ID does not exist" containerID="5592e87ce814e03fa260d84001e369aa18aac8f3a8f478b6c309b81b7db5ee5f" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.213756 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5592e87ce814e03fa260d84001e369aa18aac8f3a8f478b6c309b81b7db5ee5f"} err="failed to get container status \"5592e87ce814e03fa260d84001e369aa18aac8f3a8f478b6c309b81b7db5ee5f\": rpc error: code = NotFound desc = could not find container \"5592e87ce814e03fa260d84001e369aa18aac8f3a8f478b6c309b81b7db5ee5f\": container with ID starting with 5592e87ce814e03fa260d84001e369aa18aac8f3a8f478b6c309b81b7db5ee5f not found: ID does not exist" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.213787 4766 scope.go:117] "RemoveContainer" containerID="f2da372e0c8d91ca02e5ff2d3b6e62a853dad47179dc42c14d64a58dbe69170a" Nov 26 00:28:17 crc kubenswrapper[4766]: E1126 00:28:17.214070 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2da372e0c8d91ca02e5ff2d3b6e62a853dad47179dc42c14d64a58dbe69170a\": container with ID starting with f2da372e0c8d91ca02e5ff2d3b6e62a853dad47179dc42c14d64a58dbe69170a not found: ID does not exist" containerID="f2da372e0c8d91ca02e5ff2d3b6e62a853dad47179dc42c14d64a58dbe69170a" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.214154 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2da372e0c8d91ca02e5ff2d3b6e62a853dad47179dc42c14d64a58dbe69170a"} err="failed to get container status \"f2da372e0c8d91ca02e5ff2d3b6e62a853dad47179dc42c14d64a58dbe69170a\": rpc error: code = NotFound desc = could not find container \"f2da372e0c8d91ca02e5ff2d3b6e62a853dad47179dc42c14d64a58dbe69170a\": container with ID starting with f2da372e0c8d91ca02e5ff2d3b6e62a853dad47179dc42c14d64a58dbe69170a not found: ID does not exist" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.214174 4766 scope.go:117] "RemoveContainer" containerID="589b3e5703712a442130255630f9f46e0f819c363e3eb4fd4223fae322d7fdc5" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.225839 4766 scope.go:117] "RemoveContainer" containerID="d3daea790428b806b4c3fb6edd90c6968882cacfdc2cbffc88c3418f31ab79bc" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.239071 4766 scope.go:117] "RemoveContainer" containerID="3e83bc756ab461748f9094ecab67a4fc677a1dcc76993c47e831c2f3fe9f0f43" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.250428 4766 scope.go:117] "RemoveContainer" containerID="589b3e5703712a442130255630f9f46e0f819c363e3eb4fd4223fae322d7fdc5" Nov 26 00:28:17 crc kubenswrapper[4766]: E1126 00:28:17.250864 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"589b3e5703712a442130255630f9f46e0f819c363e3eb4fd4223fae322d7fdc5\": container with ID starting with 589b3e5703712a442130255630f9f46e0f819c363e3eb4fd4223fae322d7fdc5 not found: ID does not exist" containerID="589b3e5703712a442130255630f9f46e0f819c363e3eb4fd4223fae322d7fdc5" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.250912 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"589b3e5703712a442130255630f9f46e0f819c363e3eb4fd4223fae322d7fdc5"} err="failed to get container status \"589b3e5703712a442130255630f9f46e0f819c363e3eb4fd4223fae322d7fdc5\": rpc error: code = NotFound desc = could not find container \"589b3e5703712a442130255630f9f46e0f819c363e3eb4fd4223fae322d7fdc5\": container with ID starting with 589b3e5703712a442130255630f9f46e0f819c363e3eb4fd4223fae322d7fdc5 not found: ID does not exist" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.250939 4766 scope.go:117] "RemoveContainer" containerID="d3daea790428b806b4c3fb6edd90c6968882cacfdc2cbffc88c3418f31ab79bc" Nov 26 00:28:17 crc kubenswrapper[4766]: E1126 00:28:17.251349 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3daea790428b806b4c3fb6edd90c6968882cacfdc2cbffc88c3418f31ab79bc\": container with ID starting with d3daea790428b806b4c3fb6edd90c6968882cacfdc2cbffc88c3418f31ab79bc not found: ID does not exist" containerID="d3daea790428b806b4c3fb6edd90c6968882cacfdc2cbffc88c3418f31ab79bc" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.251387 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3daea790428b806b4c3fb6edd90c6968882cacfdc2cbffc88c3418f31ab79bc"} err="failed to get container status \"d3daea790428b806b4c3fb6edd90c6968882cacfdc2cbffc88c3418f31ab79bc\": rpc error: code = NotFound desc = could not find container \"d3daea790428b806b4c3fb6edd90c6968882cacfdc2cbffc88c3418f31ab79bc\": container with ID starting with d3daea790428b806b4c3fb6edd90c6968882cacfdc2cbffc88c3418f31ab79bc not found: ID does not exist" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.251416 4766 scope.go:117] "RemoveContainer" containerID="3e83bc756ab461748f9094ecab67a4fc677a1dcc76993c47e831c2f3fe9f0f43" Nov 26 00:28:17 crc kubenswrapper[4766]: E1126 00:28:17.252282 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e83bc756ab461748f9094ecab67a4fc677a1dcc76993c47e831c2f3fe9f0f43\": container with ID starting with 3e83bc756ab461748f9094ecab67a4fc677a1dcc76993c47e831c2f3fe9f0f43 not found: ID does not exist" containerID="3e83bc756ab461748f9094ecab67a4fc677a1dcc76993c47e831c2f3fe9f0f43" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.252312 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e83bc756ab461748f9094ecab67a4fc677a1dcc76993c47e831c2f3fe9f0f43"} err="failed to get container status \"3e83bc756ab461748f9094ecab67a4fc677a1dcc76993c47e831c2f3fe9f0f43\": rpc error: code = NotFound desc = could not find container \"3e83bc756ab461748f9094ecab67a4fc677a1dcc76993c47e831c2f3fe9f0f43\": container with ID starting with 3e83bc756ab461748f9094ecab67a4fc677a1dcc76993c47e831c2f3fe9f0f43 not found: ID does not exist" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.252335 4766 scope.go:117] "RemoveContainer" containerID="7bbaf76b46cc33ba0b12824530776175690a99499a1ca6cdce26885e22214afe" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.266746 4766 scope.go:117] "RemoveContainer" containerID="053afe7852af7f429134c0b1751a485763eb8c26ff50a42037d8eb4238d3b56d" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.278971 4766 scope.go:117] "RemoveContainer" containerID="e6a5e794fa2c82f4ac11133d96f64e1cb2b14592dd97dfc020e9863dea5b7ea9" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.290636 4766 scope.go:117] "RemoveContainer" containerID="7bbaf76b46cc33ba0b12824530776175690a99499a1ca6cdce26885e22214afe" Nov 26 00:28:17 crc kubenswrapper[4766]: E1126 00:28:17.291078 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bbaf76b46cc33ba0b12824530776175690a99499a1ca6cdce26885e22214afe\": container with ID starting with 7bbaf76b46cc33ba0b12824530776175690a99499a1ca6cdce26885e22214afe not found: ID does not exist" containerID="7bbaf76b46cc33ba0b12824530776175690a99499a1ca6cdce26885e22214afe" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.291116 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bbaf76b46cc33ba0b12824530776175690a99499a1ca6cdce26885e22214afe"} err="failed to get container status \"7bbaf76b46cc33ba0b12824530776175690a99499a1ca6cdce26885e22214afe\": rpc error: code = NotFound desc = could not find container \"7bbaf76b46cc33ba0b12824530776175690a99499a1ca6cdce26885e22214afe\": container with ID starting with 7bbaf76b46cc33ba0b12824530776175690a99499a1ca6cdce26885e22214afe not found: ID does not exist" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.291140 4766 scope.go:117] "RemoveContainer" containerID="053afe7852af7f429134c0b1751a485763eb8c26ff50a42037d8eb4238d3b56d" Nov 26 00:28:17 crc kubenswrapper[4766]: E1126 00:28:17.291698 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"053afe7852af7f429134c0b1751a485763eb8c26ff50a42037d8eb4238d3b56d\": container with ID starting with 053afe7852af7f429134c0b1751a485763eb8c26ff50a42037d8eb4238d3b56d not found: ID does not exist" containerID="053afe7852af7f429134c0b1751a485763eb8c26ff50a42037d8eb4238d3b56d" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.291725 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"053afe7852af7f429134c0b1751a485763eb8c26ff50a42037d8eb4238d3b56d"} err="failed to get container status \"053afe7852af7f429134c0b1751a485763eb8c26ff50a42037d8eb4238d3b56d\": rpc error: code = NotFound desc = could not find container \"053afe7852af7f429134c0b1751a485763eb8c26ff50a42037d8eb4238d3b56d\": container with ID starting with 053afe7852af7f429134c0b1751a485763eb8c26ff50a42037d8eb4238d3b56d not found: ID does not exist" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.291747 4766 scope.go:117] "RemoveContainer" containerID="e6a5e794fa2c82f4ac11133d96f64e1cb2b14592dd97dfc020e9863dea5b7ea9" Nov 26 00:28:17 crc kubenswrapper[4766]: E1126 00:28:17.292483 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6a5e794fa2c82f4ac11133d96f64e1cb2b14592dd97dfc020e9863dea5b7ea9\": container with ID starting with e6a5e794fa2c82f4ac11133d96f64e1cb2b14592dd97dfc020e9863dea5b7ea9 not found: ID does not exist" containerID="e6a5e794fa2c82f4ac11133d96f64e1cb2b14592dd97dfc020e9863dea5b7ea9" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.292535 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6a5e794fa2c82f4ac11133d96f64e1cb2b14592dd97dfc020e9863dea5b7ea9"} err="failed to get container status \"e6a5e794fa2c82f4ac11133d96f64e1cb2b14592dd97dfc020e9863dea5b7ea9\": rpc error: code = NotFound desc = could not find container \"e6a5e794fa2c82f4ac11133d96f64e1cb2b14592dd97dfc020e9863dea5b7ea9\": container with ID starting with e6a5e794fa2c82f4ac11133d96f64e1cb2b14592dd97dfc020e9863dea5b7ea9 not found: ID does not exist" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.416450 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s9x5n"] Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.422114 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-s9x5n"] Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.833075 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1529f474-493a-485a-80b7-ee8ec1c79ccc" path="/var/lib/kubelet/pods/1529f474-493a-485a-80b7-ee8ec1c79ccc/volumes" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.833730 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f6c20b5-b168-4529-b84a-bef5c19254d5" path="/var/lib/kubelet/pods/5f6c20b5-b168-4529-b84a-bef5c19254d5/volumes" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.834344 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90830d76-8bef-4c13-b127-23fff3e32770" path="/var/lib/kubelet/pods/90830d76-8bef-4c13-b127-23fff3e32770/volumes" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.835354 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0d4c85c-2a27-4a2f-903e-8f7473351104" path="/var/lib/kubelet/pods/e0d4c85c-2a27-4a2f-903e-8f7473351104/volumes" Nov 26 00:28:17 crc kubenswrapper[4766]: I1126 00:28:17.835932 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1e5e774-51ea-47e8-8b02-1bde1728da35" path="/var/lib/kubelet/pods/f1e5e774-51ea-47e8-8b02-1bde1728da35/volumes" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.102303 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2xsk5" event={"ID":"cad040b4-5d74-4cd0-b726-93d9c6cb73e2","Type":"ContainerStarted","Data":"8fd615be646a813060b34100eee5d38f87f1719bcc3edc4ed1e5a2c61dffea5a"} Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.102388 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-2xsk5" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.107284 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-2xsk5" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.121765 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-2xsk5" podStartSLOduration=2.121741037 podStartE2EDuration="2.121741037s" podCreationTimestamp="2025-11-26 00:28:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:28:18.117576196 +0000 UTC m=+278.966346656" watchObservedRunningTime="2025-11-26 00:28:18.121741037 +0000 UTC m=+278.970511467" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.549088 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vhw86"] Nov 26 00:28:18 crc kubenswrapper[4766]: E1126 00:28:18.549277 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f6c20b5-b168-4529-b84a-bef5c19254d5" containerName="extract-utilities" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.549290 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f6c20b5-b168-4529-b84a-bef5c19254d5" containerName="extract-utilities" Nov 26 00:28:18 crc kubenswrapper[4766]: E1126 00:28:18.549304 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1529f474-493a-485a-80b7-ee8ec1c79ccc" containerName="extract-utilities" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.549310 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="1529f474-493a-485a-80b7-ee8ec1c79ccc" containerName="extract-utilities" Nov 26 00:28:18 crc kubenswrapper[4766]: E1126 00:28:18.549317 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0d4c85c-2a27-4a2f-903e-8f7473351104" containerName="registry-server" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.549324 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0d4c85c-2a27-4a2f-903e-8f7473351104" containerName="registry-server" Nov 26 00:28:18 crc kubenswrapper[4766]: E1126 00:28:18.549331 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90830d76-8bef-4c13-b127-23fff3e32770" containerName="extract-utilities" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.549338 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="90830d76-8bef-4c13-b127-23fff3e32770" containerName="extract-utilities" Nov 26 00:28:18 crc kubenswrapper[4766]: E1126 00:28:18.549347 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90830d76-8bef-4c13-b127-23fff3e32770" containerName="extract-content" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.549353 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="90830d76-8bef-4c13-b127-23fff3e32770" containerName="extract-content" Nov 26 00:28:18 crc kubenswrapper[4766]: E1126 00:28:18.549360 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0d4c85c-2a27-4a2f-903e-8f7473351104" containerName="extract-utilities" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.549367 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0d4c85c-2a27-4a2f-903e-8f7473351104" containerName="extract-utilities" Nov 26 00:28:18 crc kubenswrapper[4766]: E1126 00:28:18.549375 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1529f474-493a-485a-80b7-ee8ec1c79ccc" containerName="extract-content" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.549380 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="1529f474-493a-485a-80b7-ee8ec1c79ccc" containerName="extract-content" Nov 26 00:28:18 crc kubenswrapper[4766]: E1126 00:28:18.549387 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f6c20b5-b168-4529-b84a-bef5c19254d5" containerName="extract-content" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.549393 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f6c20b5-b168-4529-b84a-bef5c19254d5" containerName="extract-content" Nov 26 00:28:18 crc kubenswrapper[4766]: E1126 00:28:18.549401 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90830d76-8bef-4c13-b127-23fff3e32770" containerName="registry-server" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.549406 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="90830d76-8bef-4c13-b127-23fff3e32770" containerName="registry-server" Nov 26 00:28:18 crc kubenswrapper[4766]: E1126 00:28:18.549415 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f6c20b5-b168-4529-b84a-bef5c19254d5" containerName="registry-server" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.549421 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f6c20b5-b168-4529-b84a-bef5c19254d5" containerName="registry-server" Nov 26 00:28:18 crc kubenswrapper[4766]: E1126 00:28:18.549429 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1529f474-493a-485a-80b7-ee8ec1c79ccc" containerName="registry-server" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.549435 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="1529f474-493a-485a-80b7-ee8ec1c79ccc" containerName="registry-server" Nov 26 00:28:18 crc kubenswrapper[4766]: E1126 00:28:18.549442 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0d4c85c-2a27-4a2f-903e-8f7473351104" containerName="extract-content" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.549447 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0d4c85c-2a27-4a2f-903e-8f7473351104" containerName="extract-content" Nov 26 00:28:18 crc kubenswrapper[4766]: E1126 00:28:18.549455 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1e5e774-51ea-47e8-8b02-1bde1728da35" containerName="marketplace-operator" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.549460 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1e5e774-51ea-47e8-8b02-1bde1728da35" containerName="marketplace-operator" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.549536 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0d4c85c-2a27-4a2f-903e-8f7473351104" containerName="registry-server" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.549544 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="90830d76-8bef-4c13-b127-23fff3e32770" containerName="registry-server" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.549554 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1e5e774-51ea-47e8-8b02-1bde1728da35" containerName="marketplace-operator" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.549562 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="1529f474-493a-485a-80b7-ee8ec1c79ccc" containerName="registry-server" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.549572 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f6c20b5-b168-4529-b84a-bef5c19254d5" containerName="registry-server" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.552030 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vhw86" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.556253 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.564154 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vhw86"] Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.688915 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f18c8ba9-685a-4cd8-92e9-797074b81278-utilities\") pod \"redhat-marketplace-vhw86\" (UID: \"f18c8ba9-685a-4cd8-92e9-797074b81278\") " pod="openshift-marketplace/redhat-marketplace-vhw86" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.688983 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxmmq\" (UniqueName: \"kubernetes.io/projected/f18c8ba9-685a-4cd8-92e9-797074b81278-kube-api-access-cxmmq\") pod \"redhat-marketplace-vhw86\" (UID: \"f18c8ba9-685a-4cd8-92e9-797074b81278\") " pod="openshift-marketplace/redhat-marketplace-vhw86" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.689176 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f18c8ba9-685a-4cd8-92e9-797074b81278-catalog-content\") pod \"redhat-marketplace-vhw86\" (UID: \"f18c8ba9-685a-4cd8-92e9-797074b81278\") " pod="openshift-marketplace/redhat-marketplace-vhw86" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.747872 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6vhw6"] Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.748755 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vhw6" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.754063 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.764040 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6vhw6"] Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.790310 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f18c8ba9-685a-4cd8-92e9-797074b81278-utilities\") pod \"redhat-marketplace-vhw86\" (UID: \"f18c8ba9-685a-4cd8-92e9-797074b81278\") " pod="openshift-marketplace/redhat-marketplace-vhw86" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.790372 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxmmq\" (UniqueName: \"kubernetes.io/projected/f18c8ba9-685a-4cd8-92e9-797074b81278-kube-api-access-cxmmq\") pod \"redhat-marketplace-vhw86\" (UID: \"f18c8ba9-685a-4cd8-92e9-797074b81278\") " pod="openshift-marketplace/redhat-marketplace-vhw86" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.790413 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f18c8ba9-685a-4cd8-92e9-797074b81278-catalog-content\") pod \"redhat-marketplace-vhw86\" (UID: \"f18c8ba9-685a-4cd8-92e9-797074b81278\") " pod="openshift-marketplace/redhat-marketplace-vhw86" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.790863 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f18c8ba9-685a-4cd8-92e9-797074b81278-utilities\") pod \"redhat-marketplace-vhw86\" (UID: \"f18c8ba9-685a-4cd8-92e9-797074b81278\") " pod="openshift-marketplace/redhat-marketplace-vhw86" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.790883 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f18c8ba9-685a-4cd8-92e9-797074b81278-catalog-content\") pod \"redhat-marketplace-vhw86\" (UID: \"f18c8ba9-685a-4cd8-92e9-797074b81278\") " pod="openshift-marketplace/redhat-marketplace-vhw86" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.809238 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxmmq\" (UniqueName: \"kubernetes.io/projected/f18c8ba9-685a-4cd8-92e9-797074b81278-kube-api-access-cxmmq\") pod \"redhat-marketplace-vhw86\" (UID: \"f18c8ba9-685a-4cd8-92e9-797074b81278\") " pod="openshift-marketplace/redhat-marketplace-vhw86" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.891280 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8951948c-6794-4707-b88e-9e7b506c7df0-utilities\") pod \"community-operators-6vhw6\" (UID: \"8951948c-6794-4707-b88e-9e7b506c7df0\") " pod="openshift-marketplace/community-operators-6vhw6" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.891464 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rndnx\" (UniqueName: \"kubernetes.io/projected/8951948c-6794-4707-b88e-9e7b506c7df0-kube-api-access-rndnx\") pod \"community-operators-6vhw6\" (UID: \"8951948c-6794-4707-b88e-9e7b506c7df0\") " pod="openshift-marketplace/community-operators-6vhw6" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.891493 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8951948c-6794-4707-b88e-9e7b506c7df0-catalog-content\") pod \"community-operators-6vhw6\" (UID: \"8951948c-6794-4707-b88e-9e7b506c7df0\") " pod="openshift-marketplace/community-operators-6vhw6" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.904375 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vhw86" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.995239 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8951948c-6794-4707-b88e-9e7b506c7df0-utilities\") pod \"community-operators-6vhw6\" (UID: \"8951948c-6794-4707-b88e-9e7b506c7df0\") " pod="openshift-marketplace/community-operators-6vhw6" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.995478 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rndnx\" (UniqueName: \"kubernetes.io/projected/8951948c-6794-4707-b88e-9e7b506c7df0-kube-api-access-rndnx\") pod \"community-operators-6vhw6\" (UID: \"8951948c-6794-4707-b88e-9e7b506c7df0\") " pod="openshift-marketplace/community-operators-6vhw6" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.995498 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8951948c-6794-4707-b88e-9e7b506c7df0-catalog-content\") pod \"community-operators-6vhw6\" (UID: \"8951948c-6794-4707-b88e-9e7b506c7df0\") " pod="openshift-marketplace/community-operators-6vhw6" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.995910 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8951948c-6794-4707-b88e-9e7b506c7df0-catalog-content\") pod \"community-operators-6vhw6\" (UID: \"8951948c-6794-4707-b88e-9e7b506c7df0\") " pod="openshift-marketplace/community-operators-6vhw6" Nov 26 00:28:18 crc kubenswrapper[4766]: I1126 00:28:18.996117 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8951948c-6794-4707-b88e-9e7b506c7df0-utilities\") pod \"community-operators-6vhw6\" (UID: \"8951948c-6794-4707-b88e-9e7b506c7df0\") " pod="openshift-marketplace/community-operators-6vhw6" Nov 26 00:28:19 crc kubenswrapper[4766]: I1126 00:28:19.020869 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rndnx\" (UniqueName: \"kubernetes.io/projected/8951948c-6794-4707-b88e-9e7b506c7df0-kube-api-access-rndnx\") pod \"community-operators-6vhw6\" (UID: \"8951948c-6794-4707-b88e-9e7b506c7df0\") " pod="openshift-marketplace/community-operators-6vhw6" Nov 26 00:28:19 crc kubenswrapper[4766]: I1126 00:28:19.081372 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vhw6" Nov 26 00:28:19 crc kubenswrapper[4766]: I1126 00:28:19.102696 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vhw86"] Nov 26 00:28:19 crc kubenswrapper[4766]: I1126 00:28:19.273768 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6vhw6"] Nov 26 00:28:19 crc kubenswrapper[4766]: W1126 00:28:19.352367 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8951948c_6794_4707_b88e_9e7b506c7df0.slice/crio-120d87266e1db442f2f0ec2b5b9c21d972c21dffc4badb557fb82360e4eaf848 WatchSource:0}: Error finding container 120d87266e1db442f2f0ec2b5b9c21d972c21dffc4badb557fb82360e4eaf848: Status 404 returned error can't find the container with id 120d87266e1db442f2f0ec2b5b9c21d972c21dffc4badb557fb82360e4eaf848 Nov 26 00:28:20 crc kubenswrapper[4766]: I1126 00:28:20.116962 4766 generic.go:334] "Generic (PLEG): container finished" podID="8951948c-6794-4707-b88e-9e7b506c7df0" containerID="22997b585a6d84f3f360ddb7f52b0dcc030828cdce4794210dc6451b9ee841f2" exitCode=0 Nov 26 00:28:20 crc kubenswrapper[4766]: I1126 00:28:20.117065 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vhw6" event={"ID":"8951948c-6794-4707-b88e-9e7b506c7df0","Type":"ContainerDied","Data":"22997b585a6d84f3f360ddb7f52b0dcc030828cdce4794210dc6451b9ee841f2"} Nov 26 00:28:20 crc kubenswrapper[4766]: I1126 00:28:20.117268 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vhw6" event={"ID":"8951948c-6794-4707-b88e-9e7b506c7df0","Type":"ContainerStarted","Data":"120d87266e1db442f2f0ec2b5b9c21d972c21dffc4badb557fb82360e4eaf848"} Nov 26 00:28:20 crc kubenswrapper[4766]: I1126 00:28:20.118541 4766 generic.go:334] "Generic (PLEG): container finished" podID="f18c8ba9-685a-4cd8-92e9-797074b81278" containerID="168319fc9cb9a9e56f2fde9aa1e5df3f9d538cc3e37dba3532617c8ed1a24e1e" exitCode=0 Nov 26 00:28:20 crc kubenswrapper[4766]: I1126 00:28:20.118784 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vhw86" event={"ID":"f18c8ba9-685a-4cd8-92e9-797074b81278","Type":"ContainerDied","Data":"168319fc9cb9a9e56f2fde9aa1e5df3f9d538cc3e37dba3532617c8ed1a24e1e"} Nov 26 00:28:20 crc kubenswrapper[4766]: I1126 00:28:20.118816 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vhw86" event={"ID":"f18c8ba9-685a-4cd8-92e9-797074b81278","Type":"ContainerStarted","Data":"ed66397c3d5527026bf672cfb91b3ac839ed6f3a6a6fa1db8c05aec409f03a15"} Nov 26 00:28:20 crc kubenswrapper[4766]: I1126 00:28:20.960897 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qrsgb"] Nov 26 00:28:20 crc kubenswrapper[4766]: I1126 00:28:20.966616 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qrsgb"] Nov 26 00:28:20 crc kubenswrapper[4766]: I1126 00:28:20.966734 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qrsgb" Nov 26 00:28:20 crc kubenswrapper[4766]: I1126 00:28:20.970503 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.024732 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d93cf31b-be67-48b1-a248-ba0df8870fdd-utilities\") pod \"certified-operators-qrsgb\" (UID: \"d93cf31b-be67-48b1-a248-ba0df8870fdd\") " pod="openshift-marketplace/certified-operators-qrsgb" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.025082 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gshs6\" (UniqueName: \"kubernetes.io/projected/d93cf31b-be67-48b1-a248-ba0df8870fdd-kube-api-access-gshs6\") pod \"certified-operators-qrsgb\" (UID: \"d93cf31b-be67-48b1-a248-ba0df8870fdd\") " pod="openshift-marketplace/certified-operators-qrsgb" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.025409 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d93cf31b-be67-48b1-a248-ba0df8870fdd-catalog-content\") pod \"certified-operators-qrsgb\" (UID: \"d93cf31b-be67-48b1-a248-ba0df8870fdd\") " pod="openshift-marketplace/certified-operators-qrsgb" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.125449 4766 generic.go:334] "Generic (PLEG): container finished" podID="8951948c-6794-4707-b88e-9e7b506c7df0" containerID="95a5a7028bd9bfaf61e3fee7917eb6da7156d1f2ca4fba89488f54f6d26c0bb6" exitCode=0 Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.125525 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vhw6" event={"ID":"8951948c-6794-4707-b88e-9e7b506c7df0","Type":"ContainerDied","Data":"95a5a7028bd9bfaf61e3fee7917eb6da7156d1f2ca4fba89488f54f6d26c0bb6"} Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.127463 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d93cf31b-be67-48b1-a248-ba0df8870fdd-catalog-content\") pod \"certified-operators-qrsgb\" (UID: \"d93cf31b-be67-48b1-a248-ba0df8870fdd\") " pod="openshift-marketplace/certified-operators-qrsgb" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.127509 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d93cf31b-be67-48b1-a248-ba0df8870fdd-utilities\") pod \"certified-operators-qrsgb\" (UID: \"d93cf31b-be67-48b1-a248-ba0df8870fdd\") " pod="openshift-marketplace/certified-operators-qrsgb" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.127530 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gshs6\" (UniqueName: \"kubernetes.io/projected/d93cf31b-be67-48b1-a248-ba0df8870fdd-kube-api-access-gshs6\") pod \"certified-operators-qrsgb\" (UID: \"d93cf31b-be67-48b1-a248-ba0df8870fdd\") " pod="openshift-marketplace/certified-operators-qrsgb" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.127547 4766 generic.go:334] "Generic (PLEG): container finished" podID="f18c8ba9-685a-4cd8-92e9-797074b81278" containerID="fea2745de6557173ad45a863118b46ddb1ca7a10bfe091165d7a55e1dde531b3" exitCode=0 Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.127764 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vhw86" event={"ID":"f18c8ba9-685a-4cd8-92e9-797074b81278","Type":"ContainerDied","Data":"fea2745de6557173ad45a863118b46ddb1ca7a10bfe091165d7a55e1dde531b3"} Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.128086 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d93cf31b-be67-48b1-a248-ba0df8870fdd-catalog-content\") pod \"certified-operators-qrsgb\" (UID: \"d93cf31b-be67-48b1-a248-ba0df8870fdd\") " pod="openshift-marketplace/certified-operators-qrsgb" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.128538 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d93cf31b-be67-48b1-a248-ba0df8870fdd-utilities\") pod \"certified-operators-qrsgb\" (UID: \"d93cf31b-be67-48b1-a248-ba0df8870fdd\") " pod="openshift-marketplace/certified-operators-qrsgb" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.153221 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gshs6\" (UniqueName: \"kubernetes.io/projected/d93cf31b-be67-48b1-a248-ba0df8870fdd-kube-api-access-gshs6\") pod \"certified-operators-qrsgb\" (UID: \"d93cf31b-be67-48b1-a248-ba0df8870fdd\") " pod="openshift-marketplace/certified-operators-qrsgb" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.156986 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pjtsc"] Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.158265 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pjtsc" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.161674 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.164871 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pjtsc"] Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.228395 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8qxw\" (UniqueName: \"kubernetes.io/projected/34c7586e-58a4-4e97-b559-cb805d91e9d8-kube-api-access-z8qxw\") pod \"redhat-operators-pjtsc\" (UID: \"34c7586e-58a4-4e97-b559-cb805d91e9d8\") " pod="openshift-marketplace/redhat-operators-pjtsc" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.228444 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c7586e-58a4-4e97-b559-cb805d91e9d8-utilities\") pod \"redhat-operators-pjtsc\" (UID: \"34c7586e-58a4-4e97-b559-cb805d91e9d8\") " pod="openshift-marketplace/redhat-operators-pjtsc" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.228492 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c7586e-58a4-4e97-b559-cb805d91e9d8-catalog-content\") pod \"redhat-operators-pjtsc\" (UID: \"34c7586e-58a4-4e97-b559-cb805d91e9d8\") " pod="openshift-marketplace/redhat-operators-pjtsc" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.291606 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qrsgb" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.329554 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8qxw\" (UniqueName: \"kubernetes.io/projected/34c7586e-58a4-4e97-b559-cb805d91e9d8-kube-api-access-z8qxw\") pod \"redhat-operators-pjtsc\" (UID: \"34c7586e-58a4-4e97-b559-cb805d91e9d8\") " pod="openshift-marketplace/redhat-operators-pjtsc" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.329607 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c7586e-58a4-4e97-b559-cb805d91e9d8-utilities\") pod \"redhat-operators-pjtsc\" (UID: \"34c7586e-58a4-4e97-b559-cb805d91e9d8\") " pod="openshift-marketplace/redhat-operators-pjtsc" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.329638 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c7586e-58a4-4e97-b559-cb805d91e9d8-catalog-content\") pod \"redhat-operators-pjtsc\" (UID: \"34c7586e-58a4-4e97-b559-cb805d91e9d8\") " pod="openshift-marketplace/redhat-operators-pjtsc" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.330336 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c7586e-58a4-4e97-b559-cb805d91e9d8-utilities\") pod \"redhat-operators-pjtsc\" (UID: \"34c7586e-58a4-4e97-b559-cb805d91e9d8\") " pod="openshift-marketplace/redhat-operators-pjtsc" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.330500 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c7586e-58a4-4e97-b559-cb805d91e9d8-catalog-content\") pod \"redhat-operators-pjtsc\" (UID: \"34c7586e-58a4-4e97-b559-cb805d91e9d8\") " pod="openshift-marketplace/redhat-operators-pjtsc" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.354097 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8qxw\" (UniqueName: \"kubernetes.io/projected/34c7586e-58a4-4e97-b559-cb805d91e9d8-kube-api-access-z8qxw\") pod \"redhat-operators-pjtsc\" (UID: \"34c7586e-58a4-4e97-b559-cb805d91e9d8\") " pod="openshift-marketplace/redhat-operators-pjtsc" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.494285 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pjtsc" Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.499691 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qrsgb"] Nov 26 00:28:21 crc kubenswrapper[4766]: I1126 00:28:21.682556 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pjtsc"] Nov 26 00:28:21 crc kubenswrapper[4766]: W1126 00:28:21.692453 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34c7586e_58a4_4e97_b559_cb805d91e9d8.slice/crio-2f7dfc407b1d2929d32691edafd7b4bed6492ce14fce123f5206c2bc3f6650a5 WatchSource:0}: Error finding container 2f7dfc407b1d2929d32691edafd7b4bed6492ce14fce123f5206c2bc3f6650a5: Status 404 returned error can't find the container with id 2f7dfc407b1d2929d32691edafd7b4bed6492ce14fce123f5206c2bc3f6650a5 Nov 26 00:28:22 crc kubenswrapper[4766]: I1126 00:28:22.135485 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vhw6" event={"ID":"8951948c-6794-4707-b88e-9e7b506c7df0","Type":"ContainerStarted","Data":"0d9da0dedb1cb7c18aea33f003bb734bd7f7ca3849370a469e8929e09b890095"} Nov 26 00:28:22 crc kubenswrapper[4766]: I1126 00:28:22.140813 4766 generic.go:334] "Generic (PLEG): container finished" podID="d93cf31b-be67-48b1-a248-ba0df8870fdd" containerID="46d3ce5e9c1e6baa80f11a2d9811b7c9b65d645dfac64cac84c89b7956313321" exitCode=0 Nov 26 00:28:22 crc kubenswrapper[4766]: I1126 00:28:22.140933 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qrsgb" event={"ID":"d93cf31b-be67-48b1-a248-ba0df8870fdd","Type":"ContainerDied","Data":"46d3ce5e9c1e6baa80f11a2d9811b7c9b65d645dfac64cac84c89b7956313321"} Nov 26 00:28:22 crc kubenswrapper[4766]: I1126 00:28:22.140972 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qrsgb" event={"ID":"d93cf31b-be67-48b1-a248-ba0df8870fdd","Type":"ContainerStarted","Data":"23f16556b38d18934fd8b69ccffbbdc7cd65a30a96b5a7abedfa6b7f97e7376e"} Nov 26 00:28:22 crc kubenswrapper[4766]: I1126 00:28:22.148032 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vhw86" event={"ID":"f18c8ba9-685a-4cd8-92e9-797074b81278","Type":"ContainerStarted","Data":"50be7c2bec3a404cfb048fd120462e7e4966c4e35d8b3ecbd9ba832583685750"} Nov 26 00:28:22 crc kubenswrapper[4766]: I1126 00:28:22.150408 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pjtsc" event={"ID":"34c7586e-58a4-4e97-b559-cb805d91e9d8","Type":"ContainerStarted","Data":"43246df2debc3028481015c4213fa387f0345f2a5056b06f2ff2f25523280b18"} Nov 26 00:28:22 crc kubenswrapper[4766]: I1126 00:28:22.150456 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pjtsc" event={"ID":"34c7586e-58a4-4e97-b559-cb805d91e9d8","Type":"ContainerStarted","Data":"2f7dfc407b1d2929d32691edafd7b4bed6492ce14fce123f5206c2bc3f6650a5"} Nov 26 00:28:22 crc kubenswrapper[4766]: I1126 00:28:22.161110 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6vhw6" podStartSLOduration=2.523293694 podStartE2EDuration="4.161089009s" podCreationTimestamp="2025-11-26 00:28:18 +0000 UTC" firstStartedPulling="2025-11-26 00:28:20.120258854 +0000 UTC m=+280.969029284" lastFinishedPulling="2025-11-26 00:28:21.758054169 +0000 UTC m=+282.606824599" observedRunningTime="2025-11-26 00:28:22.156510447 +0000 UTC m=+283.005280877" watchObservedRunningTime="2025-11-26 00:28:22.161089009 +0000 UTC m=+283.009859439" Nov 26 00:28:22 crc kubenswrapper[4766]: I1126 00:28:22.196241 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vhw86" podStartSLOduration=2.596109543 podStartE2EDuration="4.196217034s" podCreationTimestamp="2025-11-26 00:28:18 +0000 UTC" firstStartedPulling="2025-11-26 00:28:20.120345636 +0000 UTC m=+280.969116076" lastFinishedPulling="2025-11-26 00:28:21.720453137 +0000 UTC m=+282.569223567" observedRunningTime="2025-11-26 00:28:22.188894579 +0000 UTC m=+283.037665029" watchObservedRunningTime="2025-11-26 00:28:22.196217034 +0000 UTC m=+283.044987504" Nov 26 00:28:23 crc kubenswrapper[4766]: I1126 00:28:23.156383 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qrsgb" event={"ID":"d93cf31b-be67-48b1-a248-ba0df8870fdd","Type":"ContainerStarted","Data":"2724526881ad660ecc4eac36ae865e4ad120d04d2d05f2e34a339938afcc6da0"} Nov 26 00:28:23 crc kubenswrapper[4766]: I1126 00:28:23.158253 4766 generic.go:334] "Generic (PLEG): container finished" podID="34c7586e-58a4-4e97-b559-cb805d91e9d8" containerID="43246df2debc3028481015c4213fa387f0345f2a5056b06f2ff2f25523280b18" exitCode=0 Nov 26 00:28:23 crc kubenswrapper[4766]: I1126 00:28:23.159272 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pjtsc" event={"ID":"34c7586e-58a4-4e97-b559-cb805d91e9d8","Type":"ContainerDied","Data":"43246df2debc3028481015c4213fa387f0345f2a5056b06f2ff2f25523280b18"} Nov 26 00:28:23 crc kubenswrapper[4766]: I1126 00:28:23.159319 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pjtsc" event={"ID":"34c7586e-58a4-4e97-b559-cb805d91e9d8","Type":"ContainerStarted","Data":"71a790d4ac237a11b3a928ddf5675461bbfd89d067a7d67cb32b6fda2771063e"} Nov 26 00:28:24 crc kubenswrapper[4766]: I1126 00:28:24.166343 4766 generic.go:334] "Generic (PLEG): container finished" podID="d93cf31b-be67-48b1-a248-ba0df8870fdd" containerID="2724526881ad660ecc4eac36ae865e4ad120d04d2d05f2e34a339938afcc6da0" exitCode=0 Nov 26 00:28:24 crc kubenswrapper[4766]: I1126 00:28:24.167601 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qrsgb" event={"ID":"d93cf31b-be67-48b1-a248-ba0df8870fdd","Type":"ContainerDied","Data":"2724526881ad660ecc4eac36ae865e4ad120d04d2d05f2e34a339938afcc6da0"} Nov 26 00:28:24 crc kubenswrapper[4766]: I1126 00:28:24.170112 4766 generic.go:334] "Generic (PLEG): container finished" podID="34c7586e-58a4-4e97-b559-cb805d91e9d8" containerID="71a790d4ac237a11b3a928ddf5675461bbfd89d067a7d67cb32b6fda2771063e" exitCode=0 Nov 26 00:28:24 crc kubenswrapper[4766]: I1126 00:28:24.170160 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pjtsc" event={"ID":"34c7586e-58a4-4e97-b559-cb805d91e9d8","Type":"ContainerDied","Data":"71a790d4ac237a11b3a928ddf5675461bbfd89d067a7d67cb32b6fda2771063e"} Nov 26 00:28:26 crc kubenswrapper[4766]: I1126 00:28:26.181476 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pjtsc" event={"ID":"34c7586e-58a4-4e97-b559-cb805d91e9d8","Type":"ContainerStarted","Data":"928d382c268f34cdf7bed63f93921d82a9b14f6fe2cd1de144b071a0113b30c1"} Nov 26 00:28:26 crc kubenswrapper[4766]: I1126 00:28:26.183542 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qrsgb" event={"ID":"d93cf31b-be67-48b1-a248-ba0df8870fdd","Type":"ContainerStarted","Data":"92fd35cfdb6e079155ae81fd73f3eafb0c01a7c95934425a72564d76ba3e085b"} Nov 26 00:28:26 crc kubenswrapper[4766]: I1126 00:28:26.200640 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pjtsc" podStartSLOduration=2.75215908 podStartE2EDuration="5.200623646s" podCreationTimestamp="2025-11-26 00:28:21 +0000 UTC" firstStartedPulling="2025-11-26 00:28:22.15171699 +0000 UTC m=+283.000487450" lastFinishedPulling="2025-11-26 00:28:24.600181586 +0000 UTC m=+285.448952016" observedRunningTime="2025-11-26 00:28:26.19816174 +0000 UTC m=+287.046932180" watchObservedRunningTime="2025-11-26 00:28:26.200623646 +0000 UTC m=+287.049394076" Nov 26 00:28:26 crc kubenswrapper[4766]: I1126 00:28:26.215195 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qrsgb" podStartSLOduration=3.736331948 podStartE2EDuration="6.215171873s" podCreationTimestamp="2025-11-26 00:28:20 +0000 UTC" firstStartedPulling="2025-11-26 00:28:22.143412149 +0000 UTC m=+282.992182599" lastFinishedPulling="2025-11-26 00:28:24.622252094 +0000 UTC m=+285.471022524" observedRunningTime="2025-11-26 00:28:26.214099075 +0000 UTC m=+287.062869505" watchObservedRunningTime="2025-11-26 00:28:26.215171873 +0000 UTC m=+287.063942303" Nov 26 00:28:28 crc kubenswrapper[4766]: I1126 00:28:28.905576 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vhw86" Nov 26 00:28:28 crc kubenswrapper[4766]: I1126 00:28:28.906115 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vhw86" Nov 26 00:28:28 crc kubenswrapper[4766]: I1126 00:28:28.957076 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vhw86" Nov 26 00:28:29 crc kubenswrapper[4766]: I1126 00:28:29.082429 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6vhw6" Nov 26 00:28:29 crc kubenswrapper[4766]: I1126 00:28:29.082483 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6vhw6" Nov 26 00:28:29 crc kubenswrapper[4766]: I1126 00:28:29.119519 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6vhw6" Nov 26 00:28:29 crc kubenswrapper[4766]: I1126 00:28:29.237410 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6vhw6" Nov 26 00:28:29 crc kubenswrapper[4766]: I1126 00:28:29.240814 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vhw86" Nov 26 00:28:31 crc kubenswrapper[4766]: I1126 00:28:31.292425 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qrsgb" Nov 26 00:28:31 crc kubenswrapper[4766]: I1126 00:28:31.293570 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qrsgb" Nov 26 00:28:31 crc kubenswrapper[4766]: I1126 00:28:31.329248 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qrsgb" Nov 26 00:28:31 crc kubenswrapper[4766]: I1126 00:28:31.494794 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pjtsc" Nov 26 00:28:31 crc kubenswrapper[4766]: I1126 00:28:31.494852 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pjtsc" Nov 26 00:28:31 crc kubenswrapper[4766]: I1126 00:28:31.553224 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pjtsc" Nov 26 00:28:32 crc kubenswrapper[4766]: I1126 00:28:32.273306 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pjtsc" Nov 26 00:28:32 crc kubenswrapper[4766]: I1126 00:28:32.279354 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qrsgb" Nov 26 00:29:11 crc kubenswrapper[4766]: I1126 00:29:11.479554 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:29:11 crc kubenswrapper[4766]: I1126 00:29:11.480218 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:29:41 crc kubenswrapper[4766]: I1126 00:29:41.479995 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:29:41 crc kubenswrapper[4766]: I1126 00:29:41.480869 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:30:00 crc kubenswrapper[4766]: I1126 00:30:00.132526 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z"] Nov 26 00:30:00 crc kubenswrapper[4766]: I1126 00:30:00.133897 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z" Nov 26 00:30:00 crc kubenswrapper[4766]: I1126 00:30:00.137955 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 26 00:30:00 crc kubenswrapper[4766]: I1126 00:30:00.142143 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 26 00:30:00 crc kubenswrapper[4766]: I1126 00:30:00.146254 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z"] Nov 26 00:30:00 crc kubenswrapper[4766]: I1126 00:30:00.313698 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/216d430f-aabb-4376-a764-0b1f07ff228b-config-volume\") pod \"collect-profiles-29401950-66m4z\" (UID: \"216d430f-aabb-4376-a764-0b1f07ff228b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z" Nov 26 00:30:00 crc kubenswrapper[4766]: I1126 00:30:00.314033 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbsrt\" (UniqueName: \"kubernetes.io/projected/216d430f-aabb-4376-a764-0b1f07ff228b-kube-api-access-lbsrt\") pod \"collect-profiles-29401950-66m4z\" (UID: \"216d430f-aabb-4376-a764-0b1f07ff228b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z" Nov 26 00:30:00 crc kubenswrapper[4766]: I1126 00:30:00.314072 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/216d430f-aabb-4376-a764-0b1f07ff228b-secret-volume\") pod \"collect-profiles-29401950-66m4z\" (UID: \"216d430f-aabb-4376-a764-0b1f07ff228b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z" Nov 26 00:30:00 crc kubenswrapper[4766]: I1126 00:30:00.415346 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/216d430f-aabb-4376-a764-0b1f07ff228b-config-volume\") pod \"collect-profiles-29401950-66m4z\" (UID: \"216d430f-aabb-4376-a764-0b1f07ff228b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z" Nov 26 00:30:00 crc kubenswrapper[4766]: I1126 00:30:00.415396 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbsrt\" (UniqueName: \"kubernetes.io/projected/216d430f-aabb-4376-a764-0b1f07ff228b-kube-api-access-lbsrt\") pod \"collect-profiles-29401950-66m4z\" (UID: \"216d430f-aabb-4376-a764-0b1f07ff228b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z" Nov 26 00:30:00 crc kubenswrapper[4766]: I1126 00:30:00.415429 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/216d430f-aabb-4376-a764-0b1f07ff228b-secret-volume\") pod \"collect-profiles-29401950-66m4z\" (UID: \"216d430f-aabb-4376-a764-0b1f07ff228b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z" Nov 26 00:30:00 crc kubenswrapper[4766]: I1126 00:30:00.416209 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/216d430f-aabb-4376-a764-0b1f07ff228b-config-volume\") pod \"collect-profiles-29401950-66m4z\" (UID: \"216d430f-aabb-4376-a764-0b1f07ff228b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z" Nov 26 00:30:00 crc kubenswrapper[4766]: I1126 00:30:00.424307 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/216d430f-aabb-4376-a764-0b1f07ff228b-secret-volume\") pod \"collect-profiles-29401950-66m4z\" (UID: \"216d430f-aabb-4376-a764-0b1f07ff228b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z" Nov 26 00:30:00 crc kubenswrapper[4766]: I1126 00:30:00.443987 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbsrt\" (UniqueName: \"kubernetes.io/projected/216d430f-aabb-4376-a764-0b1f07ff228b-kube-api-access-lbsrt\") pod \"collect-profiles-29401950-66m4z\" (UID: \"216d430f-aabb-4376-a764-0b1f07ff228b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z" Nov 26 00:30:00 crc kubenswrapper[4766]: I1126 00:30:00.466173 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z" Nov 26 00:30:00 crc kubenswrapper[4766]: I1126 00:30:00.652890 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z"] Nov 26 00:30:00 crc kubenswrapper[4766]: I1126 00:30:00.751584 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z" event={"ID":"216d430f-aabb-4376-a764-0b1f07ff228b","Type":"ContainerStarted","Data":"fa110cb0548ef12c7cfc3199738b4129a6cb5c31e797860dc20fa8e0670886e9"} Nov 26 00:30:01 crc kubenswrapper[4766]: I1126 00:30:01.758020 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z" event={"ID":"216d430f-aabb-4376-a764-0b1f07ff228b","Type":"ContainerStarted","Data":"d79314a1058accc10891514c9da7003c771b394ca5edac26e6fa5b0c8e409794"} Nov 26 00:30:01 crc kubenswrapper[4766]: I1126 00:30:01.772234 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z" podStartSLOduration=1.7722186720000002 podStartE2EDuration="1.772218672s" podCreationTimestamp="2025-11-26 00:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:30:01.768967107 +0000 UTC m=+382.617737557" watchObservedRunningTime="2025-11-26 00:30:01.772218672 +0000 UTC m=+382.620989102" Nov 26 00:30:02 crc kubenswrapper[4766]: I1126 00:30:02.765328 4766 generic.go:334] "Generic (PLEG): container finished" podID="216d430f-aabb-4376-a764-0b1f07ff228b" containerID="d79314a1058accc10891514c9da7003c771b394ca5edac26e6fa5b0c8e409794" exitCode=0 Nov 26 00:30:02 crc kubenswrapper[4766]: I1126 00:30:02.765393 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z" event={"ID":"216d430f-aabb-4376-a764-0b1f07ff228b","Type":"ContainerDied","Data":"d79314a1058accc10891514c9da7003c771b394ca5edac26e6fa5b0c8e409794"} Nov 26 00:30:03 crc kubenswrapper[4766]: I1126 00:30:03.953223 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z" Nov 26 00:30:04 crc kubenswrapper[4766]: I1126 00:30:04.058889 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/216d430f-aabb-4376-a764-0b1f07ff228b-secret-volume\") pod \"216d430f-aabb-4376-a764-0b1f07ff228b\" (UID: \"216d430f-aabb-4376-a764-0b1f07ff228b\") " Nov 26 00:30:04 crc kubenswrapper[4766]: I1126 00:30:04.058946 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbsrt\" (UniqueName: \"kubernetes.io/projected/216d430f-aabb-4376-a764-0b1f07ff228b-kube-api-access-lbsrt\") pod \"216d430f-aabb-4376-a764-0b1f07ff228b\" (UID: \"216d430f-aabb-4376-a764-0b1f07ff228b\") " Nov 26 00:30:04 crc kubenswrapper[4766]: I1126 00:30:04.059139 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/216d430f-aabb-4376-a764-0b1f07ff228b-config-volume\") pod \"216d430f-aabb-4376-a764-0b1f07ff228b\" (UID: \"216d430f-aabb-4376-a764-0b1f07ff228b\") " Nov 26 00:30:04 crc kubenswrapper[4766]: I1126 00:30:04.060630 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/216d430f-aabb-4376-a764-0b1f07ff228b-config-volume" (OuterVolumeSpecName: "config-volume") pod "216d430f-aabb-4376-a764-0b1f07ff228b" (UID: "216d430f-aabb-4376-a764-0b1f07ff228b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:30:04 crc kubenswrapper[4766]: I1126 00:30:04.066532 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/216d430f-aabb-4376-a764-0b1f07ff228b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "216d430f-aabb-4376-a764-0b1f07ff228b" (UID: "216d430f-aabb-4376-a764-0b1f07ff228b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:30:04 crc kubenswrapper[4766]: I1126 00:30:04.066782 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/216d430f-aabb-4376-a764-0b1f07ff228b-kube-api-access-lbsrt" (OuterVolumeSpecName: "kube-api-access-lbsrt") pod "216d430f-aabb-4376-a764-0b1f07ff228b" (UID: "216d430f-aabb-4376-a764-0b1f07ff228b"). InnerVolumeSpecName "kube-api-access-lbsrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:30:04 crc kubenswrapper[4766]: I1126 00:30:04.160725 4766 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/216d430f-aabb-4376-a764-0b1f07ff228b-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 00:30:04 crc kubenswrapper[4766]: I1126 00:30:04.160762 4766 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/216d430f-aabb-4376-a764-0b1f07ff228b-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 26 00:30:04 crc kubenswrapper[4766]: I1126 00:30:04.160772 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbsrt\" (UniqueName: \"kubernetes.io/projected/216d430f-aabb-4376-a764-0b1f07ff228b-kube-api-access-lbsrt\") on node \"crc\" DevicePath \"\"" Nov 26 00:30:04 crc kubenswrapper[4766]: I1126 00:30:04.777681 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z" event={"ID":"216d430f-aabb-4376-a764-0b1f07ff228b","Type":"ContainerDied","Data":"fa110cb0548ef12c7cfc3199738b4129a6cb5c31e797860dc20fa8e0670886e9"} Nov 26 00:30:04 crc kubenswrapper[4766]: I1126 00:30:04.778153 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa110cb0548ef12c7cfc3199738b4129a6cb5c31e797860dc20fa8e0670886e9" Nov 26 00:30:04 crc kubenswrapper[4766]: I1126 00:30:04.777757 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z" Nov 26 00:30:11 crc kubenswrapper[4766]: I1126 00:30:11.479349 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:30:11 crc kubenswrapper[4766]: I1126 00:30:11.480135 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:30:11 crc kubenswrapper[4766]: I1126 00:30:11.480206 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:30:11 crc kubenswrapper[4766]: I1126 00:30:11.481071 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8e60726cc9ad1a2202d54a17584e9c974ee0288e42d42f37ee8591c1fcee8e0e"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 00:30:11 crc kubenswrapper[4766]: I1126 00:30:11.481172 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://8e60726cc9ad1a2202d54a17584e9c974ee0288e42d42f37ee8591c1fcee8e0e" gracePeriod=600 Nov 26 00:30:11 crc kubenswrapper[4766]: I1126 00:30:11.821580 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="8e60726cc9ad1a2202d54a17584e9c974ee0288e42d42f37ee8591c1fcee8e0e" exitCode=0 Nov 26 00:30:11 crc kubenswrapper[4766]: I1126 00:30:11.821631 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"8e60726cc9ad1a2202d54a17584e9c974ee0288e42d42f37ee8591c1fcee8e0e"} Nov 26 00:30:11 crc kubenswrapper[4766]: I1126 00:30:11.821683 4766 scope.go:117] "RemoveContainer" containerID="98c675a281d794263e5afbb564287dfa523f6fd6b3e5720996acd276206f35cf" Nov 26 00:30:13 crc kubenswrapper[4766]: I1126 00:30:13.839416 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"ba0d2e8d44b0113663280906db232a40bb0eaded73c2a35e07448bc5254dc338"} Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.221329 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-s29xf"] Nov 26 00:30:38 crc kubenswrapper[4766]: E1126 00:30:38.222817 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="216d430f-aabb-4376-a764-0b1f07ff228b" containerName="collect-profiles" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.222841 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="216d430f-aabb-4376-a764-0b1f07ff228b" containerName="collect-profiles" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.223012 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="216d430f-aabb-4376-a764-0b1f07ff228b" containerName="collect-profiles" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.223639 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.244121 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-s29xf"] Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.295847 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v28p9\" (UniqueName: \"kubernetes.io/projected/dfb3bebb-f815-47b6-89be-6d04fd900ccb-kube-api-access-v28p9\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.295891 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dfb3bebb-f815-47b6-89be-6d04fd900ccb-trusted-ca\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.295932 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dfb3bebb-f815-47b6-89be-6d04fd900ccb-bound-sa-token\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.295989 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/dfb3bebb-f815-47b6-89be-6d04fd900ccb-registry-tls\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.296026 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/dfb3bebb-f815-47b6-89be-6d04fd900ccb-ca-trust-extracted\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.296068 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/dfb3bebb-f815-47b6-89be-6d04fd900ccb-installation-pull-secrets\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.296094 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/dfb3bebb-f815-47b6-89be-6d04fd900ccb-registry-certificates\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.296148 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.316093 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.397847 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/dfb3bebb-f815-47b6-89be-6d04fd900ccb-installation-pull-secrets\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.397942 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/dfb3bebb-f815-47b6-89be-6d04fd900ccb-registry-certificates\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.398013 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v28p9\" (UniqueName: \"kubernetes.io/projected/dfb3bebb-f815-47b6-89be-6d04fd900ccb-kube-api-access-v28p9\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.398039 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dfb3bebb-f815-47b6-89be-6d04fd900ccb-trusted-ca\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.398079 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dfb3bebb-f815-47b6-89be-6d04fd900ccb-bound-sa-token\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.398112 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/dfb3bebb-f815-47b6-89be-6d04fd900ccb-registry-tls\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.398134 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/dfb3bebb-f815-47b6-89be-6d04fd900ccb-ca-trust-extracted\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.399012 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/dfb3bebb-f815-47b6-89be-6d04fd900ccb-ca-trust-extracted\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.400143 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dfb3bebb-f815-47b6-89be-6d04fd900ccb-trusted-ca\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.400511 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/dfb3bebb-f815-47b6-89be-6d04fd900ccb-registry-certificates\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.402897 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/dfb3bebb-f815-47b6-89be-6d04fd900ccb-installation-pull-secrets\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.407326 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/dfb3bebb-f815-47b6-89be-6d04fd900ccb-registry-tls\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.413733 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dfb3bebb-f815-47b6-89be-6d04fd900ccb-bound-sa-token\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.416155 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v28p9\" (UniqueName: \"kubernetes.io/projected/dfb3bebb-f815-47b6-89be-6d04fd900ccb-kube-api-access-v28p9\") pod \"image-registry-66df7c8f76-s29xf\" (UID: \"dfb3bebb-f815-47b6-89be-6d04fd900ccb\") " pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.547298 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.748340 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-s29xf"] Nov 26 00:30:38 crc kubenswrapper[4766]: W1126 00:30:38.757270 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddfb3bebb_f815_47b6_89be_6d04fd900ccb.slice/crio-7221d57a8675ecc0a9f4618b0120ffc129563913016dd6b96e1e1c6787136221 WatchSource:0}: Error finding container 7221d57a8675ecc0a9f4618b0120ffc129563913016dd6b96e1e1c6787136221: Status 404 returned error can't find the container with id 7221d57a8675ecc0a9f4618b0120ffc129563913016dd6b96e1e1c6787136221 Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.997475 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" event={"ID":"dfb3bebb-f815-47b6-89be-6d04fd900ccb","Type":"ContainerStarted","Data":"036fd3bd78af713ed5194bb927bb279b4349b34ee25dd4f30498831812941c1c"} Nov 26 00:30:38 crc kubenswrapper[4766]: I1126 00:30:38.997562 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" event={"ID":"dfb3bebb-f815-47b6-89be-6d04fd900ccb","Type":"ContainerStarted","Data":"7221d57a8675ecc0a9f4618b0120ffc129563913016dd6b96e1e1c6787136221"} Nov 26 00:30:39 crc kubenswrapper[4766]: I1126 00:30:39.017294 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" podStartSLOduration=1.017272294 podStartE2EDuration="1.017272294s" podCreationTimestamp="2025-11-26 00:30:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:30:39.012781938 +0000 UTC m=+419.861552388" watchObservedRunningTime="2025-11-26 00:30:39.017272294 +0000 UTC m=+419.866042724" Nov 26 00:30:40 crc kubenswrapper[4766]: I1126 00:30:40.001153 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:58 crc kubenswrapper[4766]: I1126 00:30:58.556935 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" Nov 26 00:30:58 crc kubenswrapper[4766]: I1126 00:30:58.625145 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-m8qgq"] Nov 26 00:31:23 crc kubenswrapper[4766]: I1126 00:31:23.670362 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" podUID="1e392e29-fdb8-450d-ac24-ba91e802139a" containerName="registry" containerID="cri-o://8bacca62923ee72b2716bbdd37654c255ec6777757d32bdf7e2550d337052e07" gracePeriod=30 Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.239939 4766 generic.go:334] "Generic (PLEG): container finished" podID="1e392e29-fdb8-450d-ac24-ba91e802139a" containerID="8bacca62923ee72b2716bbdd37654c255ec6777757d32bdf7e2550d337052e07" exitCode=0 Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.240032 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" event={"ID":"1e392e29-fdb8-450d-ac24-ba91e802139a","Type":"ContainerDied","Data":"8bacca62923ee72b2716bbdd37654c255ec6777757d32bdf7e2550d337052e07"} Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.546146 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.658857 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"1e392e29-fdb8-450d-ac24-ba91e802139a\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.658908 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1e392e29-fdb8-450d-ac24-ba91e802139a-registry-tls\") pod \"1e392e29-fdb8-450d-ac24-ba91e802139a\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.658988 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1e392e29-fdb8-450d-ac24-ba91e802139a-installation-pull-secrets\") pod \"1e392e29-fdb8-450d-ac24-ba91e802139a\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.659035 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1e392e29-fdb8-450d-ac24-ba91e802139a-ca-trust-extracted\") pod \"1e392e29-fdb8-450d-ac24-ba91e802139a\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.659068 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1e392e29-fdb8-450d-ac24-ba91e802139a-bound-sa-token\") pod \"1e392e29-fdb8-450d-ac24-ba91e802139a\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.659115 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1e392e29-fdb8-450d-ac24-ba91e802139a-trusted-ca\") pod \"1e392e29-fdb8-450d-ac24-ba91e802139a\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.659146 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvt7w\" (UniqueName: \"kubernetes.io/projected/1e392e29-fdb8-450d-ac24-ba91e802139a-kube-api-access-mvt7w\") pod \"1e392e29-fdb8-450d-ac24-ba91e802139a\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.659172 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1e392e29-fdb8-450d-ac24-ba91e802139a-registry-certificates\") pod \"1e392e29-fdb8-450d-ac24-ba91e802139a\" (UID: \"1e392e29-fdb8-450d-ac24-ba91e802139a\") " Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.659877 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e392e29-fdb8-450d-ac24-ba91e802139a-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "1e392e29-fdb8-450d-ac24-ba91e802139a" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.659931 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e392e29-fdb8-450d-ac24-ba91e802139a-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "1e392e29-fdb8-450d-ac24-ba91e802139a" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.665767 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e392e29-fdb8-450d-ac24-ba91e802139a-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "1e392e29-fdb8-450d-ac24-ba91e802139a" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.665966 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e392e29-fdb8-450d-ac24-ba91e802139a-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "1e392e29-fdb8-450d-ac24-ba91e802139a" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.666548 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e392e29-fdb8-450d-ac24-ba91e802139a-kube-api-access-mvt7w" (OuterVolumeSpecName: "kube-api-access-mvt7w") pod "1e392e29-fdb8-450d-ac24-ba91e802139a" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a"). InnerVolumeSpecName "kube-api-access-mvt7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.668142 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e392e29-fdb8-450d-ac24-ba91e802139a-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "1e392e29-fdb8-450d-ac24-ba91e802139a" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.673467 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "1e392e29-fdb8-450d-ac24-ba91e802139a" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.677262 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e392e29-fdb8-450d-ac24-ba91e802139a-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "1e392e29-fdb8-450d-ac24-ba91e802139a" (UID: "1e392e29-fdb8-450d-ac24-ba91e802139a"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.760371 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvt7w\" (UniqueName: \"kubernetes.io/projected/1e392e29-fdb8-450d-ac24-ba91e802139a-kube-api-access-mvt7w\") on node \"crc\" DevicePath \"\"" Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.760412 4766 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1e392e29-fdb8-450d-ac24-ba91e802139a-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.760422 4766 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1e392e29-fdb8-450d-ac24-ba91e802139a-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.760433 4766 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1e392e29-fdb8-450d-ac24-ba91e802139a-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.760442 4766 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1e392e29-fdb8-450d-ac24-ba91e802139a-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.760450 4766 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1e392e29-fdb8-450d-ac24-ba91e802139a-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 26 00:31:24 crc kubenswrapper[4766]: I1126 00:31:24.760457 4766 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1e392e29-fdb8-450d-ac24-ba91e802139a-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:31:25 crc kubenswrapper[4766]: I1126 00:31:25.247462 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" event={"ID":"1e392e29-fdb8-450d-ac24-ba91e802139a","Type":"ContainerDied","Data":"58f4d4bc35d4f8a388a6bfb7a6bedc2e77fe9e2a9c077d21e3a31da75826cb7b"} Nov 26 00:31:25 crc kubenswrapper[4766]: I1126 00:31:25.247519 4766 scope.go:117] "RemoveContainer" containerID="8bacca62923ee72b2716bbdd37654c255ec6777757d32bdf7e2550d337052e07" Nov 26 00:31:25 crc kubenswrapper[4766]: I1126 00:31:25.247620 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-m8qgq" Nov 26 00:31:25 crc kubenswrapper[4766]: I1126 00:31:25.274581 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-m8qgq"] Nov 26 00:31:25 crc kubenswrapper[4766]: I1126 00:31:25.278982 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-m8qgq"] Nov 26 00:31:25 crc kubenswrapper[4766]: I1126 00:31:25.837531 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e392e29-fdb8-450d-ac24-ba91e802139a" path="/var/lib/kubelet/pods/1e392e29-fdb8-450d-ac24-ba91e802139a/volumes" Nov 26 00:32:39 crc kubenswrapper[4766]: I1126 00:32:39.985054 4766 scope.go:117] "RemoveContainer" containerID="da26f486c36b0ad61f384ad1c45e78be590cd00ef690601b7ef7ee2a45548810" Nov 26 00:32:41 crc kubenswrapper[4766]: I1126 00:32:41.479912 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:32:41 crc kubenswrapper[4766]: I1126 00:32:41.480409 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:33:11 crc kubenswrapper[4766]: I1126 00:33:11.480024 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:33:11 crc kubenswrapper[4766]: I1126 00:33:11.480581 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:33:37 crc kubenswrapper[4766]: I1126 00:33:37.143475 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq"] Nov 26 00:33:37 crc kubenswrapper[4766]: E1126 00:33:37.144199 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e392e29-fdb8-450d-ac24-ba91e802139a" containerName="registry" Nov 26 00:33:37 crc kubenswrapper[4766]: I1126 00:33:37.144211 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e392e29-fdb8-450d-ac24-ba91e802139a" containerName="registry" Nov 26 00:33:37 crc kubenswrapper[4766]: I1126 00:33:37.144305 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e392e29-fdb8-450d-ac24-ba91e802139a" containerName="registry" Nov 26 00:33:37 crc kubenswrapper[4766]: I1126 00:33:37.144998 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq" Nov 26 00:33:37 crc kubenswrapper[4766]: I1126 00:33:37.148295 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 26 00:33:37 crc kubenswrapper[4766]: I1126 00:33:37.158778 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq"] Nov 26 00:33:37 crc kubenswrapper[4766]: I1126 00:33:37.301599 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4d00d8e8-b83e-4585-9678-fb64e663fec3-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq\" (UID: \"4d00d8e8-b83e-4585-9678-fb64e663fec3\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq" Nov 26 00:33:37 crc kubenswrapper[4766]: I1126 00:33:37.301739 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4d00d8e8-b83e-4585-9678-fb64e663fec3-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq\" (UID: \"4d00d8e8-b83e-4585-9678-fb64e663fec3\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq" Nov 26 00:33:37 crc kubenswrapper[4766]: I1126 00:33:37.301881 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c48c\" (UniqueName: \"kubernetes.io/projected/4d00d8e8-b83e-4585-9678-fb64e663fec3-kube-api-access-4c48c\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq\" (UID: \"4d00d8e8-b83e-4585-9678-fb64e663fec3\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq" Nov 26 00:33:37 crc kubenswrapper[4766]: I1126 00:33:37.402856 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4d00d8e8-b83e-4585-9678-fb64e663fec3-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq\" (UID: \"4d00d8e8-b83e-4585-9678-fb64e663fec3\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq" Nov 26 00:33:37 crc kubenswrapper[4766]: I1126 00:33:37.402934 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c48c\" (UniqueName: \"kubernetes.io/projected/4d00d8e8-b83e-4585-9678-fb64e663fec3-kube-api-access-4c48c\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq\" (UID: \"4d00d8e8-b83e-4585-9678-fb64e663fec3\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq" Nov 26 00:33:37 crc kubenswrapper[4766]: I1126 00:33:37.402974 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4d00d8e8-b83e-4585-9678-fb64e663fec3-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq\" (UID: \"4d00d8e8-b83e-4585-9678-fb64e663fec3\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq" Nov 26 00:33:37 crc kubenswrapper[4766]: I1126 00:33:37.403347 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4d00d8e8-b83e-4585-9678-fb64e663fec3-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq\" (UID: \"4d00d8e8-b83e-4585-9678-fb64e663fec3\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq" Nov 26 00:33:37 crc kubenswrapper[4766]: I1126 00:33:37.403457 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4d00d8e8-b83e-4585-9678-fb64e663fec3-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq\" (UID: \"4d00d8e8-b83e-4585-9678-fb64e663fec3\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq" Nov 26 00:33:37 crc kubenswrapper[4766]: I1126 00:33:37.422621 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c48c\" (UniqueName: \"kubernetes.io/projected/4d00d8e8-b83e-4585-9678-fb64e663fec3-kube-api-access-4c48c\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq\" (UID: \"4d00d8e8-b83e-4585-9678-fb64e663fec3\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq" Nov 26 00:33:37 crc kubenswrapper[4766]: I1126 00:33:37.472499 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq" Nov 26 00:33:37 crc kubenswrapper[4766]: I1126 00:33:37.697674 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq"] Nov 26 00:33:37 crc kubenswrapper[4766]: I1126 00:33:37.984916 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq" event={"ID":"4d00d8e8-b83e-4585-9678-fb64e663fec3","Type":"ContainerStarted","Data":"1699ee2d81185c637f24f368603aa1d7de17014825a24caffed00595c5fda225"} Nov 26 00:33:37 crc kubenswrapper[4766]: I1126 00:33:37.984973 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq" event={"ID":"4d00d8e8-b83e-4585-9678-fb64e663fec3","Type":"ContainerStarted","Data":"e024a2e77ab75080a537f60f1e6bca2defff890a9c2da72377f898ee6ea16bb8"} Nov 26 00:33:38 crc kubenswrapper[4766]: I1126 00:33:38.992462 4766 generic.go:334] "Generic (PLEG): container finished" podID="4d00d8e8-b83e-4585-9678-fb64e663fec3" containerID="1699ee2d81185c637f24f368603aa1d7de17014825a24caffed00595c5fda225" exitCode=0 Nov 26 00:33:38 crc kubenswrapper[4766]: I1126 00:33:38.992558 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq" event={"ID":"4d00d8e8-b83e-4585-9678-fb64e663fec3","Type":"ContainerDied","Data":"1699ee2d81185c637f24f368603aa1d7de17014825a24caffed00595c5fda225"} Nov 26 00:33:38 crc kubenswrapper[4766]: I1126 00:33:38.994721 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 00:33:40 crc kubenswrapper[4766]: I1126 00:33:40.022808 4766 scope.go:117] "RemoveContainer" containerID="cc95df5b82620cb48f2b8c099b43cbc580d65b8ead0fad0d7113134b93ce2a8b" Nov 26 00:33:40 crc kubenswrapper[4766]: I1126 00:33:40.041297 4766 scope.go:117] "RemoveContainer" containerID="cfa545f753933cfde8402e62bc238067b73040b42c1ba6861d277c2ab175906b" Nov 26 00:33:41 crc kubenswrapper[4766]: I1126 00:33:41.480109 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:33:41 crc kubenswrapper[4766]: I1126 00:33:41.480171 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:33:41 crc kubenswrapper[4766]: I1126 00:33:41.480224 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:33:41 crc kubenswrapper[4766]: I1126 00:33:41.480836 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ba0d2e8d44b0113663280906db232a40bb0eaded73c2a35e07448bc5254dc338"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 00:33:41 crc kubenswrapper[4766]: I1126 00:33:41.480891 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://ba0d2e8d44b0113663280906db232a40bb0eaded73c2a35e07448bc5254dc338" gracePeriod=600 Nov 26 00:33:42 crc kubenswrapper[4766]: I1126 00:33:42.019018 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="ba0d2e8d44b0113663280906db232a40bb0eaded73c2a35e07448bc5254dc338" exitCode=0 Nov 26 00:33:42 crc kubenswrapper[4766]: I1126 00:33:42.019126 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"ba0d2e8d44b0113663280906db232a40bb0eaded73c2a35e07448bc5254dc338"} Nov 26 00:33:42 crc kubenswrapper[4766]: I1126 00:33:42.019930 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"2f5bd86ee9ff3b132bc7816555a18731d6351ecdb67cc2c61c0c55d699b956fc"} Nov 26 00:33:42 crc kubenswrapper[4766]: I1126 00:33:42.019997 4766 scope.go:117] "RemoveContainer" containerID="8e60726cc9ad1a2202d54a17584e9c974ee0288e42d42f37ee8591c1fcee8e0e" Nov 26 00:33:43 crc kubenswrapper[4766]: I1126 00:33:43.030031 4766 generic.go:334] "Generic (PLEG): container finished" podID="4d00d8e8-b83e-4585-9678-fb64e663fec3" containerID="0ff38f7b0c3e11d97241da2841b0e82eae6bcdb11ac2598d2a80ae89b37fc69f" exitCode=0 Nov 26 00:33:43 crc kubenswrapper[4766]: I1126 00:33:43.030102 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq" event={"ID":"4d00d8e8-b83e-4585-9678-fb64e663fec3","Type":"ContainerDied","Data":"0ff38f7b0c3e11d97241da2841b0e82eae6bcdb11ac2598d2a80ae89b37fc69f"} Nov 26 00:33:44 crc kubenswrapper[4766]: I1126 00:33:44.040792 4766 generic.go:334] "Generic (PLEG): container finished" podID="4d00d8e8-b83e-4585-9678-fb64e663fec3" containerID="9f6c291f3e42058242b09447023ab181bb198401488d1f727e6a3670877c974e" exitCode=0 Nov 26 00:33:44 crc kubenswrapper[4766]: I1126 00:33:44.040876 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq" event={"ID":"4d00d8e8-b83e-4585-9678-fb64e663fec3","Type":"ContainerDied","Data":"9f6c291f3e42058242b09447023ab181bb198401488d1f727e6a3670877c974e"} Nov 26 00:33:45 crc kubenswrapper[4766]: I1126 00:33:45.315422 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq" Nov 26 00:33:45 crc kubenswrapper[4766]: I1126 00:33:45.420133 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4d00d8e8-b83e-4585-9678-fb64e663fec3-util\") pod \"4d00d8e8-b83e-4585-9678-fb64e663fec3\" (UID: \"4d00d8e8-b83e-4585-9678-fb64e663fec3\") " Nov 26 00:33:45 crc kubenswrapper[4766]: I1126 00:33:45.424976 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4d00d8e8-b83e-4585-9678-fb64e663fec3-bundle\") pod \"4d00d8e8-b83e-4585-9678-fb64e663fec3\" (UID: \"4d00d8e8-b83e-4585-9678-fb64e663fec3\") " Nov 26 00:33:45 crc kubenswrapper[4766]: I1126 00:33:45.425063 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4c48c\" (UniqueName: \"kubernetes.io/projected/4d00d8e8-b83e-4585-9678-fb64e663fec3-kube-api-access-4c48c\") pod \"4d00d8e8-b83e-4585-9678-fb64e663fec3\" (UID: \"4d00d8e8-b83e-4585-9678-fb64e663fec3\") " Nov 26 00:33:45 crc kubenswrapper[4766]: I1126 00:33:45.427107 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d00d8e8-b83e-4585-9678-fb64e663fec3-bundle" (OuterVolumeSpecName: "bundle") pod "4d00d8e8-b83e-4585-9678-fb64e663fec3" (UID: "4d00d8e8-b83e-4585-9678-fb64e663fec3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:33:45 crc kubenswrapper[4766]: I1126 00:33:45.430601 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d00d8e8-b83e-4585-9678-fb64e663fec3-util" (OuterVolumeSpecName: "util") pod "4d00d8e8-b83e-4585-9678-fb64e663fec3" (UID: "4d00d8e8-b83e-4585-9678-fb64e663fec3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:33:45 crc kubenswrapper[4766]: I1126 00:33:45.430705 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d00d8e8-b83e-4585-9678-fb64e663fec3-kube-api-access-4c48c" (OuterVolumeSpecName: "kube-api-access-4c48c") pod "4d00d8e8-b83e-4585-9678-fb64e663fec3" (UID: "4d00d8e8-b83e-4585-9678-fb64e663fec3"). InnerVolumeSpecName "kube-api-access-4c48c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:33:45 crc kubenswrapper[4766]: I1126 00:33:45.526569 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4c48c\" (UniqueName: \"kubernetes.io/projected/4d00d8e8-b83e-4585-9678-fb64e663fec3-kube-api-access-4c48c\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:45 crc kubenswrapper[4766]: I1126 00:33:45.526618 4766 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4d00d8e8-b83e-4585-9678-fb64e663fec3-util\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:45 crc kubenswrapper[4766]: I1126 00:33:45.526629 4766 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4d00d8e8-b83e-4585-9678-fb64e663fec3-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:46 crc kubenswrapper[4766]: I1126 00:33:46.054971 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq" event={"ID":"4d00d8e8-b83e-4585-9678-fb64e663fec3","Type":"ContainerDied","Data":"e024a2e77ab75080a537f60f1e6bca2defff890a9c2da72377f898ee6ea16bb8"} Nov 26 00:33:46 crc kubenswrapper[4766]: I1126 00:33:46.055010 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e024a2e77ab75080a537f60f1e6bca2defff890a9c2da72377f898ee6ea16bb8" Nov 26 00:33:46 crc kubenswrapper[4766]: I1126 00:33:46.055132 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq" Nov 26 00:33:48 crc kubenswrapper[4766]: I1126 00:33:48.576926 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fxnln"] Nov 26 00:33:48 crc kubenswrapper[4766]: I1126 00:33:48.577675 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovn-controller" containerID="cri-o://abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b" gracePeriod=30 Nov 26 00:33:48 crc kubenswrapper[4766]: I1126 00:33:48.577724 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="nbdb" containerID="cri-o://ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff" gracePeriod=30 Nov 26 00:33:48 crc kubenswrapper[4766]: I1126 00:33:48.577785 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="sbdb" containerID="cri-o://4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da" gracePeriod=30 Nov 26 00:33:48 crc kubenswrapper[4766]: I1126 00:33:48.577738 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820" gracePeriod=30 Nov 26 00:33:48 crc kubenswrapper[4766]: I1126 00:33:48.577839 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovn-acl-logging" containerID="cri-o://cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99" gracePeriod=30 Nov 26 00:33:48 crc kubenswrapper[4766]: I1126 00:33:48.577853 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="northd" containerID="cri-o://4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52" gracePeriod=30 Nov 26 00:33:48 crc kubenswrapper[4766]: I1126 00:33:48.577908 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="kube-rbac-proxy-node" containerID="cri-o://619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498" gracePeriod=30 Nov 26 00:33:48 crc kubenswrapper[4766]: I1126 00:33:48.619964 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovnkube-controller" containerID="cri-o://2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1" gracePeriod=30 Nov 26 00:33:48 crc kubenswrapper[4766]: I1126 00:33:48.929461 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fxnln_415fdacb-85c1-4265-89b7-6771a84ffc89/ovnkube-controller/3.log" Nov 26 00:33:48 crc kubenswrapper[4766]: I1126 00:33:48.944474 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fxnln_415fdacb-85c1-4265-89b7-6771a84ffc89/ovn-acl-logging/0.log" Nov 26 00:33:48 crc kubenswrapper[4766]: I1126 00:33:48.945021 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fxnln_415fdacb-85c1-4265-89b7-6771a84ffc89/ovn-controller/0.log" Nov 26 00:33:48 crc kubenswrapper[4766]: I1126 00:33:48.945422 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.090554 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/415fdacb-85c1-4265-89b7-6771a84ffc89-env-overrides\") pod \"415fdacb-85c1-4265-89b7-6771a84ffc89\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.091249 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-run-ovn-kubernetes\") pod \"415fdacb-85c1-4265-89b7-6771a84ffc89\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.091324 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l696l\" (UniqueName: \"kubernetes.io/projected/415fdacb-85c1-4265-89b7-6771a84ffc89-kube-api-access-l696l\") pod \"415fdacb-85c1-4265-89b7-6771a84ffc89\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.091355 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-log-socket\") pod \"415fdacb-85c1-4265-89b7-6771a84ffc89\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.091397 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-var-lib-openvswitch\") pod \"415fdacb-85c1-4265-89b7-6771a84ffc89\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.091433 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/415fdacb-85c1-4265-89b7-6771a84ffc89-ovnkube-script-lib\") pod \"415fdacb-85c1-4265-89b7-6771a84ffc89\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.091451 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-node-log\") pod \"415fdacb-85c1-4265-89b7-6771a84ffc89\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.091472 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-run-netns\") pod \"415fdacb-85c1-4265-89b7-6771a84ffc89\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.091501 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/415fdacb-85c1-4265-89b7-6771a84ffc89-ovnkube-config\") pod \"415fdacb-85c1-4265-89b7-6771a84ffc89\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.091536 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-cni-netd\") pod \"415fdacb-85c1-4265-89b7-6771a84ffc89\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.091560 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-run-systemd\") pod \"415fdacb-85c1-4265-89b7-6771a84ffc89\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.091593 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-run-openvswitch\") pod \"415fdacb-85c1-4265-89b7-6771a84ffc89\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.091617 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-kubelet\") pod \"415fdacb-85c1-4265-89b7-6771a84ffc89\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.091640 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-run-ovn\") pod \"415fdacb-85c1-4265-89b7-6771a84ffc89\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.091689 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-slash\") pod \"415fdacb-85c1-4265-89b7-6771a84ffc89\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.091717 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-etc-openvswitch\") pod \"415fdacb-85c1-4265-89b7-6771a84ffc89\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.091769 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/415fdacb-85c1-4265-89b7-6771a84ffc89-ovn-node-metrics-cert\") pod \"415fdacb-85c1-4265-89b7-6771a84ffc89\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.091801 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-systemd-units\") pod \"415fdacb-85c1-4265-89b7-6771a84ffc89\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.091829 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-cni-bin\") pod \"415fdacb-85c1-4265-89b7-6771a84ffc89\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.091851 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-var-lib-cni-networks-ovn-kubernetes\") pod \"415fdacb-85c1-4265-89b7-6771a84ffc89\" (UID: \"415fdacb-85c1-4265-89b7-6771a84ffc89\") " Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.091089 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/415fdacb-85c1-4265-89b7-6771a84ffc89-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "415fdacb-85c1-4265-89b7-6771a84ffc89" (UID: "415fdacb-85c1-4265-89b7-6771a84ffc89"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.092137 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/415fdacb-85c1-4265-89b7-6771a84ffc89-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "415fdacb-85c1-4265-89b7-6771a84ffc89" (UID: "415fdacb-85c1-4265-89b7-6771a84ffc89"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.092215 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "415fdacb-85c1-4265-89b7-6771a84ffc89" (UID: "415fdacb-85c1-4265-89b7-6771a84ffc89"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.092255 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "415fdacb-85c1-4265-89b7-6771a84ffc89" (UID: "415fdacb-85c1-4265-89b7-6771a84ffc89"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.092918 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "415fdacb-85c1-4265-89b7-6771a84ffc89" (UID: "415fdacb-85c1-4265-89b7-6771a84ffc89"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.092952 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "415fdacb-85c1-4265-89b7-6771a84ffc89" (UID: "415fdacb-85c1-4265-89b7-6771a84ffc89"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.092977 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "415fdacb-85c1-4265-89b7-6771a84ffc89" (UID: "415fdacb-85c1-4265-89b7-6771a84ffc89"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.093000 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-slash" (OuterVolumeSpecName: "host-slash") pod "415fdacb-85c1-4265-89b7-6771a84ffc89" (UID: "415fdacb-85c1-4265-89b7-6771a84ffc89"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.093026 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "415fdacb-85c1-4265-89b7-6771a84ffc89" (UID: "415fdacb-85c1-4265-89b7-6771a84ffc89"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.093160 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "415fdacb-85c1-4265-89b7-6771a84ffc89" (UID: "415fdacb-85c1-4265-89b7-6771a84ffc89"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.093197 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "415fdacb-85c1-4265-89b7-6771a84ffc89" (UID: "415fdacb-85c1-4265-89b7-6771a84ffc89"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.093227 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "415fdacb-85c1-4265-89b7-6771a84ffc89" (UID: "415fdacb-85c1-4265-89b7-6771a84ffc89"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.093532 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-node-log" (OuterVolumeSpecName: "node-log") pod "415fdacb-85c1-4265-89b7-6771a84ffc89" (UID: "415fdacb-85c1-4265-89b7-6771a84ffc89"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.093575 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "415fdacb-85c1-4265-89b7-6771a84ffc89" (UID: "415fdacb-85c1-4265-89b7-6771a84ffc89"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.093595 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/415fdacb-85c1-4265-89b7-6771a84ffc89-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "415fdacb-85c1-4265-89b7-6771a84ffc89" (UID: "415fdacb-85c1-4265-89b7-6771a84ffc89"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.093628 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-log-socket" (OuterVolumeSpecName: "log-socket") pod "415fdacb-85c1-4265-89b7-6771a84ffc89" (UID: "415fdacb-85c1-4265-89b7-6771a84ffc89"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.093674 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "415fdacb-85c1-4265-89b7-6771a84ffc89" (UID: "415fdacb-85c1-4265-89b7-6771a84ffc89"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.093878 4766 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.093906 4766 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/415fdacb-85c1-4265-89b7-6771a84ffc89-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.093919 4766 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.093933 4766 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.093945 4766 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.093957 4766 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.093970 4766 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-slash\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.093981 4766 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.093991 4766 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.094003 4766 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.094015 4766 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.094029 4766 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/415fdacb-85c1-4265-89b7-6771a84ffc89-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.094044 4766 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.094058 4766 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-log-socket\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.094068 4766 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.094079 4766 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/415fdacb-85c1-4265-89b7-6771a84ffc89-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.094090 4766 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-node-log\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.101442 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-59tgp"] Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.101644 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovn-controller" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.101694 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovn-controller" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.101704 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovn-acl-logging" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.101710 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovn-acl-logging" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.101718 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="nbdb" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.101724 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="nbdb" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.101730 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d00d8e8-b83e-4585-9678-fb64e663fec3" containerName="pull" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.101738 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d00d8e8-b83e-4585-9678-fb64e663fec3" containerName="pull" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.101745 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d00d8e8-b83e-4585-9678-fb64e663fec3" containerName="util" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.101751 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d00d8e8-b83e-4585-9678-fb64e663fec3" containerName="util" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.101759 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d00d8e8-b83e-4585-9678-fb64e663fec3" containerName="extract" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.101764 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d00d8e8-b83e-4585-9678-fb64e663fec3" containerName="extract" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.101772 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="kubecfg-setup" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.101778 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="kubecfg-setup" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.101787 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovnkube-controller" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.101793 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovnkube-controller" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.101798 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovnkube-controller" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.101804 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovnkube-controller" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.101810 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="northd" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.101816 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="northd" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.101826 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovnkube-controller" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.101832 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovnkube-controller" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.101842 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovnkube-controller" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.101847 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovnkube-controller" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.101857 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="kube-rbac-proxy-node" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.101863 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="kube-rbac-proxy-node" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.101874 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="kube-rbac-proxy-ovn-metrics" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.101880 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="kube-rbac-proxy-ovn-metrics" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.101888 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="sbdb" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.101893 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="sbdb" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.101977 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="sbdb" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.101989 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d00d8e8-b83e-4585-9678-fb64e663fec3" containerName="extract" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.101998 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovnkube-controller" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.102004 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="kube-rbac-proxy-ovn-metrics" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.102012 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="kube-rbac-proxy-node" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.102020 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovn-controller" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.102026 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="nbdb" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.102033 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovnkube-controller" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.102040 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="northd" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.102049 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovnkube-controller" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.102055 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovn-acl-logging" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.102062 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovnkube-controller" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.102144 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovnkube-controller" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.102152 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovnkube-controller" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.102237 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerName="ovnkube-controller" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.108967 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/415fdacb-85c1-4265-89b7-6771a84ffc89-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "415fdacb-85c1-4265-89b7-6771a84ffc89" (UID: "415fdacb-85c1-4265-89b7-6771a84ffc89"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.112159 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-prqrp_234068f7-4097-4e6b-aac3-4bafe425c1f2/kube-multus/2.log" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.112935 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-prqrp_234068f7-4097-4e6b-aac3-4bafe425c1f2/kube-multus/1.log" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.112972 4766 generic.go:334] "Generic (PLEG): container finished" podID="234068f7-4097-4e6b-aac3-4bafe425c1f2" containerID="c0c13369c6a2c56514a4088ae25d890e67379d75c2709a3f3041b422947fbcbd" exitCode=2 Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.113027 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-prqrp" event={"ID":"234068f7-4097-4e6b-aac3-4bafe425c1f2","Type":"ContainerDied","Data":"c0c13369c6a2c56514a4088ae25d890e67379d75c2709a3f3041b422947fbcbd"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.113058 4766 scope.go:117] "RemoveContainer" containerID="58d961f26d62b958e729b2bd5129790335de3019e968595091099e5e416353d1" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.118812 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/415fdacb-85c1-4265-89b7-6771a84ffc89-kube-api-access-l696l" (OuterVolumeSpecName: "kube-api-access-l696l") pod "415fdacb-85c1-4265-89b7-6771a84ffc89" (UID: "415fdacb-85c1-4265-89b7-6771a84ffc89"). InnerVolumeSpecName "kube-api-access-l696l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.118956 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.120161 4766 scope.go:117] "RemoveContainer" containerID="c0c13369c6a2c56514a4088ae25d890e67379d75c2709a3f3041b422947fbcbd" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.120438 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-prqrp_openshift-multus(234068f7-4097-4e6b-aac3-4bafe425c1f2)\"" pod="openshift-multus/multus-prqrp" podUID="234068f7-4097-4e6b-aac3-4bafe425c1f2" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.148374 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fxnln_415fdacb-85c1-4265-89b7-6771a84ffc89/ovnkube-controller/3.log" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.154219 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fxnln_415fdacb-85c1-4265-89b7-6771a84ffc89/ovn-acl-logging/0.log" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.156019 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fxnln_415fdacb-85c1-4265-89b7-6771a84ffc89/ovn-controller/0.log" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161687 4766 generic.go:334] "Generic (PLEG): container finished" podID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerID="2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1" exitCode=0 Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161715 4766 generic.go:334] "Generic (PLEG): container finished" podID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerID="4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da" exitCode=0 Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161722 4766 generic.go:334] "Generic (PLEG): container finished" podID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerID="ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff" exitCode=0 Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161730 4766 generic.go:334] "Generic (PLEG): container finished" podID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerID="4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52" exitCode=0 Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161738 4766 generic.go:334] "Generic (PLEG): container finished" podID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerID="2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820" exitCode=0 Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161744 4766 generic.go:334] "Generic (PLEG): container finished" podID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerID="619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498" exitCode=0 Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161750 4766 generic.go:334] "Generic (PLEG): container finished" podID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerID="cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99" exitCode=143 Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161757 4766 generic.go:334] "Generic (PLEG): container finished" podID="415fdacb-85c1-4265-89b7-6771a84ffc89" containerID="abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b" exitCode=143 Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161781 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerDied","Data":"2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161803 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerDied","Data":"4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161813 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerDied","Data":"ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161823 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerDied","Data":"4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161832 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerDied","Data":"2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161842 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerDied","Data":"619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161851 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161860 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161865 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161870 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161875 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161880 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161884 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161889 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161895 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161899 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161906 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerDied","Data":"cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161914 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161923 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161929 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161935 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161941 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161947 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161953 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161959 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161964 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161970 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161976 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerDied","Data":"abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161984 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161990 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161995 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.161999 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.162004 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.162009 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.162014 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.162018 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.162023 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.162027 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.162033 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" event={"ID":"415fdacb-85c1-4265-89b7-6771a84ffc89","Type":"ContainerDied","Data":"cde41151cb667db048d6f2ca5f4878532b799d90003bd6ebba008b716686f3c2"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.162042 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.162048 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.162054 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.162060 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.162066 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.162072 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.162078 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.162084 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.162088 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.162094 4766 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7"} Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.162174 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fxnln" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.163148 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "415fdacb-85c1-4265-89b7-6771a84ffc89" (UID: "415fdacb-85c1-4265-89b7-6771a84ffc89"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.173889 4766 scope.go:117] "RemoveContainer" containerID="2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.194836 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-run-ovn\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.194871 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-host-kubelet\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.194889 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-host-cni-bin\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.194914 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-host-run-ovn-kubernetes\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.194935 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-ovnkube-script-lib\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.194951 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-systemd-units\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.194970 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-ovn-node-metrics-cert\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.194984 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-run-openvswitch\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.195028 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-log-socket\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.195043 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-ovnkube-config\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.195061 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-host-cni-netd\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.195074 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-env-overrides\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.195090 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.195106 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-host-slash\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.195129 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l9k4\" (UniqueName: \"kubernetes.io/projected/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-kube-api-access-9l9k4\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.195148 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-var-lib-openvswitch\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.195165 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-run-systemd\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.195179 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-node-log\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.195216 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-etc-openvswitch\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.195234 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-host-run-netns\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.195265 4766 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/415fdacb-85c1-4265-89b7-6771a84ffc89-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.195276 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l696l\" (UniqueName: \"kubernetes.io/projected/415fdacb-85c1-4265-89b7-6771a84ffc89-kube-api-access-l696l\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.195285 4766 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/415fdacb-85c1-4265-89b7-6771a84ffc89-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.220758 4766 scope.go:117] "RemoveContainer" containerID="3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.242470 4766 scope.go:117] "RemoveContainer" containerID="4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.261314 4766 scope.go:117] "RemoveContainer" containerID="ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.273281 4766 scope.go:117] "RemoveContainer" containerID="4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.284552 4766 scope.go:117] "RemoveContainer" containerID="2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.296519 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-var-lib-openvswitch\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.296557 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l9k4\" (UniqueName: \"kubernetes.io/projected/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-kube-api-access-9l9k4\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.296578 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-run-systemd\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.296593 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-node-log\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.296618 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-etc-openvswitch\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.296635 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-host-run-netns\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.296701 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-run-ovn\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.296715 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-host-kubelet\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.296731 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-host-cni-bin\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.296751 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-host-run-ovn-kubernetes\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.296772 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-ovnkube-script-lib\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.296788 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-systemd-units\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.296805 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-ovn-node-metrics-cert\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.296822 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-run-openvswitch\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.296840 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-log-socket\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.296856 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-ovnkube-config\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.296914 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-run-ovn\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.296946 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-run-systemd\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.296963 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-host-kubelet\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.297007 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-host-run-ovn-kubernetes\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.297020 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-var-lib-openvswitch\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.297050 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-host-cni-bin\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.297059 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-node-log\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.297071 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-etc-openvswitch\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.297128 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-host-cni-netd\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.297196 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-run-openvswitch\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.297216 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-systemd-units\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.297230 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-host-run-netns\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.297257 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-log-socket\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.297285 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-env-overrides\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.297442 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-host-cni-netd\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.297504 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-ovnkube-script-lib\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.297520 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.297547 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.297553 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-host-slash\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.297601 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-host-slash\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.297728 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-env-overrides\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.297799 4766 scope.go:117] "RemoveContainer" containerID="619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.298204 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-ovnkube-config\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.303119 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-ovn-node-metrics-cert\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.313851 4766 scope.go:117] "RemoveContainer" containerID="cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.326110 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l9k4\" (UniqueName: \"kubernetes.io/projected/e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5-kube-api-access-9l9k4\") pod \"ovnkube-node-59tgp\" (UID: \"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5\") " pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.326966 4766 scope.go:117] "RemoveContainer" containerID="abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.341869 4766 scope.go:117] "RemoveContainer" containerID="d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.353773 4766 scope.go:117] "RemoveContainer" containerID="2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.356280 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1\": container with ID starting with 2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1 not found: ID does not exist" containerID="2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.356320 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1"} err="failed to get container status \"2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1\": rpc error: code = NotFound desc = could not find container \"2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1\": container with ID starting with 2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.356349 4766 scope.go:117] "RemoveContainer" containerID="3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.358276 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d\": container with ID starting with 3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d not found: ID does not exist" containerID="3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.358311 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d"} err="failed to get container status \"3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d\": rpc error: code = NotFound desc = could not find container \"3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d\": container with ID starting with 3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.358330 4766 scope.go:117] "RemoveContainer" containerID="4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.358704 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\": container with ID starting with 4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da not found: ID does not exist" containerID="4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.358742 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da"} err="failed to get container status \"4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\": rpc error: code = NotFound desc = could not find container \"4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\": container with ID starting with 4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.358765 4766 scope.go:117] "RemoveContainer" containerID="ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.359026 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\": container with ID starting with ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff not found: ID does not exist" containerID="ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.359049 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff"} err="failed to get container status \"ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\": rpc error: code = NotFound desc = could not find container \"ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\": container with ID starting with ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.359061 4766 scope.go:117] "RemoveContainer" containerID="4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.359257 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\": container with ID starting with 4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52 not found: ID does not exist" containerID="4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.359283 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52"} err="failed to get container status \"4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\": rpc error: code = NotFound desc = could not find container \"4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\": container with ID starting with 4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.359297 4766 scope.go:117] "RemoveContainer" containerID="2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.359511 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\": container with ID starting with 2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820 not found: ID does not exist" containerID="2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.359529 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820"} err="failed to get container status \"2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\": rpc error: code = NotFound desc = could not find container \"2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\": container with ID starting with 2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.359540 4766 scope.go:117] "RemoveContainer" containerID="619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.359756 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\": container with ID starting with 619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498 not found: ID does not exist" containerID="619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.359776 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498"} err="failed to get container status \"619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\": rpc error: code = NotFound desc = could not find container \"619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\": container with ID starting with 619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.359792 4766 scope.go:117] "RemoveContainer" containerID="cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.359950 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\": container with ID starting with cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99 not found: ID does not exist" containerID="cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.359973 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99"} err="failed to get container status \"cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\": rpc error: code = NotFound desc = could not find container \"cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\": container with ID starting with cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.359986 4766 scope.go:117] "RemoveContainer" containerID="abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.360170 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\": container with ID starting with abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b not found: ID does not exist" containerID="abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.360192 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b"} err="failed to get container status \"abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\": rpc error: code = NotFound desc = could not find container \"abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\": container with ID starting with abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.360207 4766 scope.go:117] "RemoveContainer" containerID="d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7" Nov 26 00:33:49 crc kubenswrapper[4766]: E1126 00:33:49.360392 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\": container with ID starting with d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7 not found: ID does not exist" containerID="d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.360414 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7"} err="failed to get container status \"d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\": rpc error: code = NotFound desc = could not find container \"d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\": container with ID starting with d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.360427 4766 scope.go:117] "RemoveContainer" containerID="2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.360623 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1"} err="failed to get container status \"2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1\": rpc error: code = NotFound desc = could not find container \"2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1\": container with ID starting with 2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.360643 4766 scope.go:117] "RemoveContainer" containerID="3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.361071 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d"} err="failed to get container status \"3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d\": rpc error: code = NotFound desc = could not find container \"3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d\": container with ID starting with 3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.361095 4766 scope.go:117] "RemoveContainer" containerID="4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.361290 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da"} err="failed to get container status \"4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\": rpc error: code = NotFound desc = could not find container \"4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\": container with ID starting with 4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.361309 4766 scope.go:117] "RemoveContainer" containerID="ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.361452 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff"} err="failed to get container status \"ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\": rpc error: code = NotFound desc = could not find container \"ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\": container with ID starting with ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.361468 4766 scope.go:117] "RemoveContainer" containerID="4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.361664 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52"} err="failed to get container status \"4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\": rpc error: code = NotFound desc = could not find container \"4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\": container with ID starting with 4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.361684 4766 scope.go:117] "RemoveContainer" containerID="2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.361872 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820"} err="failed to get container status \"2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\": rpc error: code = NotFound desc = could not find container \"2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\": container with ID starting with 2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.361892 4766 scope.go:117] "RemoveContainer" containerID="619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.362089 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498"} err="failed to get container status \"619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\": rpc error: code = NotFound desc = could not find container \"619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\": container with ID starting with 619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.362108 4766 scope.go:117] "RemoveContainer" containerID="cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.362311 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99"} err="failed to get container status \"cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\": rpc error: code = NotFound desc = could not find container \"cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\": container with ID starting with cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.362332 4766 scope.go:117] "RemoveContainer" containerID="abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.362509 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b"} err="failed to get container status \"abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\": rpc error: code = NotFound desc = could not find container \"abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\": container with ID starting with abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.362527 4766 scope.go:117] "RemoveContainer" containerID="d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.362717 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7"} err="failed to get container status \"d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\": rpc error: code = NotFound desc = could not find container \"d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\": container with ID starting with d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.362742 4766 scope.go:117] "RemoveContainer" containerID="2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.362914 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1"} err="failed to get container status \"2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1\": rpc error: code = NotFound desc = could not find container \"2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1\": container with ID starting with 2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.362932 4766 scope.go:117] "RemoveContainer" containerID="3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.363149 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d"} err="failed to get container status \"3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d\": rpc error: code = NotFound desc = could not find container \"3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d\": container with ID starting with 3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.363168 4766 scope.go:117] "RemoveContainer" containerID="4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.363355 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da"} err="failed to get container status \"4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\": rpc error: code = NotFound desc = could not find container \"4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\": container with ID starting with 4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.363371 4766 scope.go:117] "RemoveContainer" containerID="ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.363559 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff"} err="failed to get container status \"ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\": rpc error: code = NotFound desc = could not find container \"ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\": container with ID starting with ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.363574 4766 scope.go:117] "RemoveContainer" containerID="4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.363819 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52"} err="failed to get container status \"4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\": rpc error: code = NotFound desc = could not find container \"4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\": container with ID starting with 4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.363846 4766 scope.go:117] "RemoveContainer" containerID="2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.364057 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820"} err="failed to get container status \"2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\": rpc error: code = NotFound desc = could not find container \"2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\": container with ID starting with 2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.364078 4766 scope.go:117] "RemoveContainer" containerID="619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.364270 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498"} err="failed to get container status \"619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\": rpc error: code = NotFound desc = could not find container \"619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\": container with ID starting with 619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.364287 4766 scope.go:117] "RemoveContainer" containerID="cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.364614 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99"} err="failed to get container status \"cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\": rpc error: code = NotFound desc = could not find container \"cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\": container with ID starting with cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.364636 4766 scope.go:117] "RemoveContainer" containerID="abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.364827 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b"} err="failed to get container status \"abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\": rpc error: code = NotFound desc = could not find container \"abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\": container with ID starting with abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.364849 4766 scope.go:117] "RemoveContainer" containerID="d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.365058 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7"} err="failed to get container status \"d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\": rpc error: code = NotFound desc = could not find container \"d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\": container with ID starting with d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.365075 4766 scope.go:117] "RemoveContainer" containerID="2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.365245 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1"} err="failed to get container status \"2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1\": rpc error: code = NotFound desc = could not find container \"2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1\": container with ID starting with 2b5261479540b0342fe6da8e3166d873a1fe9ef175377bd122de7ed24dbf80e1 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.365264 4766 scope.go:117] "RemoveContainer" containerID="3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.365417 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d"} err="failed to get container status \"3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d\": rpc error: code = NotFound desc = could not find container \"3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d\": container with ID starting with 3d0e550fd486488429ebe3ea587fafa7dceef16df6793e0852471e1bbd87651d not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.365434 4766 scope.go:117] "RemoveContainer" containerID="4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.365606 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da"} err="failed to get container status \"4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\": rpc error: code = NotFound desc = could not find container \"4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da\": container with ID starting with 4a37460c90d2855a15ff8cc561bc813f60c19ac47051e59b8eb741549fcd47da not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.365629 4766 scope.go:117] "RemoveContainer" containerID="ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.365930 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff"} err="failed to get container status \"ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\": rpc error: code = NotFound desc = could not find container \"ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff\": container with ID starting with ba19a952708edc327e1cb9d8df2c2ae937ef1dbe10f35e1ce8b764836373a4ff not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.365947 4766 scope.go:117] "RemoveContainer" containerID="4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.366132 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52"} err="failed to get container status \"4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\": rpc error: code = NotFound desc = could not find container \"4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52\": container with ID starting with 4e21540d459a053fde1857ba68d1953547a2f29043d94b3a7ebe6bd94f8fea52 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.366150 4766 scope.go:117] "RemoveContainer" containerID="2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.366335 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820"} err="failed to get container status \"2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\": rpc error: code = NotFound desc = could not find container \"2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820\": container with ID starting with 2757b072e687ee6304ba4363b4825bb2ffebbb24af48f4441915ef528e700820 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.366352 4766 scope.go:117] "RemoveContainer" containerID="619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.366889 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498"} err="failed to get container status \"619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\": rpc error: code = NotFound desc = could not find container \"619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498\": container with ID starting with 619d1c415204d431dfc45018a7142ec1612b10d63ca9132d45e851ce4b9e1498 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.366907 4766 scope.go:117] "RemoveContainer" containerID="cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.367072 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99"} err="failed to get container status \"cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\": rpc error: code = NotFound desc = could not find container \"cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99\": container with ID starting with cd2ade8aed0db5e9194eac623728115a19604592cceb0dc1b60701162918fe99 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.367090 4766 scope.go:117] "RemoveContainer" containerID="abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.367286 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b"} err="failed to get container status \"abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\": rpc error: code = NotFound desc = could not find container \"abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b\": container with ID starting with abe20c4c25b841eee23ef700d9187dda1d07c9a97ab076365551f78a476eba3b not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.367303 4766 scope.go:117] "RemoveContainer" containerID="d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.367506 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7"} err="failed to get container status \"d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\": rpc error: code = NotFound desc = could not find container \"d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7\": container with ID starting with d11c347729fc5610dd20e8fdacb7b1e9a66f7d14a75121d3e50458ac520d7ff7 not found: ID does not exist" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.447364 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.500480 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fxnln"] Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.514441 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fxnln"] Nov 26 00:33:49 crc kubenswrapper[4766]: I1126 00:33:49.833327 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="415fdacb-85c1-4265-89b7-6771a84ffc89" path="/var/lib/kubelet/pods/415fdacb-85c1-4265-89b7-6771a84ffc89/volumes" Nov 26 00:33:50 crc kubenswrapper[4766]: I1126 00:33:50.184163 4766 generic.go:334] "Generic (PLEG): container finished" podID="e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5" containerID="5b4305cb58f46bbe4a43f12ec2afd9c21f44849c63dcab24aae43efd23aade6c" exitCode=0 Nov 26 00:33:50 crc kubenswrapper[4766]: I1126 00:33:50.184219 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" event={"ID":"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5","Type":"ContainerDied","Data":"5b4305cb58f46bbe4a43f12ec2afd9c21f44849c63dcab24aae43efd23aade6c"} Nov 26 00:33:50 crc kubenswrapper[4766]: I1126 00:33:50.184242 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" event={"ID":"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5","Type":"ContainerStarted","Data":"27d1fce9331b823d52435f8a720b12fc3c22d45838345bac9226961971b18d6f"} Nov 26 00:33:50 crc kubenswrapper[4766]: I1126 00:33:50.187031 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-prqrp_234068f7-4097-4e6b-aac3-4bafe425c1f2/kube-multus/2.log" Nov 26 00:33:51 crc kubenswrapper[4766]: I1126 00:33:51.197061 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" event={"ID":"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5","Type":"ContainerStarted","Data":"66dd5d8d87f468dcfa6f1e5a16c9fc129ec86531f1ed3fb8d1a0f7f3bc1096ff"} Nov 26 00:33:51 crc kubenswrapper[4766]: I1126 00:33:51.197484 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" event={"ID":"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5","Type":"ContainerStarted","Data":"8a6cb393ab261a5abbc0b191db0f49de32a56d06c67062633a1bf04f2921b21d"} Nov 26 00:33:51 crc kubenswrapper[4766]: I1126 00:33:51.197495 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" event={"ID":"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5","Type":"ContainerStarted","Data":"df776608cd7ee9c2d58118a51c458aafe744d02be39103c71850145e17d2045c"} Nov 26 00:33:51 crc kubenswrapper[4766]: I1126 00:33:51.197503 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" event={"ID":"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5","Type":"ContainerStarted","Data":"49be04bf440e149f7452a143fb7208ad3c336f515a95c0d073e80df424a03a72"} Nov 26 00:33:51 crc kubenswrapper[4766]: I1126 00:33:51.197511 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" event={"ID":"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5","Type":"ContainerStarted","Data":"35ae13a45efac7c1c9fde5152bfe344cfd525cc4c9b05159afd4ace2ca822aa2"} Nov 26 00:33:51 crc kubenswrapper[4766]: I1126 00:33:51.197520 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" event={"ID":"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5","Type":"ContainerStarted","Data":"716c44614d935bf9102a54373949cbf7f5873894d0afe05f8b04636246379efa"} Nov 26 00:33:54 crc kubenswrapper[4766]: I1126 00:33:54.214308 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" event={"ID":"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5","Type":"ContainerStarted","Data":"edf56c0adb455ad0b49d96851bedbc32f150b985d79ad2da7351c1c309d95692"} Nov 26 00:33:54 crc kubenswrapper[4766]: I1126 00:33:54.841275 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977"] Nov 26 00:33:54 crc kubenswrapper[4766]: I1126 00:33:54.841983 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" Nov 26 00:33:54 crc kubenswrapper[4766]: I1126 00:33:54.844960 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Nov 26 00:33:54 crc kubenswrapper[4766]: I1126 00:33:54.846885 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-t2z57" Nov 26 00:33:54 crc kubenswrapper[4766]: I1126 00:33:54.846946 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Nov 26 00:33:54 crc kubenswrapper[4766]: I1126 00:33:54.960205 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt4wx\" (UniqueName: \"kubernetes.io/projected/bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70-kube-api-access-kt4wx\") pod \"obo-prometheus-operator-668cf9dfbb-bt977\" (UID: \"bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" Nov 26 00:33:54 crc kubenswrapper[4766]: I1126 00:33:54.978608 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx"] Nov 26 00:33:54 crc kubenswrapper[4766]: I1126 00:33:54.979229 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" Nov 26 00:33:54 crc kubenswrapper[4766]: I1126 00:33:54.981021 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-whslv" Nov 26 00:33:54 crc kubenswrapper[4766]: I1126 00:33:54.981182 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Nov 26 00:33:54 crc kubenswrapper[4766]: I1126 00:33:54.994907 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs"] Nov 26 00:33:54 crc kubenswrapper[4766]: I1126 00:33:54.995534 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.061534 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt4wx\" (UniqueName: \"kubernetes.io/projected/bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70-kube-api-access-kt4wx\") pod \"obo-prometheus-operator-668cf9dfbb-bt977\" (UID: \"bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.074625 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-b8qgs"] Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.076296 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.079130 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-7b2rn" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.083445 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.104403 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt4wx\" (UniqueName: \"kubernetes.io/projected/bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70-kube-api-access-kt4wx\") pod \"obo-prometheus-operator-668cf9dfbb-bt977\" (UID: \"bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.157115 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.162580 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1125f245-de34-4508-a4b0-2a468e73df36-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs\" (UID: \"1125f245-de34-4508-a4b0-2a468e73df36\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.162634 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1125f245-de34-4508-a4b0-2a468e73df36-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs\" (UID: \"1125f245-de34-4508-a4b0-2a468e73df36\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.162678 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/490105d8-e5f5-4fe9-89b9-d67ee8e07b87-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx\" (UID: \"490105d8-e5f5-4fe9-89b9-d67ee8e07b87\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.163173 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/490105d8-e5f5-4fe9-89b9-d67ee8e07b87-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx\" (UID: \"490105d8-e5f5-4fe9-89b9-d67ee8e07b87\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.172257 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-twhqz"] Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.173055 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.174751 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-84mld" Nov 26 00:33:55 crc kubenswrapper[4766]: E1126 00:33:55.179107 4766 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-bt977_openshift-operators_bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70_0(48b8bd074ded384bbb5a7f4f2866c5e7244792fc6244be76747c814161eef47c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 00:33:55 crc kubenswrapper[4766]: E1126 00:33:55.179170 4766 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-bt977_openshift-operators_bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70_0(48b8bd074ded384bbb5a7f4f2866c5e7244792fc6244be76747c814161eef47c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" Nov 26 00:33:55 crc kubenswrapper[4766]: E1126 00:33:55.179192 4766 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-bt977_openshift-operators_bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70_0(48b8bd074ded384bbb5a7f4f2866c5e7244792fc6244be76747c814161eef47c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" Nov 26 00:33:55 crc kubenswrapper[4766]: E1126 00:33:55.179236 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-bt977_openshift-operators(bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-bt977_openshift-operators(bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-bt977_openshift-operators_bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70_0(48b8bd074ded384bbb5a7f4f2866c5e7244792fc6244be76747c814161eef47c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" podUID="bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.263914 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/490105d8-e5f5-4fe9-89b9-d67ee8e07b87-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx\" (UID: \"490105d8-e5f5-4fe9-89b9-d67ee8e07b87\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.264013 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1125f245-de34-4508-a4b0-2a468e73df36-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs\" (UID: \"1125f245-de34-4508-a4b0-2a468e73df36\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.264206 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1125f245-de34-4508-a4b0-2a468e73df36-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs\" (UID: \"1125f245-de34-4508-a4b0-2a468e73df36\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.264250 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/490105d8-e5f5-4fe9-89b9-d67ee8e07b87-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx\" (UID: \"490105d8-e5f5-4fe9-89b9-d67ee8e07b87\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.264292 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr4bd\" (UniqueName: \"kubernetes.io/projected/a6dba801-0cee-4c4b-856c-db3be51c0ed9-kube-api-access-qr4bd\") pod \"observability-operator-d8bb48f5d-b8qgs\" (UID: \"a6dba801-0cee-4c4b-856c-db3be51c0ed9\") " pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.264324 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/a6dba801-0cee-4c4b-856c-db3be51c0ed9-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-b8qgs\" (UID: \"a6dba801-0cee-4c4b-856c-db3be51c0ed9\") " pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.266938 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1125f245-de34-4508-a4b0-2a468e73df36-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs\" (UID: \"1125f245-de34-4508-a4b0-2a468e73df36\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.267153 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/490105d8-e5f5-4fe9-89b9-d67ee8e07b87-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx\" (UID: \"490105d8-e5f5-4fe9-89b9-d67ee8e07b87\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.281075 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/490105d8-e5f5-4fe9-89b9-d67ee8e07b87-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx\" (UID: \"490105d8-e5f5-4fe9-89b9-d67ee8e07b87\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.282605 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1125f245-de34-4508-a4b0-2a468e73df36-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs\" (UID: \"1125f245-de34-4508-a4b0-2a468e73df36\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.296038 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.310288 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" Nov 26 00:33:55 crc kubenswrapper[4766]: E1126 00:33:55.333196 4766 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx_openshift-operators_490105d8-e5f5-4fe9-89b9-d67ee8e07b87_0(728db035271a51ca56b3014f6b7b663b1ef107d28501037008767afec2ea28e1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 00:33:55 crc kubenswrapper[4766]: E1126 00:33:55.333274 4766 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx_openshift-operators_490105d8-e5f5-4fe9-89b9-d67ee8e07b87_0(728db035271a51ca56b3014f6b7b663b1ef107d28501037008767afec2ea28e1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" Nov 26 00:33:55 crc kubenswrapper[4766]: E1126 00:33:55.333296 4766 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx_openshift-operators_490105d8-e5f5-4fe9-89b9-d67ee8e07b87_0(728db035271a51ca56b3014f6b7b663b1ef107d28501037008767afec2ea28e1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" Nov 26 00:33:55 crc kubenswrapper[4766]: E1126 00:33:55.333341 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx_openshift-operators(490105d8-e5f5-4fe9-89b9-d67ee8e07b87)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx_openshift-operators(490105d8-e5f5-4fe9-89b9-d67ee8e07b87)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx_openshift-operators_490105d8-e5f5-4fe9-89b9-d67ee8e07b87_0(728db035271a51ca56b3014f6b7b663b1ef107d28501037008767afec2ea28e1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" podUID="490105d8-e5f5-4fe9-89b9-d67ee8e07b87" Nov 26 00:33:55 crc kubenswrapper[4766]: E1126 00:33:55.344496 4766 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs_openshift-operators_1125f245-de34-4508-a4b0-2a468e73df36_0(5227cb3d83a6b0d8010a6b63d68c596cbbf4382f79e5df86307f4d5ed6aabdcb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 00:33:55 crc kubenswrapper[4766]: E1126 00:33:55.344602 4766 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs_openshift-operators_1125f245-de34-4508-a4b0-2a468e73df36_0(5227cb3d83a6b0d8010a6b63d68c596cbbf4382f79e5df86307f4d5ed6aabdcb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" Nov 26 00:33:55 crc kubenswrapper[4766]: E1126 00:33:55.344624 4766 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs_openshift-operators_1125f245-de34-4508-a4b0-2a468e73df36_0(5227cb3d83a6b0d8010a6b63d68c596cbbf4382f79e5df86307f4d5ed6aabdcb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" Nov 26 00:33:55 crc kubenswrapper[4766]: E1126 00:33:55.344702 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs_openshift-operators(1125f245-de34-4508-a4b0-2a468e73df36)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs_openshift-operators(1125f245-de34-4508-a4b0-2a468e73df36)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs_openshift-operators_1125f245-de34-4508-a4b0-2a468e73df36_0(5227cb3d83a6b0d8010a6b63d68c596cbbf4382f79e5df86307f4d5ed6aabdcb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" podUID="1125f245-de34-4508-a4b0-2a468e73df36" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.365477 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr4bd\" (UniqueName: \"kubernetes.io/projected/a6dba801-0cee-4c4b-856c-db3be51c0ed9-kube-api-access-qr4bd\") pod \"observability-operator-d8bb48f5d-b8qgs\" (UID: \"a6dba801-0cee-4c4b-856c-db3be51c0ed9\") " pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.365523 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/a6dba801-0cee-4c4b-856c-db3be51c0ed9-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-b8qgs\" (UID: \"a6dba801-0cee-4c4b-856c-db3be51c0ed9\") " pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.365549 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npmjt\" (UniqueName: \"kubernetes.io/projected/0be2b756-c941-49e3-b8f0-0977eaaf59d0-kube-api-access-npmjt\") pod \"perses-operator-5446b9c989-twhqz\" (UID: \"0be2b756-c941-49e3-b8f0-0977eaaf59d0\") " pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.365575 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/0be2b756-c941-49e3-b8f0-0977eaaf59d0-openshift-service-ca\") pod \"perses-operator-5446b9c989-twhqz\" (UID: \"0be2b756-c941-49e3-b8f0-0977eaaf59d0\") " pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.368185 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/a6dba801-0cee-4c4b-856c-db3be51c0ed9-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-b8qgs\" (UID: \"a6dba801-0cee-4c4b-856c-db3be51c0ed9\") " pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.380054 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr4bd\" (UniqueName: \"kubernetes.io/projected/a6dba801-0cee-4c4b-856c-db3be51c0ed9-kube-api-access-qr4bd\") pod \"observability-operator-d8bb48f5d-b8qgs\" (UID: \"a6dba801-0cee-4c4b-856c-db3be51c0ed9\") " pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.399865 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:33:55 crc kubenswrapper[4766]: E1126 00:33:55.418357 4766 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-b8qgs_openshift-operators_a6dba801-0cee-4c4b-856c-db3be51c0ed9_0(1e7245caeb919254fac735343f83f9ee34f8fb950dee83eee1972ad4f11297d8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 00:33:55 crc kubenswrapper[4766]: E1126 00:33:55.418411 4766 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-b8qgs_openshift-operators_a6dba801-0cee-4c4b-856c-db3be51c0ed9_0(1e7245caeb919254fac735343f83f9ee34f8fb950dee83eee1972ad4f11297d8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:33:55 crc kubenswrapper[4766]: E1126 00:33:55.418438 4766 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-b8qgs_openshift-operators_a6dba801-0cee-4c4b-856c-db3be51c0ed9_0(1e7245caeb919254fac735343f83f9ee34f8fb950dee83eee1972ad4f11297d8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:33:55 crc kubenswrapper[4766]: E1126 00:33:55.418479 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-b8qgs_openshift-operators(a6dba801-0cee-4c4b-856c-db3be51c0ed9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-b8qgs_openshift-operators(a6dba801-0cee-4c4b-856c-db3be51c0ed9)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-b8qgs_openshift-operators_a6dba801-0cee-4c4b-856c-db3be51c0ed9_0(1e7245caeb919254fac735343f83f9ee34f8fb950dee83eee1972ad4f11297d8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" podUID="a6dba801-0cee-4c4b-856c-db3be51c0ed9" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.466555 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npmjt\" (UniqueName: \"kubernetes.io/projected/0be2b756-c941-49e3-b8f0-0977eaaf59d0-kube-api-access-npmjt\") pod \"perses-operator-5446b9c989-twhqz\" (UID: \"0be2b756-c941-49e3-b8f0-0977eaaf59d0\") " pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.467021 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/0be2b756-c941-49e3-b8f0-0977eaaf59d0-openshift-service-ca\") pod \"perses-operator-5446b9c989-twhqz\" (UID: \"0be2b756-c941-49e3-b8f0-0977eaaf59d0\") " pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.467991 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/0be2b756-c941-49e3-b8f0-0977eaaf59d0-openshift-service-ca\") pod \"perses-operator-5446b9c989-twhqz\" (UID: \"0be2b756-c941-49e3-b8f0-0977eaaf59d0\") " pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.486037 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npmjt\" (UniqueName: \"kubernetes.io/projected/0be2b756-c941-49e3-b8f0-0977eaaf59d0-kube-api-access-npmjt\") pod \"perses-operator-5446b9c989-twhqz\" (UID: \"0be2b756-c941-49e3-b8f0-0977eaaf59d0\") " pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:33:55 crc kubenswrapper[4766]: I1126 00:33:55.517590 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:33:55 crc kubenswrapper[4766]: E1126 00:33:55.538088 4766 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-twhqz_openshift-operators_0be2b756-c941-49e3-b8f0-0977eaaf59d0_0(0daa61120250622d723963d2639c89fef0c67378b4f3a3e5ae25a72d577be88f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 00:33:55 crc kubenswrapper[4766]: E1126 00:33:55.538148 4766 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-twhqz_openshift-operators_0be2b756-c941-49e3-b8f0-0977eaaf59d0_0(0daa61120250622d723963d2639c89fef0c67378b4f3a3e5ae25a72d577be88f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:33:55 crc kubenswrapper[4766]: E1126 00:33:55.538174 4766 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-twhqz_openshift-operators_0be2b756-c941-49e3-b8f0-0977eaaf59d0_0(0daa61120250622d723963d2639c89fef0c67378b4f3a3e5ae25a72d577be88f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:33:55 crc kubenswrapper[4766]: E1126 00:33:55.538229 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-twhqz_openshift-operators(0be2b756-c941-49e3-b8f0-0977eaaf59d0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-twhqz_openshift-operators(0be2b756-c941-49e3-b8f0-0977eaaf59d0)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-twhqz_openshift-operators_0be2b756-c941-49e3-b8f0-0977eaaf59d0_0(0daa61120250622d723963d2639c89fef0c67378b4f3a3e5ae25a72d577be88f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-twhqz" podUID="0be2b756-c941-49e3-b8f0-0977eaaf59d0" Nov 26 00:33:56 crc kubenswrapper[4766]: I1126 00:33:56.226538 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" event={"ID":"e46e6fbb-6a86-4d7d-91d8-edf6b51f15d5","Type":"ContainerStarted","Data":"4a0da13076d131dc1f006ed4dd267f0065c6cc62073f8e50fbbc29497ba7ab04"} Nov 26 00:33:56 crc kubenswrapper[4766]: I1126 00:33:56.227116 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:56 crc kubenswrapper[4766]: I1126 00:33:56.227310 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:56 crc kubenswrapper[4766]: I1126 00:33:56.256468 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:56 crc kubenswrapper[4766]: I1126 00:33:56.260852 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" podStartSLOduration=7.260832079 podStartE2EDuration="7.260832079s" podCreationTimestamp="2025-11-26 00:33:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:33:56.256767325 +0000 UTC m=+617.105537755" watchObservedRunningTime="2025-11-26 00:33:56.260832079 +0000 UTC m=+617.109602509" Nov 26 00:33:56 crc kubenswrapper[4766]: I1126 00:33:56.552511 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977"] Nov 26 00:33:56 crc kubenswrapper[4766]: I1126 00:33:56.552784 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" Nov 26 00:33:56 crc kubenswrapper[4766]: I1126 00:33:56.553160 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" Nov 26 00:33:56 crc kubenswrapper[4766]: I1126 00:33:56.557229 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx"] Nov 26 00:33:56 crc kubenswrapper[4766]: I1126 00:33:56.557472 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" Nov 26 00:33:56 crc kubenswrapper[4766]: I1126 00:33:56.558045 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" Nov 26 00:33:56 crc kubenswrapper[4766]: I1126 00:33:56.583464 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-b8qgs"] Nov 26 00:33:56 crc kubenswrapper[4766]: I1126 00:33:56.583538 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:33:56 crc kubenswrapper[4766]: I1126 00:33:56.583801 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:33:56 crc kubenswrapper[4766]: E1126 00:33:56.590512 4766 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx_openshift-operators_490105d8-e5f5-4fe9-89b9-d67ee8e07b87_0(d31c56f48455a3afc7a146560ed260497ed59bf15d1d2e7e21aae2ab24342c03): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 00:33:56 crc kubenswrapper[4766]: E1126 00:33:56.590579 4766 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx_openshift-operators_490105d8-e5f5-4fe9-89b9-d67ee8e07b87_0(d31c56f48455a3afc7a146560ed260497ed59bf15d1d2e7e21aae2ab24342c03): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" Nov 26 00:33:56 crc kubenswrapper[4766]: E1126 00:33:56.590608 4766 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx_openshift-operators_490105d8-e5f5-4fe9-89b9-d67ee8e07b87_0(d31c56f48455a3afc7a146560ed260497ed59bf15d1d2e7e21aae2ab24342c03): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" Nov 26 00:33:56 crc kubenswrapper[4766]: E1126 00:33:56.590683 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx_openshift-operators(490105d8-e5f5-4fe9-89b9-d67ee8e07b87)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx_openshift-operators(490105d8-e5f5-4fe9-89b9-d67ee8e07b87)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx_openshift-operators_490105d8-e5f5-4fe9-89b9-d67ee8e07b87_0(d31c56f48455a3afc7a146560ed260497ed59bf15d1d2e7e21aae2ab24342c03): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" podUID="490105d8-e5f5-4fe9-89b9-d67ee8e07b87" Nov 26 00:33:56 crc kubenswrapper[4766]: E1126 00:33:56.601459 4766 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-bt977_openshift-operators_bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70_0(ae61f1dedb787c1fafa5796c6fa9112b41f979d9b62ac5690ea05a0c1b7a8f06): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 00:33:56 crc kubenswrapper[4766]: E1126 00:33:56.601521 4766 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-bt977_openshift-operators_bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70_0(ae61f1dedb787c1fafa5796c6fa9112b41f979d9b62ac5690ea05a0c1b7a8f06): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" Nov 26 00:33:56 crc kubenswrapper[4766]: E1126 00:33:56.601543 4766 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-bt977_openshift-operators_bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70_0(ae61f1dedb787c1fafa5796c6fa9112b41f979d9b62ac5690ea05a0c1b7a8f06): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" Nov 26 00:33:56 crc kubenswrapper[4766]: E1126 00:33:56.601584 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-bt977_openshift-operators(bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-bt977_openshift-operators(bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-bt977_openshift-operators_bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70_0(ae61f1dedb787c1fafa5796c6fa9112b41f979d9b62ac5690ea05a0c1b7a8f06): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" podUID="bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70" Nov 26 00:33:56 crc kubenswrapper[4766]: I1126 00:33:56.605812 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs"] Nov 26 00:33:56 crc kubenswrapper[4766]: I1126 00:33:56.605931 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" Nov 26 00:33:56 crc kubenswrapper[4766]: I1126 00:33:56.606395 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" Nov 26 00:33:56 crc kubenswrapper[4766]: I1126 00:33:56.614642 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-twhqz"] Nov 26 00:33:56 crc kubenswrapper[4766]: I1126 00:33:56.614759 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:33:56 crc kubenswrapper[4766]: I1126 00:33:56.615166 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:33:56 crc kubenswrapper[4766]: E1126 00:33:56.615530 4766 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-b8qgs_openshift-operators_a6dba801-0cee-4c4b-856c-db3be51c0ed9_0(820d876bdd88d627d41f547f78b1b50b221738b4e7193e486f4e2a6af3e5b684): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 00:33:56 crc kubenswrapper[4766]: E1126 00:33:56.615577 4766 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-b8qgs_openshift-operators_a6dba801-0cee-4c4b-856c-db3be51c0ed9_0(820d876bdd88d627d41f547f78b1b50b221738b4e7193e486f4e2a6af3e5b684): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:33:56 crc kubenswrapper[4766]: E1126 00:33:56.615602 4766 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-b8qgs_openshift-operators_a6dba801-0cee-4c4b-856c-db3be51c0ed9_0(820d876bdd88d627d41f547f78b1b50b221738b4e7193e486f4e2a6af3e5b684): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:33:56 crc kubenswrapper[4766]: E1126 00:33:56.615642 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-b8qgs_openshift-operators(a6dba801-0cee-4c4b-856c-db3be51c0ed9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-b8qgs_openshift-operators(a6dba801-0cee-4c4b-856c-db3be51c0ed9)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-b8qgs_openshift-operators_a6dba801-0cee-4c4b-856c-db3be51c0ed9_0(820d876bdd88d627d41f547f78b1b50b221738b4e7193e486f4e2a6af3e5b684): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" podUID="a6dba801-0cee-4c4b-856c-db3be51c0ed9" Nov 26 00:33:56 crc kubenswrapper[4766]: E1126 00:33:56.655605 4766 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs_openshift-operators_1125f245-de34-4508-a4b0-2a468e73df36_0(6a3669e55dbe39c4a52dfa5ebaa2725d6ced6a4ab1f386afdccd0db2de58ae91): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 00:33:56 crc kubenswrapper[4766]: E1126 00:33:56.655889 4766 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs_openshift-operators_1125f245-de34-4508-a4b0-2a468e73df36_0(6a3669e55dbe39c4a52dfa5ebaa2725d6ced6a4ab1f386afdccd0db2de58ae91): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" Nov 26 00:33:56 crc kubenswrapper[4766]: E1126 00:33:56.655919 4766 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs_openshift-operators_1125f245-de34-4508-a4b0-2a468e73df36_0(6a3669e55dbe39c4a52dfa5ebaa2725d6ced6a4ab1f386afdccd0db2de58ae91): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" Nov 26 00:33:56 crc kubenswrapper[4766]: E1126 00:33:56.655988 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs_openshift-operators(1125f245-de34-4508-a4b0-2a468e73df36)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs_openshift-operators(1125f245-de34-4508-a4b0-2a468e73df36)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs_openshift-operators_1125f245-de34-4508-a4b0-2a468e73df36_0(6a3669e55dbe39c4a52dfa5ebaa2725d6ced6a4ab1f386afdccd0db2de58ae91): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" podUID="1125f245-de34-4508-a4b0-2a468e73df36" Nov 26 00:33:56 crc kubenswrapper[4766]: E1126 00:33:56.662415 4766 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-twhqz_openshift-operators_0be2b756-c941-49e3-b8f0-0977eaaf59d0_0(a86a0aae1955f8aa3bcea658c6c4861fecc6b030288dbd4ed9474db8d3062914): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 00:33:56 crc kubenswrapper[4766]: E1126 00:33:56.662478 4766 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-twhqz_openshift-operators_0be2b756-c941-49e3-b8f0-0977eaaf59d0_0(a86a0aae1955f8aa3bcea658c6c4861fecc6b030288dbd4ed9474db8d3062914): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:33:56 crc kubenswrapper[4766]: E1126 00:33:56.662502 4766 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-twhqz_openshift-operators_0be2b756-c941-49e3-b8f0-0977eaaf59d0_0(a86a0aae1955f8aa3bcea658c6c4861fecc6b030288dbd4ed9474db8d3062914): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:33:56 crc kubenswrapper[4766]: E1126 00:33:56.662550 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-twhqz_openshift-operators(0be2b756-c941-49e3-b8f0-0977eaaf59d0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-twhqz_openshift-operators(0be2b756-c941-49e3-b8f0-0977eaaf59d0)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-twhqz_openshift-operators_0be2b756-c941-49e3-b8f0-0977eaaf59d0_0(a86a0aae1955f8aa3bcea658c6c4861fecc6b030288dbd4ed9474db8d3062914): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-twhqz" podUID="0be2b756-c941-49e3-b8f0-0977eaaf59d0" Nov 26 00:33:57 crc kubenswrapper[4766]: I1126 00:33:57.231683 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:33:57 crc kubenswrapper[4766]: I1126 00:33:57.260243 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:34:03 crc kubenswrapper[4766]: I1126 00:34:03.826596 4766 scope.go:117] "RemoveContainer" containerID="c0c13369c6a2c56514a4088ae25d890e67379d75c2709a3f3041b422947fbcbd" Nov 26 00:34:03 crc kubenswrapper[4766]: E1126 00:34:03.827365 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-prqrp_openshift-multus(234068f7-4097-4e6b-aac3-4bafe425c1f2)\"" pod="openshift-multus/multus-prqrp" podUID="234068f7-4097-4e6b-aac3-4bafe425c1f2" Nov 26 00:34:07 crc kubenswrapper[4766]: I1126 00:34:07.826237 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:34:07 crc kubenswrapper[4766]: I1126 00:34:07.828333 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:34:07 crc kubenswrapper[4766]: E1126 00:34:07.894183 4766 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-b8qgs_openshift-operators_a6dba801-0cee-4c4b-856c-db3be51c0ed9_0(fb73982b4075aceb04a4a1a2fa214b54dbfd3edd658cc10df39bcb505c4421e0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 00:34:07 crc kubenswrapper[4766]: E1126 00:34:07.894722 4766 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-b8qgs_openshift-operators_a6dba801-0cee-4c4b-856c-db3be51c0ed9_0(fb73982b4075aceb04a4a1a2fa214b54dbfd3edd658cc10df39bcb505c4421e0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:34:07 crc kubenswrapper[4766]: E1126 00:34:07.894819 4766 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-b8qgs_openshift-operators_a6dba801-0cee-4c4b-856c-db3be51c0ed9_0(fb73982b4075aceb04a4a1a2fa214b54dbfd3edd658cc10df39bcb505c4421e0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:34:07 crc kubenswrapper[4766]: E1126 00:34:07.894945 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-b8qgs_openshift-operators(a6dba801-0cee-4c4b-856c-db3be51c0ed9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-b8qgs_openshift-operators(a6dba801-0cee-4c4b-856c-db3be51c0ed9)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-b8qgs_openshift-operators_a6dba801-0cee-4c4b-856c-db3be51c0ed9_0(fb73982b4075aceb04a4a1a2fa214b54dbfd3edd658cc10df39bcb505c4421e0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" podUID="a6dba801-0cee-4c4b-856c-db3be51c0ed9" Nov 26 00:34:08 crc kubenswrapper[4766]: I1126 00:34:08.826920 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" Nov 26 00:34:08 crc kubenswrapper[4766]: I1126 00:34:08.827515 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" Nov 26 00:34:08 crc kubenswrapper[4766]: E1126 00:34:08.862783 4766 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-bt977_openshift-operators_bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70_0(80775c21a67fcad354ff26b69d2b20fff77ad143079648b5ff93d3ea4ee3cb00): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 00:34:08 crc kubenswrapper[4766]: E1126 00:34:08.862942 4766 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-bt977_openshift-operators_bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70_0(80775c21a67fcad354ff26b69d2b20fff77ad143079648b5ff93d3ea4ee3cb00): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" Nov 26 00:34:08 crc kubenswrapper[4766]: E1126 00:34:08.863049 4766 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-bt977_openshift-operators_bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70_0(80775c21a67fcad354ff26b69d2b20fff77ad143079648b5ff93d3ea4ee3cb00): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" Nov 26 00:34:08 crc kubenswrapper[4766]: E1126 00:34:08.863164 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-bt977_openshift-operators(bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-bt977_openshift-operators(bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-bt977_openshift-operators_bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70_0(80775c21a67fcad354ff26b69d2b20fff77ad143079648b5ff93d3ea4ee3cb00): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" podUID="bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70" Nov 26 00:34:10 crc kubenswrapper[4766]: I1126 00:34:10.826676 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:34:10 crc kubenswrapper[4766]: I1126 00:34:10.826729 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" Nov 26 00:34:10 crc kubenswrapper[4766]: I1126 00:34:10.828201 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" Nov 26 00:34:10 crc kubenswrapper[4766]: I1126 00:34:10.828419 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:34:10 crc kubenswrapper[4766]: E1126 00:34:10.858488 4766 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-twhqz_openshift-operators_0be2b756-c941-49e3-b8f0-0977eaaf59d0_0(610160780d939d3e407785960340b70f5d10b26c4d34f8f4ec566693754ab34a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 00:34:10 crc kubenswrapper[4766]: E1126 00:34:10.858703 4766 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-twhqz_openshift-operators_0be2b756-c941-49e3-b8f0-0977eaaf59d0_0(610160780d939d3e407785960340b70f5d10b26c4d34f8f4ec566693754ab34a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:34:10 crc kubenswrapper[4766]: E1126 00:34:10.858943 4766 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-twhqz_openshift-operators_0be2b756-c941-49e3-b8f0-0977eaaf59d0_0(610160780d939d3e407785960340b70f5d10b26c4d34f8f4ec566693754ab34a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:34:10 crc kubenswrapper[4766]: E1126 00:34:10.859099 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-twhqz_openshift-operators(0be2b756-c941-49e3-b8f0-0977eaaf59d0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-twhqz_openshift-operators(0be2b756-c941-49e3-b8f0-0977eaaf59d0)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-twhqz_openshift-operators_0be2b756-c941-49e3-b8f0-0977eaaf59d0_0(610160780d939d3e407785960340b70f5d10b26c4d34f8f4ec566693754ab34a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-twhqz" podUID="0be2b756-c941-49e3-b8f0-0977eaaf59d0" Nov 26 00:34:10 crc kubenswrapper[4766]: E1126 00:34:10.864520 4766 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx_openshift-operators_490105d8-e5f5-4fe9-89b9-d67ee8e07b87_0(ae0213e786706e7cd572a36866fb8efdc0f955b5eba18062221d1614fb3ad248): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 00:34:10 crc kubenswrapper[4766]: E1126 00:34:10.864568 4766 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx_openshift-operators_490105d8-e5f5-4fe9-89b9-d67ee8e07b87_0(ae0213e786706e7cd572a36866fb8efdc0f955b5eba18062221d1614fb3ad248): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" Nov 26 00:34:10 crc kubenswrapper[4766]: E1126 00:34:10.864587 4766 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx_openshift-operators_490105d8-e5f5-4fe9-89b9-d67ee8e07b87_0(ae0213e786706e7cd572a36866fb8efdc0f955b5eba18062221d1614fb3ad248): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" Nov 26 00:34:10 crc kubenswrapper[4766]: E1126 00:34:10.864630 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx_openshift-operators(490105d8-e5f5-4fe9-89b9-d67ee8e07b87)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx_openshift-operators(490105d8-e5f5-4fe9-89b9-d67ee8e07b87)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx_openshift-operators_490105d8-e5f5-4fe9-89b9-d67ee8e07b87_0(ae0213e786706e7cd572a36866fb8efdc0f955b5eba18062221d1614fb3ad248): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" podUID="490105d8-e5f5-4fe9-89b9-d67ee8e07b87" Nov 26 00:34:11 crc kubenswrapper[4766]: I1126 00:34:11.826714 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" Nov 26 00:34:11 crc kubenswrapper[4766]: I1126 00:34:11.827223 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" Nov 26 00:34:11 crc kubenswrapper[4766]: E1126 00:34:11.870558 4766 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs_openshift-operators_1125f245-de34-4508-a4b0-2a468e73df36_0(625766f9db9f621cb0add3ce3a9e422b7aeb9798ee11c399de852e4d447b9a34): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 00:34:11 crc kubenswrapper[4766]: E1126 00:34:11.870627 4766 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs_openshift-operators_1125f245-de34-4508-a4b0-2a468e73df36_0(625766f9db9f621cb0add3ce3a9e422b7aeb9798ee11c399de852e4d447b9a34): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" Nov 26 00:34:11 crc kubenswrapper[4766]: E1126 00:34:11.870663 4766 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs_openshift-operators_1125f245-de34-4508-a4b0-2a468e73df36_0(625766f9db9f621cb0add3ce3a9e422b7aeb9798ee11c399de852e4d447b9a34): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" Nov 26 00:34:11 crc kubenswrapper[4766]: E1126 00:34:11.870710 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs_openshift-operators(1125f245-de34-4508-a4b0-2a468e73df36)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs_openshift-operators(1125f245-de34-4508-a4b0-2a468e73df36)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs_openshift-operators_1125f245-de34-4508-a4b0-2a468e73df36_0(625766f9db9f621cb0add3ce3a9e422b7aeb9798ee11c399de852e4d447b9a34): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" podUID="1125f245-de34-4508-a4b0-2a468e73df36" Nov 26 00:34:16 crc kubenswrapper[4766]: I1126 00:34:16.826301 4766 scope.go:117] "RemoveContainer" containerID="c0c13369c6a2c56514a4088ae25d890e67379d75c2709a3f3041b422947fbcbd" Nov 26 00:34:17 crc kubenswrapper[4766]: I1126 00:34:17.330752 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-prqrp_234068f7-4097-4e6b-aac3-4bafe425c1f2/kube-multus/2.log" Nov 26 00:34:17 crc kubenswrapper[4766]: I1126 00:34:17.331355 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-prqrp" event={"ID":"234068f7-4097-4e6b-aac3-4bafe425c1f2","Type":"ContainerStarted","Data":"652dc8353c89ccc47c20752d1d208918fee3edb291252bf4f76d86e7e66796d2"} Nov 26 00:34:19 crc kubenswrapper[4766]: I1126 00:34:19.475534 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-59tgp" Nov 26 00:34:22 crc kubenswrapper[4766]: I1126 00:34:22.826545 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:34:22 crc kubenswrapper[4766]: I1126 00:34:22.826583 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" Nov 26 00:34:22 crc kubenswrapper[4766]: I1126 00:34:22.826756 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" Nov 26 00:34:22 crc kubenswrapper[4766]: I1126 00:34:22.826742 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:34:22 crc kubenswrapper[4766]: I1126 00:34:22.827049 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:34:22 crc kubenswrapper[4766]: I1126 00:34:22.827104 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" Nov 26 00:34:22 crc kubenswrapper[4766]: I1126 00:34:22.827736 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:34:22 crc kubenswrapper[4766]: I1126 00:34:22.829751 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" Nov 26 00:34:23 crc kubenswrapper[4766]: I1126 00:34:23.132398 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977"] Nov 26 00:34:23 crc kubenswrapper[4766]: W1126 00:34:23.142294 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf6b72c0_e2e6_4f3d_89f8_c18157e1ca70.slice/crio-5b54319dd53baad9c38c267a5f250329bca97033cd55fed117c7e407b9e6ec64 WatchSource:0}: Error finding container 5b54319dd53baad9c38c267a5f250329bca97033cd55fed117c7e407b9e6ec64: Status 404 returned error can't find the container with id 5b54319dd53baad9c38c267a5f250329bca97033cd55fed117c7e407b9e6ec64 Nov 26 00:34:23 crc kubenswrapper[4766]: I1126 00:34:23.168845 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx"] Nov 26 00:34:23 crc kubenswrapper[4766]: I1126 00:34:23.297590 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-twhqz"] Nov 26 00:34:23 crc kubenswrapper[4766]: I1126 00:34:23.302358 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-b8qgs"] Nov 26 00:34:23 crc kubenswrapper[4766]: W1126 00:34:23.306188 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6dba801_0cee_4c4b_856c_db3be51c0ed9.slice/crio-1e3dc9f0b785456c589c173cc802398075c00db41f48a07d07e88afcc43c3036 WatchSource:0}: Error finding container 1e3dc9f0b785456c589c173cc802398075c00db41f48a07d07e88afcc43c3036: Status 404 returned error can't find the container with id 1e3dc9f0b785456c589c173cc802398075c00db41f48a07d07e88afcc43c3036 Nov 26 00:34:23 crc kubenswrapper[4766]: I1126 00:34:23.363220 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" event={"ID":"a6dba801-0cee-4c4b-856c-db3be51c0ed9","Type":"ContainerStarted","Data":"1e3dc9f0b785456c589c173cc802398075c00db41f48a07d07e88afcc43c3036"} Nov 26 00:34:23 crc kubenswrapper[4766]: I1126 00:34:23.365622 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-twhqz" event={"ID":"0be2b756-c941-49e3-b8f0-0977eaaf59d0","Type":"ContainerStarted","Data":"6d9c050c3e5cb0e033dd48f3c9481a0e1e605473dfe696842f102b69a61691c1"} Nov 26 00:34:23 crc kubenswrapper[4766]: I1126 00:34:23.366893 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" event={"ID":"bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70","Type":"ContainerStarted","Data":"5b54319dd53baad9c38c267a5f250329bca97033cd55fed117c7e407b9e6ec64"} Nov 26 00:34:23 crc kubenswrapper[4766]: I1126 00:34:23.368057 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" event={"ID":"490105d8-e5f5-4fe9-89b9-d67ee8e07b87","Type":"ContainerStarted","Data":"b9eeb60335ac1a5a1e4ce7471f0e82e44520329adb594c4af903536599d575f4"} Nov 26 00:34:24 crc kubenswrapper[4766]: I1126 00:34:24.826558 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" Nov 26 00:34:24 crc kubenswrapper[4766]: I1126 00:34:24.827066 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" Nov 26 00:34:25 crc kubenswrapper[4766]: I1126 00:34:25.288238 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs"] Nov 26 00:34:25 crc kubenswrapper[4766]: W1126 00:34:25.306212 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1125f245_de34_4508_a4b0_2a468e73df36.slice/crio-ca81c34bf1c320a214931fd17bcc4acf399a5e753ea775abe1fb7fb00c041d49 WatchSource:0}: Error finding container ca81c34bf1c320a214931fd17bcc4acf399a5e753ea775abe1fb7fb00c041d49: Status 404 returned error can't find the container with id ca81c34bf1c320a214931fd17bcc4acf399a5e753ea775abe1fb7fb00c041d49 Nov 26 00:34:25 crc kubenswrapper[4766]: I1126 00:34:25.383089 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" event={"ID":"1125f245-de34-4508-a4b0-2a468e73df36","Type":"ContainerStarted","Data":"ca81c34bf1c320a214931fd17bcc4acf399a5e753ea775abe1fb7fb00c041d49"} Nov 26 00:34:38 crc kubenswrapper[4766]: E1126 00:34:38.480068 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec" Nov 26 00:34:38 crc kubenswrapper[4766]: E1126 00:34:38.480808 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator-admission-webhook,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec,Command:[],Args:[--web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{209715200 0} {} BinarySI},},Requests:ResourceList{cpu: {{50 -3} {} 50m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx_openshift-operators(490105d8-e5f5-4fe9-89b9-d67ee8e07b87): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 00:34:38 crc kubenswrapper[4766]: E1126 00:34:38.482013 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" podUID="490105d8-e5f5-4fe9-89b9-d67ee8e07b87" Nov 26 00:34:38 crc kubenswrapper[4766]: E1126 00:34:38.950256 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec" Nov 26 00:34:38 crc kubenswrapper[4766]: E1126 00:34:38.950482 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator-admission-webhook,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec,Command:[],Args:[--web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{209715200 0} {} BinarySI},},Requests:ResourceList{cpu: {{50 -3} {} 50m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs_openshift-operators(1125f245-de34-4508-a4b0-2a468e73df36): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 00:34:38 crc kubenswrapper[4766]: E1126 00:34:38.951517 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:203cf5b9dc1460f09e75f58d8b5cf7df5e57c18c8c6a41c14b5e8977d83263f3" Nov 26 00:34:38 crc kubenswrapper[4766]: E1126 00:34:38.951627 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" podUID="1125f245-de34-4508-a4b0-2a468e73df36" Nov 26 00:34:38 crc kubenswrapper[4766]: E1126 00:34:38.951724 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:203cf5b9dc1460f09e75f58d8b5cf7df5e57c18c8c6a41c14b5e8977d83263f3,Command:[],Args:[--prometheus-config-reloader=$(RELATED_IMAGE_PROMETHEUS_CONFIG_RELOADER) --prometheus-instance-selector=app.kubernetes.io/managed-by=observability-operator --alertmanager-instance-selector=app.kubernetes.io/managed-by=observability-operator --thanos-ruler-instance-selector=app.kubernetes.io/managed-by=observability-operator],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:GOGC,Value:30,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS_CONFIG_RELOADER,Value:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:1133c973c7472c665f910a722e19c8e2e27accb34b90fab67f14548627ce9c62,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{157286400 0} {} 150Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kt4wx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-668cf9dfbb-bt977_openshift-operators(bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 00:34:38 crc kubenswrapper[4766]: E1126 00:34:38.953728 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" podUID="bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70" Nov 26 00:34:39 crc kubenswrapper[4766]: E1126 00:34:39.431340 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385" Nov 26 00:34:39 crc kubenswrapper[4766]: E1126 00:34:39.431516 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:perses-operator,Image:registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{134217728 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openshift-service-ca,ReadOnly:true,MountPath:/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-npmjt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod perses-operator-5446b9c989-twhqz_openshift-operators(0be2b756-c941-49e3-b8f0-0977eaaf59d0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 00:34:39 crc kubenswrapper[4766]: E1126 00:34:39.432773 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/perses-operator-5446b9c989-twhqz" podUID="0be2b756-c941-49e3-b8f0-0977eaaf59d0" Nov 26 00:34:39 crc kubenswrapper[4766]: E1126 00:34:39.476996 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec\\\"\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" podUID="490105d8-e5f5-4fe9-89b9-d67ee8e07b87" Nov 26 00:34:39 crc kubenswrapper[4766]: E1126 00:34:39.477021 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385\\\"\"" pod="openshift-operators/perses-operator-5446b9c989-twhqz" podUID="0be2b756-c941-49e3-b8f0-0977eaaf59d0" Nov 26 00:34:39 crc kubenswrapper[4766]: E1126 00:34:39.477054 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec\\\"\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" podUID="1125f245-de34-4508-a4b0-2a468e73df36" Nov 26 00:34:39 crc kubenswrapper[4766]: E1126 00:34:39.477098 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:203cf5b9dc1460f09e75f58d8b5cf7df5e57c18c8c6a41c14b5e8977d83263f3\\\"\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" podUID="bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70" Nov 26 00:34:41 crc kubenswrapper[4766]: E1126 00:34:41.161829 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:ce7d2904f7b238aa37dfe74a0b76bf73629e7a14fa52bf54b0ecf030ca36f1bb" Nov 26 00:34:41 crc kubenswrapper[4766]: E1126 00:34:41.162387 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:ce7d2904f7b238aa37dfe74a0b76bf73629e7a14fa52bf54b0ecf030ca36f1bb,Command:[],Args:[--namespace=$(NAMESPACE) --images=perses=$(RELATED_IMAGE_PERSES) --images=alertmanager=$(RELATED_IMAGE_ALERTMANAGER) --images=prometheus=$(RELATED_IMAGE_PROMETHEUS) --images=thanos=$(RELATED_IMAGE_THANOS) --images=ui-dashboards=$(RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN) --images=ui-distributed-tracing=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN) --images=ui-distributed-tracing-pf5=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5) --images=ui-distributed-tracing-pf4=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4) --images=ui-logging=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN) --images=ui-logging-pf4=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4) --images=ui-troubleshooting-panel=$(RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN) --images=ui-monitoring=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN) --images=ui-monitoring-pf5=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5) --images=korrel8r=$(RELATED_IMAGE_KORREL8R) --images=health-analyzer=$(RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER) --openshift.enabled=true],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:RELATED_IMAGE_ALERTMANAGER,Value:registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:e718854a7d6ca8accf0fa72db0eb902e46c44d747ad51dc3f06bba0cefaa3c01,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS,Value:registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:17ea20be390a94ab39f5cdd7f0cbc2498046eebcf77fe3dec9aa288d5c2cf46b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_THANOS,Value:registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:d972f4faa5e9c121402d23ed85002f26af48ec36b1b71a7489d677b3913d08b4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PERSES,Value:registry.redhat.io/cluster-observability-operator/perses-rhel9@sha256:91531137fc1dcd740e277e0f65e120a0176a16f788c14c27925b61aa0b792ade,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-rhel9@sha256:a69da8bbca8a28dd2925f864d51cc31cf761b10532c553095ba40b242ef701cb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-rhel9@sha256:897e1bfad1187062725b54d87107bd0155972257a50d8335dd29e1999b828a4f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-pf5-rhel9@sha256:95fe5b5746ca8c07ac9217ce2d8ac8e6afad17af210f9d8e0074df1310b209a8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-pf4-rhel9@sha256:e9d9a89e4d8126a62b1852055482258ee528cac6398dd5d43ebad75ace0f33c9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-rhel9@sha256:ec684a0645ceb917b019af7ddba68c3533416e356ab0d0320a30e75ca7ebb31b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-pf4-rhel9@sha256:3b9693fcde9b3a9494fb04735b1f7cfd0426f10be820fdc3f024175c0d3df1c9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-rhel9@sha256:580606f194180accc8abba099e17a26dca7522ec6d233fa2fdd40312771703e3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-rhel9@sha256:e03777be39e71701935059cd877603874a13ac94daa73219d4e5e545599d78a9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-pf5-rhel9@sha256:aa47256193cfd2877853878e1ae97d2ab8b8e5deae62b387cbfad02b284d379c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KORREL8R,Value:registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:c595ff56b2cb85514bf4784db6ddb82e4e657e3e708a7fb695fc4997379a94d4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER,Value:registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:45a4ec2a519bcec99e886aa91596d5356a2414a2bd103baaef9fa7838c672eb2,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{400 -3} {} 400m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:observability-operator-tls,ReadOnly:true,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qr4bd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod observability-operator-d8bb48f5d-b8qgs_openshift-operators(a6dba801-0cee-4c4b-856c-db3be51c0ed9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 00:34:41 crc kubenswrapper[4766]: E1126 00:34:41.164114 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" podUID="a6dba801-0cee-4c4b-856c-db3be51c0ed9" Nov 26 00:34:41 crc kubenswrapper[4766]: E1126 00:34:41.479313 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:ce7d2904f7b238aa37dfe74a0b76bf73629e7a14fa52bf54b0ecf030ca36f1bb\\\"\"" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" podUID="a6dba801-0cee-4c4b-856c-db3be51c0ed9" Nov 26 00:34:52 crc kubenswrapper[4766]: I1126 00:34:52.541925 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" event={"ID":"1125f245-de34-4508-a4b0-2a468e73df36","Type":"ContainerStarted","Data":"2a3c7bcaf494d3d67fa25f24026a374d5c6472838c4ec47a838c1a68cf596456"} Nov 26 00:34:52 crc kubenswrapper[4766]: I1126 00:34:52.558434 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs" podStartSLOduration=32.53607693 podStartE2EDuration="58.558417893s" podCreationTimestamp="2025-11-26 00:33:54 +0000 UTC" firstStartedPulling="2025-11-26 00:34:25.31626024 +0000 UTC m=+646.165030670" lastFinishedPulling="2025-11-26 00:34:51.338601203 +0000 UTC m=+672.187371633" observedRunningTime="2025-11-26 00:34:52.556637118 +0000 UTC m=+673.405407558" watchObservedRunningTime="2025-11-26 00:34:52.558417893 +0000 UTC m=+673.407188323" Nov 26 00:34:53 crc kubenswrapper[4766]: I1126 00:34:53.548750 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" event={"ID":"490105d8-e5f5-4fe9-89b9-d67ee8e07b87","Type":"ContainerStarted","Data":"991d314a36e5e01c854e584739030f381b66a711c63d9e7d6da7d36dcac30406"} Nov 26 00:34:53 crc kubenswrapper[4766]: I1126 00:34:53.568716 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx" podStartSLOduration=-9223371977.286081 podStartE2EDuration="59.568694431s" podCreationTimestamp="2025-11-26 00:33:54 +0000 UTC" firstStartedPulling="2025-11-26 00:34:23.181377106 +0000 UTC m=+644.030147546" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:34:53.566585988 +0000 UTC m=+674.415356428" watchObservedRunningTime="2025-11-26 00:34:53.568694431 +0000 UTC m=+674.417464861" Nov 26 00:34:56 crc kubenswrapper[4766]: I1126 00:34:56.564012 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-twhqz" event={"ID":"0be2b756-c941-49e3-b8f0-0977eaaf59d0","Type":"ContainerStarted","Data":"19a07a737737cd2d34b3f96a342947de91267e4c3a8d81ad04c5a5673c6f211d"} Nov 26 00:34:56 crc kubenswrapper[4766]: I1126 00:34:56.564716 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:34:56 crc kubenswrapper[4766]: I1126 00:34:56.566617 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" event={"ID":"bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70","Type":"ContainerStarted","Data":"fc2b553ff9747cde68f8d26b81acc750d5d30b4415d57ed02bfc5d3035b5f97d"} Nov 26 00:34:56 crc kubenswrapper[4766]: I1126 00:34:56.567832 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" event={"ID":"a6dba801-0cee-4c4b-856c-db3be51c0ed9","Type":"ContainerStarted","Data":"d8a9591ba21970ab046000df4edd257c38db06a3e3537f7c1ed5d368ad569001"} Nov 26 00:34:56 crc kubenswrapper[4766]: I1126 00:34:56.568163 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:34:56 crc kubenswrapper[4766]: I1126 00:34:56.582574 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-twhqz" podStartSLOduration=29.207362482 podStartE2EDuration="1m1.582556864s" podCreationTimestamp="2025-11-26 00:33:55 +0000 UTC" firstStartedPulling="2025-11-26 00:34:23.310844748 +0000 UTC m=+644.159615178" lastFinishedPulling="2025-11-26 00:34:55.68603913 +0000 UTC m=+676.534809560" observedRunningTime="2025-11-26 00:34:56.58080401 +0000 UTC m=+677.429574450" watchObservedRunningTime="2025-11-26 00:34:56.582556864 +0000 UTC m=+677.431327294" Nov 26 00:34:56 crc kubenswrapper[4766]: I1126 00:34:56.607296 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" Nov 26 00:34:56 crc kubenswrapper[4766]: I1126 00:34:56.609023 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-b8qgs" podStartSLOduration=29.234074541 podStartE2EDuration="1m1.609005567s" podCreationTimestamp="2025-11-26 00:33:55 +0000 UTC" firstStartedPulling="2025-11-26 00:34:23.310800307 +0000 UTC m=+644.159570737" lastFinishedPulling="2025-11-26 00:34:55.685731303 +0000 UTC m=+676.534501763" observedRunningTime="2025-11-26 00:34:56.607446698 +0000 UTC m=+677.456217118" watchObservedRunningTime="2025-11-26 00:34:56.609005567 +0000 UTC m=+677.457775997" Nov 26 00:34:56 crc kubenswrapper[4766]: I1126 00:34:56.626589 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-bt977" podStartSLOduration=30.092325847 podStartE2EDuration="1m2.626572274s" podCreationTimestamp="2025-11-26 00:33:54 +0000 UTC" firstStartedPulling="2025-11-26 00:34:23.150551752 +0000 UTC m=+643.999322182" lastFinishedPulling="2025-11-26 00:34:55.684798179 +0000 UTC m=+676.533568609" observedRunningTime="2025-11-26 00:34:56.623984268 +0000 UTC m=+677.472754708" watchObservedRunningTime="2025-11-26 00:34:56.626572274 +0000 UTC m=+677.475342704" Nov 26 00:35:05 crc kubenswrapper[4766]: I1126 00:35:05.521201 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-twhqz" Nov 26 00:35:05 crc kubenswrapper[4766]: I1126 00:35:05.863052 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qjmjr"] Nov 26 00:35:05 crc kubenswrapper[4766]: I1126 00:35:05.863877 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-qjmjr" Nov 26 00:35:05 crc kubenswrapper[4766]: I1126 00:35:05.866338 4766 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-wt42m" Nov 26 00:35:05 crc kubenswrapper[4766]: I1126 00:35:05.866576 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 26 00:35:05 crc kubenswrapper[4766]: I1126 00:35:05.866909 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 26 00:35:05 crc kubenswrapper[4766]: I1126 00:35:05.875252 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qjmjr"] Nov 26 00:35:05 crc kubenswrapper[4766]: I1126 00:35:05.885706 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-nd9m2"] Nov 26 00:35:05 crc kubenswrapper[4766]: I1126 00:35:05.887197 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-nd9m2" Nov 26 00:35:05 crc kubenswrapper[4766]: I1126 00:35:05.892304 4766 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-v4fsg" Nov 26 00:35:05 crc kubenswrapper[4766]: I1126 00:35:05.903050 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-nd9m2"] Nov 26 00:35:05 crc kubenswrapper[4766]: I1126 00:35:05.906517 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrd4d\" (UniqueName: \"kubernetes.io/projected/954f9d95-f1ab-4e36-9b7a-d16fbe282e53-kube-api-access-vrd4d\") pod \"cert-manager-5b446d88c5-nd9m2\" (UID: \"954f9d95-f1ab-4e36-9b7a-d16fbe282e53\") " pod="cert-manager/cert-manager-5b446d88c5-nd9m2" Nov 26 00:35:05 crc kubenswrapper[4766]: I1126 00:35:05.906586 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfb6n\" (UniqueName: \"kubernetes.io/projected/e6ee9725-56d1-426a-afeb-c00b1abb9e2a-kube-api-access-sfb6n\") pod \"cert-manager-cainjector-7f985d654d-qjmjr\" (UID: \"e6ee9725-56d1-426a-afeb-c00b1abb9e2a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qjmjr" Nov 26 00:35:05 crc kubenswrapper[4766]: I1126 00:35:05.908029 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-btczr"] Nov 26 00:35:05 crc kubenswrapper[4766]: I1126 00:35:05.910014 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-btczr" Nov 26 00:35:05 crc kubenswrapper[4766]: I1126 00:35:05.912976 4766 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-hxpmg" Nov 26 00:35:05 crc kubenswrapper[4766]: I1126 00:35:05.932490 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-btczr"] Nov 26 00:35:06 crc kubenswrapper[4766]: I1126 00:35:06.007600 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrd4d\" (UniqueName: \"kubernetes.io/projected/954f9d95-f1ab-4e36-9b7a-d16fbe282e53-kube-api-access-vrd4d\") pod \"cert-manager-5b446d88c5-nd9m2\" (UID: \"954f9d95-f1ab-4e36-9b7a-d16fbe282e53\") " pod="cert-manager/cert-manager-5b446d88c5-nd9m2" Nov 26 00:35:06 crc kubenswrapper[4766]: I1126 00:35:06.007793 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfb6n\" (UniqueName: \"kubernetes.io/projected/e6ee9725-56d1-426a-afeb-c00b1abb9e2a-kube-api-access-sfb6n\") pod \"cert-manager-cainjector-7f985d654d-qjmjr\" (UID: \"e6ee9725-56d1-426a-afeb-c00b1abb9e2a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qjmjr" Nov 26 00:35:06 crc kubenswrapper[4766]: I1126 00:35:06.026949 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrd4d\" (UniqueName: \"kubernetes.io/projected/954f9d95-f1ab-4e36-9b7a-d16fbe282e53-kube-api-access-vrd4d\") pod \"cert-manager-5b446d88c5-nd9m2\" (UID: \"954f9d95-f1ab-4e36-9b7a-d16fbe282e53\") " pod="cert-manager/cert-manager-5b446d88c5-nd9m2" Nov 26 00:35:06 crc kubenswrapper[4766]: I1126 00:35:06.027556 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfb6n\" (UniqueName: \"kubernetes.io/projected/e6ee9725-56d1-426a-afeb-c00b1abb9e2a-kube-api-access-sfb6n\") pod \"cert-manager-cainjector-7f985d654d-qjmjr\" (UID: \"e6ee9725-56d1-426a-afeb-c00b1abb9e2a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qjmjr" Nov 26 00:35:06 crc kubenswrapper[4766]: I1126 00:35:06.108937 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5grhd\" (UniqueName: \"kubernetes.io/projected/403aadd6-8bd1-4446-b0e7-f342f31e54c9-kube-api-access-5grhd\") pod \"cert-manager-webhook-5655c58dd6-btczr\" (UID: \"403aadd6-8bd1-4446-b0e7-f342f31e54c9\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-btczr" Nov 26 00:35:06 crc kubenswrapper[4766]: I1126 00:35:06.180812 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-qjmjr" Nov 26 00:35:06 crc kubenswrapper[4766]: I1126 00:35:06.209805 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-nd9m2" Nov 26 00:35:06 crc kubenswrapper[4766]: I1126 00:35:06.210160 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5grhd\" (UniqueName: \"kubernetes.io/projected/403aadd6-8bd1-4446-b0e7-f342f31e54c9-kube-api-access-5grhd\") pod \"cert-manager-webhook-5655c58dd6-btczr\" (UID: \"403aadd6-8bd1-4446-b0e7-f342f31e54c9\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-btczr" Nov 26 00:35:06 crc kubenswrapper[4766]: I1126 00:35:06.236766 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5grhd\" (UniqueName: \"kubernetes.io/projected/403aadd6-8bd1-4446-b0e7-f342f31e54c9-kube-api-access-5grhd\") pod \"cert-manager-webhook-5655c58dd6-btczr\" (UID: \"403aadd6-8bd1-4446-b0e7-f342f31e54c9\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-btczr" Nov 26 00:35:06 crc kubenswrapper[4766]: I1126 00:35:06.429131 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qjmjr"] Nov 26 00:35:06 crc kubenswrapper[4766]: W1126 00:35:06.430046 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6ee9725_56d1_426a_afeb_c00b1abb9e2a.slice/crio-89a53bb2b84d7cad5ea65037744e6a716f32e2622fd6fc02807e4749f5f16857 WatchSource:0}: Error finding container 89a53bb2b84d7cad5ea65037744e6a716f32e2622fd6fc02807e4749f5f16857: Status 404 returned error can't find the container with id 89a53bb2b84d7cad5ea65037744e6a716f32e2622fd6fc02807e4749f5f16857 Nov 26 00:35:06 crc kubenswrapper[4766]: I1126 00:35:06.452055 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-nd9m2"] Nov 26 00:35:06 crc kubenswrapper[4766]: I1126 00:35:06.522028 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-btczr" Nov 26 00:35:06 crc kubenswrapper[4766]: I1126 00:35:06.627183 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-qjmjr" event={"ID":"e6ee9725-56d1-426a-afeb-c00b1abb9e2a","Type":"ContainerStarted","Data":"89a53bb2b84d7cad5ea65037744e6a716f32e2622fd6fc02807e4749f5f16857"} Nov 26 00:35:06 crc kubenswrapper[4766]: I1126 00:35:06.645034 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-nd9m2" event={"ID":"954f9d95-f1ab-4e36-9b7a-d16fbe282e53","Type":"ContainerStarted","Data":"7af3bff062387196de6b86af176713e779abc61a92e2465300c9af5de5efe677"} Nov 26 00:35:06 crc kubenswrapper[4766]: I1126 00:35:06.828108 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-btczr"] Nov 26 00:35:06 crc kubenswrapper[4766]: W1126 00:35:06.831341 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod403aadd6_8bd1_4446_b0e7_f342f31e54c9.slice/crio-d58f73e0582526baa8f8096edd4afc653c9928104fc6f724bb90772357fefc31 WatchSource:0}: Error finding container d58f73e0582526baa8f8096edd4afc653c9928104fc6f724bb90772357fefc31: Status 404 returned error can't find the container with id d58f73e0582526baa8f8096edd4afc653c9928104fc6f724bb90772357fefc31 Nov 26 00:35:07 crc kubenswrapper[4766]: I1126 00:35:07.652880 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-btczr" event={"ID":"403aadd6-8bd1-4446-b0e7-f342f31e54c9","Type":"ContainerStarted","Data":"d58f73e0582526baa8f8096edd4afc653c9928104fc6f724bb90772357fefc31"} Nov 26 00:35:24 crc kubenswrapper[4766]: I1126 00:35:24.765568 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-btczr" event={"ID":"403aadd6-8bd1-4446-b0e7-f342f31e54c9","Type":"ContainerStarted","Data":"e5445f77485b969c92fee36e4e6420413d80089d3b9e9aebe41d8862aaa7b983"} Nov 26 00:35:24 crc kubenswrapper[4766]: I1126 00:35:24.766097 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-btczr" Nov 26 00:35:24 crc kubenswrapper[4766]: I1126 00:35:24.767435 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-nd9m2" event={"ID":"954f9d95-f1ab-4e36-9b7a-d16fbe282e53","Type":"ContainerStarted","Data":"cace71eb37005730736ea0818605949369df8b7df51bfed5483cf0e70336e9ce"} Nov 26 00:35:24 crc kubenswrapper[4766]: I1126 00:35:24.786877 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-btczr" podStartSLOduration=2.073905625 podStartE2EDuration="19.786524315s" podCreationTimestamp="2025-11-26 00:35:05 +0000 UTC" firstStartedPulling="2025-11-26 00:35:06.833877887 +0000 UTC m=+687.682648317" lastFinishedPulling="2025-11-26 00:35:24.546496577 +0000 UTC m=+705.395267007" observedRunningTime="2025-11-26 00:35:24.778577112 +0000 UTC m=+705.627347542" watchObservedRunningTime="2025-11-26 00:35:24.786524315 +0000 UTC m=+705.635294745" Nov 26 00:35:24 crc kubenswrapper[4766]: I1126 00:35:24.802007 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-nd9m2" podStartSLOduration=1.712126949 podStartE2EDuration="19.801987958s" podCreationTimestamp="2025-11-26 00:35:05 +0000 UTC" firstStartedPulling="2025-11-26 00:35:06.46304236 +0000 UTC m=+687.311812800" lastFinishedPulling="2025-11-26 00:35:24.552903379 +0000 UTC m=+705.401673809" observedRunningTime="2025-11-26 00:35:24.797866623 +0000 UTC m=+705.646637053" watchObservedRunningTime="2025-11-26 00:35:24.801987958 +0000 UTC m=+705.650758388" Nov 26 00:35:25 crc kubenswrapper[4766]: I1126 00:35:25.775197 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-qjmjr" event={"ID":"e6ee9725-56d1-426a-afeb-c00b1abb9e2a","Type":"ContainerStarted","Data":"ceb1128e55c9097f6588cfcbabbdea9f4c403f0b45eb3918905bd6f8328ffb40"} Nov 26 00:35:25 crc kubenswrapper[4766]: I1126 00:35:25.789719 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-qjmjr" podStartSLOduration=2.420482165 podStartE2EDuration="20.789693892s" podCreationTimestamp="2025-11-26 00:35:05 +0000 UTC" firstStartedPulling="2025-11-26 00:35:06.431926609 +0000 UTC m=+687.280697039" lastFinishedPulling="2025-11-26 00:35:24.801138336 +0000 UTC m=+705.649908766" observedRunningTime="2025-11-26 00:35:25.788930873 +0000 UTC m=+706.637701323" watchObservedRunningTime="2025-11-26 00:35:25.789693892 +0000 UTC m=+706.638464342" Nov 26 00:35:31 crc kubenswrapper[4766]: I1126 00:35:31.525629 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-btczr" Nov 26 00:35:41 crc kubenswrapper[4766]: I1126 00:35:41.479608 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:35:41 crc kubenswrapper[4766]: I1126 00:35:41.480194 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:36:02 crc kubenswrapper[4766]: I1126 00:36:02.669496 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9"] Nov 26 00:36:02 crc kubenswrapper[4766]: I1126 00:36:02.672003 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9" Nov 26 00:36:02 crc kubenswrapper[4766]: I1126 00:36:02.675072 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 26 00:36:02 crc kubenswrapper[4766]: I1126 00:36:02.679062 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9"] Nov 26 00:36:02 crc kubenswrapper[4766]: I1126 00:36:02.824403 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/36c1a0d9-7873-4ce0-b926-17e535470807-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9\" (UID: \"36c1a0d9-7873-4ce0-b926-17e535470807\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9" Nov 26 00:36:02 crc kubenswrapper[4766]: I1126 00:36:02.824485 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/36c1a0d9-7873-4ce0-b926-17e535470807-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9\" (UID: \"36c1a0d9-7873-4ce0-b926-17e535470807\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9" Nov 26 00:36:02 crc kubenswrapper[4766]: I1126 00:36:02.824553 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clk2f\" (UniqueName: \"kubernetes.io/projected/36c1a0d9-7873-4ce0-b926-17e535470807-kube-api-access-clk2f\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9\" (UID: \"36c1a0d9-7873-4ce0-b926-17e535470807\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9" Nov 26 00:36:02 crc kubenswrapper[4766]: I1126 00:36:02.925778 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clk2f\" (UniqueName: \"kubernetes.io/projected/36c1a0d9-7873-4ce0-b926-17e535470807-kube-api-access-clk2f\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9\" (UID: \"36c1a0d9-7873-4ce0-b926-17e535470807\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9" Nov 26 00:36:02 crc kubenswrapper[4766]: I1126 00:36:02.925886 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/36c1a0d9-7873-4ce0-b926-17e535470807-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9\" (UID: \"36c1a0d9-7873-4ce0-b926-17e535470807\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9" Nov 26 00:36:02 crc kubenswrapper[4766]: I1126 00:36:02.925941 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/36c1a0d9-7873-4ce0-b926-17e535470807-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9\" (UID: \"36c1a0d9-7873-4ce0-b926-17e535470807\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9" Nov 26 00:36:02 crc kubenswrapper[4766]: I1126 00:36:02.926451 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/36c1a0d9-7873-4ce0-b926-17e535470807-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9\" (UID: \"36c1a0d9-7873-4ce0-b926-17e535470807\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9" Nov 26 00:36:02 crc kubenswrapper[4766]: I1126 00:36:02.926506 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/36c1a0d9-7873-4ce0-b926-17e535470807-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9\" (UID: \"36c1a0d9-7873-4ce0-b926-17e535470807\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9" Nov 26 00:36:02 crc kubenswrapper[4766]: I1126 00:36:02.944465 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clk2f\" (UniqueName: \"kubernetes.io/projected/36c1a0d9-7873-4ce0-b926-17e535470807-kube-api-access-clk2f\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9\" (UID: \"36c1a0d9-7873-4ce0-b926-17e535470807\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9" Nov 26 00:36:02 crc kubenswrapper[4766]: I1126 00:36:02.988690 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9" Nov 26 00:36:03 crc kubenswrapper[4766]: I1126 00:36:03.088352 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58"] Nov 26 00:36:03 crc kubenswrapper[4766]: I1126 00:36:03.090128 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58" Nov 26 00:36:03 crc kubenswrapper[4766]: I1126 00:36:03.092093 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58"] Nov 26 00:36:03 crc kubenswrapper[4766]: I1126 00:36:03.138395 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64a27a6e-0ec0-42e2-8606-370bafc71c04-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58\" (UID: \"64a27a6e-0ec0-42e2-8606-370bafc71c04\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58" Nov 26 00:36:03 crc kubenswrapper[4766]: I1126 00:36:03.138444 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdgzg\" (UniqueName: \"kubernetes.io/projected/64a27a6e-0ec0-42e2-8606-370bafc71c04-kube-api-access-hdgzg\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58\" (UID: \"64a27a6e-0ec0-42e2-8606-370bafc71c04\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58" Nov 26 00:36:03 crc kubenswrapper[4766]: I1126 00:36:03.138487 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64a27a6e-0ec0-42e2-8606-370bafc71c04-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58\" (UID: \"64a27a6e-0ec0-42e2-8606-370bafc71c04\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58" Nov 26 00:36:03 crc kubenswrapper[4766]: I1126 00:36:03.239346 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64a27a6e-0ec0-42e2-8606-370bafc71c04-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58\" (UID: \"64a27a6e-0ec0-42e2-8606-370bafc71c04\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58" Nov 26 00:36:03 crc kubenswrapper[4766]: I1126 00:36:03.239435 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64a27a6e-0ec0-42e2-8606-370bafc71c04-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58\" (UID: \"64a27a6e-0ec0-42e2-8606-370bafc71c04\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58" Nov 26 00:36:03 crc kubenswrapper[4766]: I1126 00:36:03.239474 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdgzg\" (UniqueName: \"kubernetes.io/projected/64a27a6e-0ec0-42e2-8606-370bafc71c04-kube-api-access-hdgzg\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58\" (UID: \"64a27a6e-0ec0-42e2-8606-370bafc71c04\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58" Nov 26 00:36:03 crc kubenswrapper[4766]: I1126 00:36:03.240673 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64a27a6e-0ec0-42e2-8606-370bafc71c04-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58\" (UID: \"64a27a6e-0ec0-42e2-8606-370bafc71c04\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58" Nov 26 00:36:03 crc kubenswrapper[4766]: I1126 00:36:03.241459 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64a27a6e-0ec0-42e2-8606-370bafc71c04-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58\" (UID: \"64a27a6e-0ec0-42e2-8606-370bafc71c04\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58" Nov 26 00:36:03 crc kubenswrapper[4766]: I1126 00:36:03.259179 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdgzg\" (UniqueName: \"kubernetes.io/projected/64a27a6e-0ec0-42e2-8606-370bafc71c04-kube-api-access-hdgzg\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58\" (UID: \"64a27a6e-0ec0-42e2-8606-370bafc71c04\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58" Nov 26 00:36:03 crc kubenswrapper[4766]: I1126 00:36:03.392381 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9"] Nov 26 00:36:03 crc kubenswrapper[4766]: I1126 00:36:03.414524 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58" Nov 26 00:36:03 crc kubenswrapper[4766]: I1126 00:36:03.611576 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58"] Nov 26 00:36:03 crc kubenswrapper[4766]: W1126 00:36:03.615950 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64a27a6e_0ec0_42e2_8606_370bafc71c04.slice/crio-c37895135af12bb57f111c200fd8f9b8b88828f25e716012335aeffcaea0893d WatchSource:0}: Error finding container c37895135af12bb57f111c200fd8f9b8b88828f25e716012335aeffcaea0893d: Status 404 returned error can't find the container with id c37895135af12bb57f111c200fd8f9b8b88828f25e716012335aeffcaea0893d Nov 26 00:36:04 crc kubenswrapper[4766]: I1126 00:36:04.001091 4766 generic.go:334] "Generic (PLEG): container finished" podID="36c1a0d9-7873-4ce0-b926-17e535470807" containerID="7f8351d81360e4ff2473e14c8b7c4a917beaafbf32531f9513555184d819be3a" exitCode=0 Nov 26 00:36:04 crc kubenswrapper[4766]: I1126 00:36:04.001194 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9" event={"ID":"36c1a0d9-7873-4ce0-b926-17e535470807","Type":"ContainerDied","Data":"7f8351d81360e4ff2473e14c8b7c4a917beaafbf32531f9513555184d819be3a"} Nov 26 00:36:04 crc kubenswrapper[4766]: I1126 00:36:04.001223 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9" event={"ID":"36c1a0d9-7873-4ce0-b926-17e535470807","Type":"ContainerStarted","Data":"95c4c802d94f935a6bedb455b2d30c82a9fd841dba1bb9adbaf116689c840e1d"} Nov 26 00:36:04 crc kubenswrapper[4766]: I1126 00:36:04.002950 4766 generic.go:334] "Generic (PLEG): container finished" podID="64a27a6e-0ec0-42e2-8606-370bafc71c04" containerID="974fa048fc899ef8519cca3a389d7824b87e1c29cd89bd9b1efae3e07d609b14" exitCode=0 Nov 26 00:36:04 crc kubenswrapper[4766]: I1126 00:36:04.002985 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58" event={"ID":"64a27a6e-0ec0-42e2-8606-370bafc71c04","Type":"ContainerDied","Data":"974fa048fc899ef8519cca3a389d7824b87e1c29cd89bd9b1efae3e07d609b14"} Nov 26 00:36:04 crc kubenswrapper[4766]: I1126 00:36:04.003036 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58" event={"ID":"64a27a6e-0ec0-42e2-8606-370bafc71c04","Type":"ContainerStarted","Data":"c37895135af12bb57f111c200fd8f9b8b88828f25e716012335aeffcaea0893d"} Nov 26 00:36:06 crc kubenswrapper[4766]: I1126 00:36:06.014751 4766 generic.go:334] "Generic (PLEG): container finished" podID="36c1a0d9-7873-4ce0-b926-17e535470807" containerID="4bdd22cd9f05384c7cd4b40d450db7182707c6385e28323ed623381201ec482b" exitCode=0 Nov 26 00:36:06 crc kubenswrapper[4766]: I1126 00:36:06.014832 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9" event={"ID":"36c1a0d9-7873-4ce0-b926-17e535470807","Type":"ContainerDied","Data":"4bdd22cd9f05384c7cd4b40d450db7182707c6385e28323ed623381201ec482b"} Nov 26 00:36:06 crc kubenswrapper[4766]: I1126 00:36:06.017364 4766 generic.go:334] "Generic (PLEG): container finished" podID="64a27a6e-0ec0-42e2-8606-370bafc71c04" containerID="1dc72efb0b0ef91e8157d33a42b1c7b4bf5945607ed337a2164225a801a2f467" exitCode=0 Nov 26 00:36:06 crc kubenswrapper[4766]: I1126 00:36:06.017427 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58" event={"ID":"64a27a6e-0ec0-42e2-8606-370bafc71c04","Type":"ContainerDied","Data":"1dc72efb0b0ef91e8157d33a42b1c7b4bf5945607ed337a2164225a801a2f467"} Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.024956 4766 generic.go:334] "Generic (PLEG): container finished" podID="64a27a6e-0ec0-42e2-8606-370bafc71c04" containerID="7032ef70f5fcbfa3c320f2b43073833db6246c2f02b4fcf3149b5cd40e7b69da" exitCode=0 Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.025061 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58" event={"ID":"64a27a6e-0ec0-42e2-8606-370bafc71c04","Type":"ContainerDied","Data":"7032ef70f5fcbfa3c320f2b43073833db6246c2f02b4fcf3149b5cd40e7b69da"} Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.027454 4766 generic.go:334] "Generic (PLEG): container finished" podID="36c1a0d9-7873-4ce0-b926-17e535470807" containerID="4edc2c289745e173c8a7ec53ceb30ee073b1aa43647d8973a5cf23589894fd40" exitCode=0 Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.027498 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9" event={"ID":"36c1a0d9-7873-4ce0-b926-17e535470807","Type":"ContainerDied","Data":"4edc2c289745e173c8a7ec53ceb30ee073b1aa43647d8973a5cf23589894fd40"} Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.061433 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mjvdt"] Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.061712 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" podUID="9aaa727c-565e-4819-801d-9f1787118d20" containerName="controller-manager" containerID="cri-o://04db4c3f77a9f02e92b110390dbb87ad59837534850bb190c39f1178091e033e" gracePeriod=30 Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.157142 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn"] Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.157377 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" podUID="2e6903aa-3f06-44f0-807c-4098ba5fa6d0" containerName="route-controller-manager" containerID="cri-o://0075cab8c7badd77fdfb6505d91198e3e29dcd7ba3fd761756ac9efca6dd1ba1" gracePeriod=30 Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.457319 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.488181 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aaa727c-565e-4819-801d-9f1787118d20-config\") pod \"9aaa727c-565e-4819-801d-9f1787118d20\" (UID: \"9aaa727c-565e-4819-801d-9f1787118d20\") " Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.488225 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9aaa727c-565e-4819-801d-9f1787118d20-client-ca\") pod \"9aaa727c-565e-4819-801d-9f1787118d20\" (UID: \"9aaa727c-565e-4819-801d-9f1787118d20\") " Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.488252 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9aaa727c-565e-4819-801d-9f1787118d20-serving-cert\") pod \"9aaa727c-565e-4819-801d-9f1787118d20\" (UID: \"9aaa727c-565e-4819-801d-9f1787118d20\") " Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.488282 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6j56\" (UniqueName: \"kubernetes.io/projected/9aaa727c-565e-4819-801d-9f1787118d20-kube-api-access-r6j56\") pod \"9aaa727c-565e-4819-801d-9f1787118d20\" (UID: \"9aaa727c-565e-4819-801d-9f1787118d20\") " Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.488303 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9aaa727c-565e-4819-801d-9f1787118d20-proxy-ca-bundles\") pod \"9aaa727c-565e-4819-801d-9f1787118d20\" (UID: \"9aaa727c-565e-4819-801d-9f1787118d20\") " Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.489085 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aaa727c-565e-4819-801d-9f1787118d20-client-ca" (OuterVolumeSpecName: "client-ca") pod "9aaa727c-565e-4819-801d-9f1787118d20" (UID: "9aaa727c-565e-4819-801d-9f1787118d20"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.489122 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aaa727c-565e-4819-801d-9f1787118d20-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9aaa727c-565e-4819-801d-9f1787118d20" (UID: "9aaa727c-565e-4819-801d-9f1787118d20"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.489601 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aaa727c-565e-4819-801d-9f1787118d20-config" (OuterVolumeSpecName: "config") pod "9aaa727c-565e-4819-801d-9f1787118d20" (UID: "9aaa727c-565e-4819-801d-9f1787118d20"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.489867 4766 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9aaa727c-565e-4819-801d-9f1787118d20-client-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.489882 4766 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9aaa727c-565e-4819-801d-9f1787118d20-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.494989 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aaa727c-565e-4819-801d-9f1787118d20-kube-api-access-r6j56" (OuterVolumeSpecName: "kube-api-access-r6j56") pod "9aaa727c-565e-4819-801d-9f1787118d20" (UID: "9aaa727c-565e-4819-801d-9f1787118d20"). InnerVolumeSpecName "kube-api-access-r6j56". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.496571 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aaa727c-565e-4819-801d-9f1787118d20-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9aaa727c-565e-4819-801d-9f1787118d20" (UID: "9aaa727c-565e-4819-801d-9f1787118d20"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.532199 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.590420 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-config\") pod \"2e6903aa-3f06-44f0-807c-4098ba5fa6d0\" (UID: \"2e6903aa-3f06-44f0-807c-4098ba5fa6d0\") " Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.590465 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-client-ca\") pod \"2e6903aa-3f06-44f0-807c-4098ba5fa6d0\" (UID: \"2e6903aa-3f06-44f0-807c-4098ba5fa6d0\") " Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.590515 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwt9w\" (UniqueName: \"kubernetes.io/projected/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-kube-api-access-wwt9w\") pod \"2e6903aa-3f06-44f0-807c-4098ba5fa6d0\" (UID: \"2e6903aa-3f06-44f0-807c-4098ba5fa6d0\") " Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.590534 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-serving-cert\") pod \"2e6903aa-3f06-44f0-807c-4098ba5fa6d0\" (UID: \"2e6903aa-3f06-44f0-807c-4098ba5fa6d0\") " Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.590681 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aaa727c-565e-4819-801d-9f1787118d20-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.590693 4766 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9aaa727c-565e-4819-801d-9f1787118d20-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.590701 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6j56\" (UniqueName: \"kubernetes.io/projected/9aaa727c-565e-4819-801d-9f1787118d20-kube-api-access-r6j56\") on node \"crc\" DevicePath \"\"" Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.591890 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-client-ca" (OuterVolumeSpecName: "client-ca") pod "2e6903aa-3f06-44f0-807c-4098ba5fa6d0" (UID: "2e6903aa-3f06-44f0-807c-4098ba5fa6d0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.591975 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-config" (OuterVolumeSpecName: "config") pod "2e6903aa-3f06-44f0-807c-4098ba5fa6d0" (UID: "2e6903aa-3f06-44f0-807c-4098ba5fa6d0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.593746 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2e6903aa-3f06-44f0-807c-4098ba5fa6d0" (UID: "2e6903aa-3f06-44f0-807c-4098ba5fa6d0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.594455 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-kube-api-access-wwt9w" (OuterVolumeSpecName: "kube-api-access-wwt9w") pod "2e6903aa-3f06-44f0-807c-4098ba5fa6d0" (UID: "2e6903aa-3f06-44f0-807c-4098ba5fa6d0"). InnerVolumeSpecName "kube-api-access-wwt9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.691469 4766 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-client-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.691542 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwt9w\" (UniqueName: \"kubernetes.io/projected/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-kube-api-access-wwt9w\") on node \"crc\" DevicePath \"\"" Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.691568 4766 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:36:07 crc kubenswrapper[4766]: I1126 00:36:07.691589 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e6903aa-3f06-44f0-807c-4098ba5fa6d0-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.035465 4766 generic.go:334] "Generic (PLEG): container finished" podID="9aaa727c-565e-4819-801d-9f1787118d20" containerID="04db4c3f77a9f02e92b110390dbb87ad59837534850bb190c39f1178091e033e" exitCode=0 Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.035540 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.035554 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" event={"ID":"9aaa727c-565e-4819-801d-9f1787118d20","Type":"ContainerDied","Data":"04db4c3f77a9f02e92b110390dbb87ad59837534850bb190c39f1178091e033e"} Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.035588 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-mjvdt" event={"ID":"9aaa727c-565e-4819-801d-9f1787118d20","Type":"ContainerDied","Data":"456addb0ffb294f97e1e8e7b98ebaa82a4171c9db8aa4baf100f0e30643e4309"} Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.035608 4766 scope.go:117] "RemoveContainer" containerID="04db4c3f77a9f02e92b110390dbb87ad59837534850bb190c39f1178091e033e" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.039252 4766 generic.go:334] "Generic (PLEG): container finished" podID="2e6903aa-3f06-44f0-807c-4098ba5fa6d0" containerID="0075cab8c7badd77fdfb6505d91198e3e29dcd7ba3fd761756ac9efca6dd1ba1" exitCode=0 Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.039341 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" event={"ID":"2e6903aa-3f06-44f0-807c-4098ba5fa6d0","Type":"ContainerDied","Data":"0075cab8c7badd77fdfb6505d91198e3e29dcd7ba3fd761756ac9efca6dd1ba1"} Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.039396 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" event={"ID":"2e6903aa-3f06-44f0-807c-4098ba5fa6d0","Type":"ContainerDied","Data":"ad4d720b99f936bd05c9b42318d94b4dbf0397839577c56bdfd3994cbb69713b"} Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.039538 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.063445 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mjvdt"] Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.066709 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mjvdt"] Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.074812 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn"] Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.077688 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dsmkn"] Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.080916 4766 scope.go:117] "RemoveContainer" containerID="04db4c3f77a9f02e92b110390dbb87ad59837534850bb190c39f1178091e033e" Nov 26 00:36:08 crc kubenswrapper[4766]: E1126 00:36:08.081398 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04db4c3f77a9f02e92b110390dbb87ad59837534850bb190c39f1178091e033e\": container with ID starting with 04db4c3f77a9f02e92b110390dbb87ad59837534850bb190c39f1178091e033e not found: ID does not exist" containerID="04db4c3f77a9f02e92b110390dbb87ad59837534850bb190c39f1178091e033e" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.081434 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04db4c3f77a9f02e92b110390dbb87ad59837534850bb190c39f1178091e033e"} err="failed to get container status \"04db4c3f77a9f02e92b110390dbb87ad59837534850bb190c39f1178091e033e\": rpc error: code = NotFound desc = could not find container \"04db4c3f77a9f02e92b110390dbb87ad59837534850bb190c39f1178091e033e\": container with ID starting with 04db4c3f77a9f02e92b110390dbb87ad59837534850bb190c39f1178091e033e not found: ID does not exist" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.081453 4766 scope.go:117] "RemoveContainer" containerID="0075cab8c7badd77fdfb6505d91198e3e29dcd7ba3fd761756ac9efca6dd1ba1" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.103308 4766 scope.go:117] "RemoveContainer" containerID="0075cab8c7badd77fdfb6505d91198e3e29dcd7ba3fd761756ac9efca6dd1ba1" Nov 26 00:36:08 crc kubenswrapper[4766]: E1126 00:36:08.104086 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0075cab8c7badd77fdfb6505d91198e3e29dcd7ba3fd761756ac9efca6dd1ba1\": container with ID starting with 0075cab8c7badd77fdfb6505d91198e3e29dcd7ba3fd761756ac9efca6dd1ba1 not found: ID does not exist" containerID="0075cab8c7badd77fdfb6505d91198e3e29dcd7ba3fd761756ac9efca6dd1ba1" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.104128 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0075cab8c7badd77fdfb6505d91198e3e29dcd7ba3fd761756ac9efca6dd1ba1"} err="failed to get container status \"0075cab8c7badd77fdfb6505d91198e3e29dcd7ba3fd761756ac9efca6dd1ba1\": rpc error: code = NotFound desc = could not find container \"0075cab8c7badd77fdfb6505d91198e3e29dcd7ba3fd761756ac9efca6dd1ba1\": container with ID starting with 0075cab8c7badd77fdfb6505d91198e3e29dcd7ba3fd761756ac9efca6dd1ba1 not found: ID does not exist" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.239339 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-666464cb58-h5zh7"] Nov 26 00:36:08 crc kubenswrapper[4766]: E1126 00:36:08.239584 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e6903aa-3f06-44f0-807c-4098ba5fa6d0" containerName="route-controller-manager" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.239599 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e6903aa-3f06-44f0-807c-4098ba5fa6d0" containerName="route-controller-manager" Nov 26 00:36:08 crc kubenswrapper[4766]: E1126 00:36:08.239617 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aaa727c-565e-4819-801d-9f1787118d20" containerName="controller-manager" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.239626 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aaa727c-565e-4819-801d-9f1787118d20" containerName="controller-manager" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.239767 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e6903aa-3f06-44f0-807c-4098ba5fa6d0" containerName="route-controller-manager" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.239796 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aaa727c-565e-4819-801d-9f1787118d20" containerName="controller-manager" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.240528 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.246977 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.247827 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.248155 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.251009 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.251022 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.251618 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.257470 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.259225 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-666464cb58-h5zh7"] Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.343338 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.394951 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.400370 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fe3b8b5-aaeb-4937-854f-78df48f7a990-config\") pod \"controller-manager-666464cb58-h5zh7\" (UID: \"9fe3b8b5-aaeb-4937-854f-78df48f7a990\") " pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.400448 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9fe3b8b5-aaeb-4937-854f-78df48f7a990-proxy-ca-bundles\") pod \"controller-manager-666464cb58-h5zh7\" (UID: \"9fe3b8b5-aaeb-4937-854f-78df48f7a990\") " pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.400518 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9fe3b8b5-aaeb-4937-854f-78df48f7a990-client-ca\") pod \"controller-manager-666464cb58-h5zh7\" (UID: \"9fe3b8b5-aaeb-4937-854f-78df48f7a990\") " pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.400544 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fe3b8b5-aaeb-4937-854f-78df48f7a990-serving-cert\") pod \"controller-manager-666464cb58-h5zh7\" (UID: \"9fe3b8b5-aaeb-4937-854f-78df48f7a990\") " pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.400570 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h96jf\" (UniqueName: \"kubernetes.io/projected/9fe3b8b5-aaeb-4937-854f-78df48f7a990-kube-api-access-h96jf\") pod \"controller-manager-666464cb58-h5zh7\" (UID: \"9fe3b8b5-aaeb-4937-854f-78df48f7a990\") " pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.501945 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clk2f\" (UniqueName: \"kubernetes.io/projected/36c1a0d9-7873-4ce0-b926-17e535470807-kube-api-access-clk2f\") pod \"36c1a0d9-7873-4ce0-b926-17e535470807\" (UID: \"36c1a0d9-7873-4ce0-b926-17e535470807\") " Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.502001 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/36c1a0d9-7873-4ce0-b926-17e535470807-util\") pod \"36c1a0d9-7873-4ce0-b926-17e535470807\" (UID: \"36c1a0d9-7873-4ce0-b926-17e535470807\") " Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.502025 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64a27a6e-0ec0-42e2-8606-370bafc71c04-util\") pod \"64a27a6e-0ec0-42e2-8606-370bafc71c04\" (UID: \"64a27a6e-0ec0-42e2-8606-370bafc71c04\") " Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.502052 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/36c1a0d9-7873-4ce0-b926-17e535470807-bundle\") pod \"36c1a0d9-7873-4ce0-b926-17e535470807\" (UID: \"36c1a0d9-7873-4ce0-b926-17e535470807\") " Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.502094 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64a27a6e-0ec0-42e2-8606-370bafc71c04-bundle\") pod \"64a27a6e-0ec0-42e2-8606-370bafc71c04\" (UID: \"64a27a6e-0ec0-42e2-8606-370bafc71c04\") " Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.502117 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdgzg\" (UniqueName: \"kubernetes.io/projected/64a27a6e-0ec0-42e2-8606-370bafc71c04-kube-api-access-hdgzg\") pod \"64a27a6e-0ec0-42e2-8606-370bafc71c04\" (UID: \"64a27a6e-0ec0-42e2-8606-370bafc71c04\") " Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.502252 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fe3b8b5-aaeb-4937-854f-78df48f7a990-config\") pod \"controller-manager-666464cb58-h5zh7\" (UID: \"9fe3b8b5-aaeb-4937-854f-78df48f7a990\") " pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.502292 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9fe3b8b5-aaeb-4937-854f-78df48f7a990-proxy-ca-bundles\") pod \"controller-manager-666464cb58-h5zh7\" (UID: \"9fe3b8b5-aaeb-4937-854f-78df48f7a990\") " pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.502329 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9fe3b8b5-aaeb-4937-854f-78df48f7a990-client-ca\") pod \"controller-manager-666464cb58-h5zh7\" (UID: \"9fe3b8b5-aaeb-4937-854f-78df48f7a990\") " pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.502348 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fe3b8b5-aaeb-4937-854f-78df48f7a990-serving-cert\") pod \"controller-manager-666464cb58-h5zh7\" (UID: \"9fe3b8b5-aaeb-4937-854f-78df48f7a990\") " pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.502371 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h96jf\" (UniqueName: \"kubernetes.io/projected/9fe3b8b5-aaeb-4937-854f-78df48f7a990-kube-api-access-h96jf\") pod \"controller-manager-666464cb58-h5zh7\" (UID: \"9fe3b8b5-aaeb-4937-854f-78df48f7a990\") " pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.503719 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64a27a6e-0ec0-42e2-8606-370bafc71c04-bundle" (OuterVolumeSpecName: "bundle") pod "64a27a6e-0ec0-42e2-8606-370bafc71c04" (UID: "64a27a6e-0ec0-42e2-8606-370bafc71c04"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.503922 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9fe3b8b5-aaeb-4937-854f-78df48f7a990-proxy-ca-bundles\") pod \"controller-manager-666464cb58-h5zh7\" (UID: \"9fe3b8b5-aaeb-4937-854f-78df48f7a990\") " pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.504507 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9fe3b8b5-aaeb-4937-854f-78df48f7a990-client-ca\") pod \"controller-manager-666464cb58-h5zh7\" (UID: \"9fe3b8b5-aaeb-4937-854f-78df48f7a990\") " pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.506776 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64a27a6e-0ec0-42e2-8606-370bafc71c04-kube-api-access-hdgzg" (OuterVolumeSpecName: "kube-api-access-hdgzg") pod "64a27a6e-0ec0-42e2-8606-370bafc71c04" (UID: "64a27a6e-0ec0-42e2-8606-370bafc71c04"). InnerVolumeSpecName "kube-api-access-hdgzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.507108 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36c1a0d9-7873-4ce0-b926-17e535470807-bundle" (OuterVolumeSpecName: "bundle") pod "36c1a0d9-7873-4ce0-b926-17e535470807" (UID: "36c1a0d9-7873-4ce0-b926-17e535470807"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.507988 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fe3b8b5-aaeb-4937-854f-78df48f7a990-config\") pod \"controller-manager-666464cb58-h5zh7\" (UID: \"9fe3b8b5-aaeb-4937-854f-78df48f7a990\") " pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.509812 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36c1a0d9-7873-4ce0-b926-17e535470807-kube-api-access-clk2f" (OuterVolumeSpecName: "kube-api-access-clk2f") pod "36c1a0d9-7873-4ce0-b926-17e535470807" (UID: "36c1a0d9-7873-4ce0-b926-17e535470807"). InnerVolumeSpecName "kube-api-access-clk2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.510041 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fe3b8b5-aaeb-4937-854f-78df48f7a990-serving-cert\") pod \"controller-manager-666464cb58-h5zh7\" (UID: \"9fe3b8b5-aaeb-4937-854f-78df48f7a990\") " pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.518033 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h96jf\" (UniqueName: \"kubernetes.io/projected/9fe3b8b5-aaeb-4937-854f-78df48f7a990-kube-api-access-h96jf\") pod \"controller-manager-666464cb58-h5zh7\" (UID: \"9fe3b8b5-aaeb-4937-854f-78df48f7a990\") " pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.519273 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36c1a0d9-7873-4ce0-b926-17e535470807-util" (OuterVolumeSpecName: "util") pod "36c1a0d9-7873-4ce0-b926-17e535470807" (UID: "36c1a0d9-7873-4ce0-b926-17e535470807"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.519552 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64a27a6e-0ec0-42e2-8606-370bafc71c04-util" (OuterVolumeSpecName: "util") pod "64a27a6e-0ec0-42e2-8606-370bafc71c04" (UID: "64a27a6e-0ec0-42e2-8606-370bafc71c04"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.572495 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.604093 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clk2f\" (UniqueName: \"kubernetes.io/projected/36c1a0d9-7873-4ce0-b926-17e535470807-kube-api-access-clk2f\") on node \"crc\" DevicePath \"\"" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.604129 4766 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/36c1a0d9-7873-4ce0-b926-17e535470807-util\") on node \"crc\" DevicePath \"\"" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.604141 4766 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64a27a6e-0ec0-42e2-8606-370bafc71c04-util\") on node \"crc\" DevicePath \"\"" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.604152 4766 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/36c1a0d9-7873-4ce0-b926-17e535470807-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.604162 4766 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64a27a6e-0ec0-42e2-8606-370bafc71c04-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.604175 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdgzg\" (UniqueName: \"kubernetes.io/projected/64a27a6e-0ec0-42e2-8606-370bafc71c04-kube-api-access-hdgzg\") on node \"crc\" DevicePath \"\"" Nov 26 00:36:08 crc kubenswrapper[4766]: I1126 00:36:08.756545 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-666464cb58-h5zh7"] Nov 26 00:36:08 crc kubenswrapper[4766]: W1126 00:36:08.765211 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9fe3b8b5_aaeb_4937_854f_78df48f7a990.slice/crio-07d89417c4fad19ea21440d6e12be35a4eeedda92d148e0c49fd5b354c93ef40 WatchSource:0}: Error finding container 07d89417c4fad19ea21440d6e12be35a4eeedda92d148e0c49fd5b354c93ef40: Status 404 returned error can't find the container with id 07d89417c4fad19ea21440d6e12be35a4eeedda92d148e0c49fd5b354c93ef40 Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.047108 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58" event={"ID":"64a27a6e-0ec0-42e2-8606-370bafc71c04","Type":"ContainerDied","Data":"c37895135af12bb57f111c200fd8f9b8b88828f25e716012335aeffcaea0893d"} Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.047160 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c37895135af12bb57f111c200fd8f9b8b88828f25e716012335aeffcaea0893d" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.047122 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.048284 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" event={"ID":"9fe3b8b5-aaeb-4937-854f-78df48f7a990","Type":"ContainerStarted","Data":"062298d9955c912b5d55e5d10e6b7dbab9b20cbc36ddaf866631b74747b6c0ae"} Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.048339 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" event={"ID":"9fe3b8b5-aaeb-4937-854f-78df48f7a990","Type":"ContainerStarted","Data":"07d89417c4fad19ea21440d6e12be35a4eeedda92d148e0c49fd5b354c93ef40"} Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.048678 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.052149 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.052152 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9" event={"ID":"36c1a0d9-7873-4ce0-b926-17e535470807","Type":"ContainerDied","Data":"95c4c802d94f935a6bedb455b2d30c82a9fd841dba1bb9adbaf116689c840e1d"} Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.052202 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95c4c802d94f935a6bedb455b2d30c82a9fd841dba1bb9adbaf116689c840e1d" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.071916 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.088294 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-666464cb58-h5zh7" podStartSLOduration=2.088277607 podStartE2EDuration="2.088277607s" podCreationTimestamp="2025-11-26 00:36:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:36:09.082497081 +0000 UTC m=+749.931267501" watchObservedRunningTime="2025-11-26 00:36:09.088277607 +0000 UTC m=+749.937048037" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.241067 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m"] Nov 26 00:36:09 crc kubenswrapper[4766]: E1126 00:36:09.241303 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36c1a0d9-7873-4ce0-b926-17e535470807" containerName="pull" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.241318 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="36c1a0d9-7873-4ce0-b926-17e535470807" containerName="pull" Nov 26 00:36:09 crc kubenswrapper[4766]: E1126 00:36:09.241328 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64a27a6e-0ec0-42e2-8606-370bafc71c04" containerName="extract" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.241335 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="64a27a6e-0ec0-42e2-8606-370bafc71c04" containerName="extract" Nov 26 00:36:09 crc kubenswrapper[4766]: E1126 00:36:09.241343 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36c1a0d9-7873-4ce0-b926-17e535470807" containerName="util" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.241350 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="36c1a0d9-7873-4ce0-b926-17e535470807" containerName="util" Nov 26 00:36:09 crc kubenswrapper[4766]: E1126 00:36:09.241358 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64a27a6e-0ec0-42e2-8606-370bafc71c04" containerName="util" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.241363 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="64a27a6e-0ec0-42e2-8606-370bafc71c04" containerName="util" Nov 26 00:36:09 crc kubenswrapper[4766]: E1126 00:36:09.241372 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64a27a6e-0ec0-42e2-8606-370bafc71c04" containerName="pull" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.241378 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="64a27a6e-0ec0-42e2-8606-370bafc71c04" containerName="pull" Nov 26 00:36:09 crc kubenswrapper[4766]: E1126 00:36:09.241387 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36c1a0d9-7873-4ce0-b926-17e535470807" containerName="extract" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.241392 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="36c1a0d9-7873-4ce0-b926-17e535470807" containerName="extract" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.241495 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="36c1a0d9-7873-4ce0-b926-17e535470807" containerName="extract" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.241505 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="64a27a6e-0ec0-42e2-8606-370bafc71c04" containerName="extract" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.241937 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.263547 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.263701 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.264253 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.264284 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.264413 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.265722 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.268721 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m"] Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.416572 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/175486f1-eead-4c06-b0d4-d24c5b50512b-client-ca\") pod \"route-controller-manager-8596487f4c-fdx4m\" (UID: \"175486f1-eead-4c06-b0d4-d24c5b50512b\") " pod="openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.416702 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/175486f1-eead-4c06-b0d4-d24c5b50512b-serving-cert\") pod \"route-controller-manager-8596487f4c-fdx4m\" (UID: \"175486f1-eead-4c06-b0d4-d24c5b50512b\") " pod="openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.416723 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5rz2\" (UniqueName: \"kubernetes.io/projected/175486f1-eead-4c06-b0d4-d24c5b50512b-kube-api-access-s5rz2\") pod \"route-controller-manager-8596487f4c-fdx4m\" (UID: \"175486f1-eead-4c06-b0d4-d24c5b50512b\") " pod="openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.416785 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/175486f1-eead-4c06-b0d4-d24c5b50512b-config\") pod \"route-controller-manager-8596487f4c-fdx4m\" (UID: \"175486f1-eead-4c06-b0d4-d24c5b50512b\") " pod="openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.518204 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/175486f1-eead-4c06-b0d4-d24c5b50512b-serving-cert\") pod \"route-controller-manager-8596487f4c-fdx4m\" (UID: \"175486f1-eead-4c06-b0d4-d24c5b50512b\") " pod="openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.518252 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5rz2\" (UniqueName: \"kubernetes.io/projected/175486f1-eead-4c06-b0d4-d24c5b50512b-kube-api-access-s5rz2\") pod \"route-controller-manager-8596487f4c-fdx4m\" (UID: \"175486f1-eead-4c06-b0d4-d24c5b50512b\") " pod="openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.518297 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/175486f1-eead-4c06-b0d4-d24c5b50512b-config\") pod \"route-controller-manager-8596487f4c-fdx4m\" (UID: \"175486f1-eead-4c06-b0d4-d24c5b50512b\") " pod="openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.518374 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/175486f1-eead-4c06-b0d4-d24c5b50512b-client-ca\") pod \"route-controller-manager-8596487f4c-fdx4m\" (UID: \"175486f1-eead-4c06-b0d4-d24c5b50512b\") " pod="openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.520646 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/175486f1-eead-4c06-b0d4-d24c5b50512b-client-ca\") pod \"route-controller-manager-8596487f4c-fdx4m\" (UID: \"175486f1-eead-4c06-b0d4-d24c5b50512b\") " pod="openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.521609 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/175486f1-eead-4c06-b0d4-d24c5b50512b-config\") pod \"route-controller-manager-8596487f4c-fdx4m\" (UID: \"175486f1-eead-4c06-b0d4-d24c5b50512b\") " pod="openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.525035 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/175486f1-eead-4c06-b0d4-d24c5b50512b-serving-cert\") pod \"route-controller-manager-8596487f4c-fdx4m\" (UID: \"175486f1-eead-4c06-b0d4-d24c5b50512b\") " pod="openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.543980 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5rz2\" (UniqueName: \"kubernetes.io/projected/175486f1-eead-4c06-b0d4-d24c5b50512b-kube-api-access-s5rz2\") pod \"route-controller-manager-8596487f4c-fdx4m\" (UID: \"175486f1-eead-4c06-b0d4-d24c5b50512b\") " pod="openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.556727 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.834467 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e6903aa-3f06-44f0-807c-4098ba5fa6d0" path="/var/lib/kubelet/pods/2e6903aa-3f06-44f0-807c-4098ba5fa6d0/volumes" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.835394 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aaa727c-565e-4819-801d-9f1787118d20" path="/var/lib/kubelet/pods/9aaa727c-565e-4819-801d-9f1787118d20/volumes" Nov 26 00:36:09 crc kubenswrapper[4766]: I1126 00:36:09.959243 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m"] Nov 26 00:36:09 crc kubenswrapper[4766]: W1126 00:36:09.963463 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod175486f1_eead_4c06_b0d4_d24c5b50512b.slice/crio-364b1392fe4522712d2ba9f4a7d00054acc454d3a59904521e52460d4f78e42d WatchSource:0}: Error finding container 364b1392fe4522712d2ba9f4a7d00054acc454d3a59904521e52460d4f78e42d: Status 404 returned error can't find the container with id 364b1392fe4522712d2ba9f4a7d00054acc454d3a59904521e52460d4f78e42d Nov 26 00:36:10 crc kubenswrapper[4766]: I1126 00:36:10.108677 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m" event={"ID":"175486f1-eead-4c06-b0d4-d24c5b50512b","Type":"ContainerStarted","Data":"364b1392fe4522712d2ba9f4a7d00054acc454d3a59904521e52460d4f78e42d"} Nov 26 00:36:11 crc kubenswrapper[4766]: I1126 00:36:11.116151 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m" event={"ID":"175486f1-eead-4c06-b0d4-d24c5b50512b","Type":"ContainerStarted","Data":"6a9bbf358658c0a7f531b047c0ebe2301890dad7dfeb1f633c4993fc1db65788"} Nov 26 00:36:11 crc kubenswrapper[4766]: I1126 00:36:11.140423 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m" podStartSLOduration=4.140404258 podStartE2EDuration="4.140404258s" podCreationTimestamp="2025-11-26 00:36:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:36:11.138053288 +0000 UTC m=+751.986823728" watchObservedRunningTime="2025-11-26 00:36:11.140404258 +0000 UTC m=+751.989174698" Nov 26 00:36:11 crc kubenswrapper[4766]: I1126 00:36:11.479993 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:36:11 crc kubenswrapper[4766]: I1126 00:36:11.480054 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.394167 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m" Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.404411 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-8596487f4c-fdx4m" Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.444186 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9wnr8"] Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.445546 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9wnr8" Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.469928 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9wnr8"] Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.488734 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-mq2w5"] Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.489620 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-mq2w5" Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.497842 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"kube-root-ca.crt" Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.498138 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"openshift-service-ca.crt" Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.498762 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"cluster-logging-operator-dockercfg-5hkff" Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.518144 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-mq2w5"] Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.591571 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5afbc53d-2420-468c-b993-2ca253705495-catalog-content\") pod \"redhat-operators-9wnr8\" (UID: \"5afbc53d-2420-468c-b993-2ca253705495\") " pod="openshift-marketplace/redhat-operators-9wnr8" Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.591618 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdmft\" (UniqueName: \"kubernetes.io/projected/2850bfaf-5303-44e6-8fc6-616612cb68e1-kube-api-access-pdmft\") pod \"cluster-logging-operator-ff9846bd-mq2w5\" (UID: \"2850bfaf-5303-44e6-8fc6-616612cb68e1\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-mq2w5" Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.591696 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bphcz\" (UniqueName: \"kubernetes.io/projected/5afbc53d-2420-468c-b993-2ca253705495-kube-api-access-bphcz\") pod \"redhat-operators-9wnr8\" (UID: \"5afbc53d-2420-468c-b993-2ca253705495\") " pod="openshift-marketplace/redhat-operators-9wnr8" Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.591715 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5afbc53d-2420-468c-b993-2ca253705495-utilities\") pod \"redhat-operators-9wnr8\" (UID: \"5afbc53d-2420-468c-b993-2ca253705495\") " pod="openshift-marketplace/redhat-operators-9wnr8" Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.692554 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5afbc53d-2420-468c-b993-2ca253705495-catalog-content\") pod \"redhat-operators-9wnr8\" (UID: \"5afbc53d-2420-468c-b993-2ca253705495\") " pod="openshift-marketplace/redhat-operators-9wnr8" Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.692624 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdmft\" (UniqueName: \"kubernetes.io/projected/2850bfaf-5303-44e6-8fc6-616612cb68e1-kube-api-access-pdmft\") pod \"cluster-logging-operator-ff9846bd-mq2w5\" (UID: \"2850bfaf-5303-44e6-8fc6-616612cb68e1\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-mq2w5" Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.692674 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bphcz\" (UniqueName: \"kubernetes.io/projected/5afbc53d-2420-468c-b993-2ca253705495-kube-api-access-bphcz\") pod \"redhat-operators-9wnr8\" (UID: \"5afbc53d-2420-468c-b993-2ca253705495\") " pod="openshift-marketplace/redhat-operators-9wnr8" Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.692702 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5afbc53d-2420-468c-b993-2ca253705495-utilities\") pod \"redhat-operators-9wnr8\" (UID: \"5afbc53d-2420-468c-b993-2ca253705495\") " pod="openshift-marketplace/redhat-operators-9wnr8" Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.693146 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5afbc53d-2420-468c-b993-2ca253705495-catalog-content\") pod \"redhat-operators-9wnr8\" (UID: \"5afbc53d-2420-468c-b993-2ca253705495\") " pod="openshift-marketplace/redhat-operators-9wnr8" Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.693226 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5afbc53d-2420-468c-b993-2ca253705495-utilities\") pod \"redhat-operators-9wnr8\" (UID: \"5afbc53d-2420-468c-b993-2ca253705495\") " pod="openshift-marketplace/redhat-operators-9wnr8" Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.714691 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bphcz\" (UniqueName: \"kubernetes.io/projected/5afbc53d-2420-468c-b993-2ca253705495-kube-api-access-bphcz\") pod \"redhat-operators-9wnr8\" (UID: \"5afbc53d-2420-468c-b993-2ca253705495\") " pod="openshift-marketplace/redhat-operators-9wnr8" Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.714956 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdmft\" (UniqueName: \"kubernetes.io/projected/2850bfaf-5303-44e6-8fc6-616612cb68e1-kube-api-access-pdmft\") pod \"cluster-logging-operator-ff9846bd-mq2w5\" (UID: \"2850bfaf-5303-44e6-8fc6-616612cb68e1\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-mq2w5" Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.771843 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9wnr8" Nov 26 00:36:12 crc kubenswrapper[4766]: I1126 00:36:12.818882 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-mq2w5" Nov 26 00:36:13 crc kubenswrapper[4766]: I1126 00:36:13.186070 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9wnr8"] Nov 26 00:36:13 crc kubenswrapper[4766]: W1126 00:36:13.196017 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5afbc53d_2420_468c_b993_2ca253705495.slice/crio-fd30c6f73e68262ce0aa773a9f2c1bba572f40cbc67f6e8989dca53f2a89cbfa WatchSource:0}: Error finding container fd30c6f73e68262ce0aa773a9f2c1bba572f40cbc67f6e8989dca53f2a89cbfa: Status 404 returned error can't find the container with id fd30c6f73e68262ce0aa773a9f2c1bba572f40cbc67f6e8989dca53f2a89cbfa Nov 26 00:36:13 crc kubenswrapper[4766]: I1126 00:36:13.399785 4766 generic.go:334] "Generic (PLEG): container finished" podID="5afbc53d-2420-468c-b993-2ca253705495" containerID="1dc7216892c9d2e9031f41f09daa351b7076a9fe2dac1bb4bae49f552a7f90d9" exitCode=0 Nov 26 00:36:13 crc kubenswrapper[4766]: I1126 00:36:13.400268 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9wnr8" event={"ID":"5afbc53d-2420-468c-b993-2ca253705495","Type":"ContainerDied","Data":"1dc7216892c9d2e9031f41f09daa351b7076a9fe2dac1bb4bae49f552a7f90d9"} Nov 26 00:36:13 crc kubenswrapper[4766]: I1126 00:36:13.400316 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9wnr8" event={"ID":"5afbc53d-2420-468c-b993-2ca253705495","Type":"ContainerStarted","Data":"fd30c6f73e68262ce0aa773a9f2c1bba572f40cbc67f6e8989dca53f2a89cbfa"} Nov 26 00:36:13 crc kubenswrapper[4766]: I1126 00:36:13.426357 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-mq2w5"] Nov 26 00:36:13 crc kubenswrapper[4766]: W1126 00:36:13.431246 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2850bfaf_5303_44e6_8fc6_616612cb68e1.slice/crio-247a0f88bb39acd3bdfc6978092d79bd0f2aecb9062076ac98686fe2dcfbf699 WatchSource:0}: Error finding container 247a0f88bb39acd3bdfc6978092d79bd0f2aecb9062076ac98686fe2dcfbf699: Status 404 returned error can't find the container with id 247a0f88bb39acd3bdfc6978092d79bd0f2aecb9062076ac98686fe2dcfbf699 Nov 26 00:36:13 crc kubenswrapper[4766]: I1126 00:36:13.820677 4766 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 26 00:36:14 crc kubenswrapper[4766]: I1126 00:36:14.406461 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9wnr8" event={"ID":"5afbc53d-2420-468c-b993-2ca253705495","Type":"ContainerStarted","Data":"e705713ce7a2d4cd62d4544a3d21cc829bdcd866b7d0b2c681b2ec62354a9e02"} Nov 26 00:36:14 crc kubenswrapper[4766]: I1126 00:36:14.407519 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-mq2w5" event={"ID":"2850bfaf-5303-44e6-8fc6-616612cb68e1","Type":"ContainerStarted","Data":"247a0f88bb39acd3bdfc6978092d79bd0f2aecb9062076ac98686fe2dcfbf699"} Nov 26 00:36:15 crc kubenswrapper[4766]: I1126 00:36:15.422020 4766 generic.go:334] "Generic (PLEG): container finished" podID="5afbc53d-2420-468c-b993-2ca253705495" containerID="e705713ce7a2d4cd62d4544a3d21cc829bdcd866b7d0b2c681b2ec62354a9e02" exitCode=0 Nov 26 00:36:15 crc kubenswrapper[4766]: I1126 00:36:15.422075 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9wnr8" event={"ID":"5afbc53d-2420-468c-b993-2ca253705495","Type":"ContainerDied","Data":"e705713ce7a2d4cd62d4544a3d21cc829bdcd866b7d0b2c681b2ec62354a9e02"} Nov 26 00:36:19 crc kubenswrapper[4766]: I1126 00:36:19.445938 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9wnr8" event={"ID":"5afbc53d-2420-468c-b993-2ca253705495","Type":"ContainerStarted","Data":"2296ed958a42582968b336519b8f6156036edb45ce18d088faf8c38bf11b973b"} Nov 26 00:36:19 crc kubenswrapper[4766]: I1126 00:36:19.448606 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-mq2w5" event={"ID":"2850bfaf-5303-44e6-8fc6-616612cb68e1","Type":"ContainerStarted","Data":"b0bd64949825d706bbf947aa82fdcadcb40d20ecd234531176fba60a0f6cbd19"} Nov 26 00:36:19 crc kubenswrapper[4766]: I1126 00:36:19.474339 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9wnr8" podStartSLOduration=1.748305607 podStartE2EDuration="7.474318511s" podCreationTimestamp="2025-11-26 00:36:12 +0000 UTC" firstStartedPulling="2025-11-26 00:36:13.401402835 +0000 UTC m=+754.250173265" lastFinishedPulling="2025-11-26 00:36:19.127415739 +0000 UTC m=+759.976186169" observedRunningTime="2025-11-26 00:36:19.470066064 +0000 UTC m=+760.318836494" watchObservedRunningTime="2025-11-26 00:36:19.474318511 +0000 UTC m=+760.323088941" Nov 26 00:36:19 crc kubenswrapper[4766]: I1126 00:36:19.498462 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/cluster-logging-operator-ff9846bd-mq2w5" podStartSLOduration=1.780642354 podStartE2EDuration="7.498442781s" podCreationTimestamp="2025-11-26 00:36:12 +0000 UTC" firstStartedPulling="2025-11-26 00:36:13.435484557 +0000 UTC m=+754.284254987" lastFinishedPulling="2025-11-26 00:36:19.153284984 +0000 UTC m=+760.002055414" observedRunningTime="2025-11-26 00:36:19.490030959 +0000 UTC m=+760.338801389" watchObservedRunningTime="2025-11-26 00:36:19.498442781 +0000 UTC m=+760.347213201" Nov 26 00:36:22 crc kubenswrapper[4766]: I1126 00:36:22.772891 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9wnr8" Nov 26 00:36:22 crc kubenswrapper[4766]: I1126 00:36:22.773273 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9wnr8" Nov 26 00:36:23 crc kubenswrapper[4766]: I1126 00:36:23.824938 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9wnr8" podUID="5afbc53d-2420-468c-b993-2ca253705495" containerName="registry-server" probeResult="failure" output=< Nov 26 00:36:23 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 00:36:23 crc kubenswrapper[4766]: > Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.278010 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx"] Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.278937 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.282245 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.282665 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.284091 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.289258 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.289365 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.290206 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-c8mws" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.313495 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx"] Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.481468 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/b90f7164-a5e9-452f-8733-f549f15e17ec-manager-config\") pod \"loki-operator-controller-manager-7bbf4564c5-m77vx\" (UID: \"b90f7164-a5e9-452f-8733-f549f15e17ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.481542 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b90f7164-a5e9-452f-8733-f549f15e17ec-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-7bbf4564c5-m77vx\" (UID: \"b90f7164-a5e9-452f-8733-f549f15e17ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.481577 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b90f7164-a5e9-452f-8733-f549f15e17ec-apiservice-cert\") pod \"loki-operator-controller-manager-7bbf4564c5-m77vx\" (UID: \"b90f7164-a5e9-452f-8733-f549f15e17ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.481610 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp84z\" (UniqueName: \"kubernetes.io/projected/b90f7164-a5e9-452f-8733-f549f15e17ec-kube-api-access-bp84z\") pod \"loki-operator-controller-manager-7bbf4564c5-m77vx\" (UID: \"b90f7164-a5e9-452f-8733-f549f15e17ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.481703 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b90f7164-a5e9-452f-8733-f549f15e17ec-webhook-cert\") pod \"loki-operator-controller-manager-7bbf4564c5-m77vx\" (UID: \"b90f7164-a5e9-452f-8733-f549f15e17ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.582542 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b90f7164-a5e9-452f-8733-f549f15e17ec-webhook-cert\") pod \"loki-operator-controller-manager-7bbf4564c5-m77vx\" (UID: \"b90f7164-a5e9-452f-8733-f549f15e17ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.582672 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/b90f7164-a5e9-452f-8733-f549f15e17ec-manager-config\") pod \"loki-operator-controller-manager-7bbf4564c5-m77vx\" (UID: \"b90f7164-a5e9-452f-8733-f549f15e17ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.582704 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b90f7164-a5e9-452f-8733-f549f15e17ec-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-7bbf4564c5-m77vx\" (UID: \"b90f7164-a5e9-452f-8733-f549f15e17ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.582726 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b90f7164-a5e9-452f-8733-f549f15e17ec-apiservice-cert\") pod \"loki-operator-controller-manager-7bbf4564c5-m77vx\" (UID: \"b90f7164-a5e9-452f-8733-f549f15e17ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.582753 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp84z\" (UniqueName: \"kubernetes.io/projected/b90f7164-a5e9-452f-8733-f549f15e17ec-kube-api-access-bp84z\") pod \"loki-operator-controller-manager-7bbf4564c5-m77vx\" (UID: \"b90f7164-a5e9-452f-8733-f549f15e17ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.584317 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/b90f7164-a5e9-452f-8733-f549f15e17ec-manager-config\") pod \"loki-operator-controller-manager-7bbf4564c5-m77vx\" (UID: \"b90f7164-a5e9-452f-8733-f549f15e17ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.593557 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b90f7164-a5e9-452f-8733-f549f15e17ec-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-7bbf4564c5-m77vx\" (UID: \"b90f7164-a5e9-452f-8733-f549f15e17ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.593575 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b90f7164-a5e9-452f-8733-f549f15e17ec-webhook-cert\") pod \"loki-operator-controller-manager-7bbf4564c5-m77vx\" (UID: \"b90f7164-a5e9-452f-8733-f549f15e17ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.605014 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b90f7164-a5e9-452f-8733-f549f15e17ec-apiservice-cert\") pod \"loki-operator-controller-manager-7bbf4564c5-m77vx\" (UID: \"b90f7164-a5e9-452f-8733-f549f15e17ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.607502 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bp84z\" (UniqueName: \"kubernetes.io/projected/b90f7164-a5e9-452f-8733-f549f15e17ec-kube-api-access-bp84z\") pod \"loki-operator-controller-manager-7bbf4564c5-m77vx\" (UID: \"b90f7164-a5e9-452f-8733-f549f15e17ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" Nov 26 00:36:24 crc kubenswrapper[4766]: I1126 00:36:24.901307 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" Nov 26 00:36:25 crc kubenswrapper[4766]: I1126 00:36:25.319403 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx"] Nov 26 00:36:25 crc kubenswrapper[4766]: W1126 00:36:25.324380 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb90f7164_a5e9_452f_8733_f549f15e17ec.slice/crio-3e4fce83f2acc33c33f38730529a6b73b950295e08a8698a8245116eb1c81fd0 WatchSource:0}: Error finding container 3e4fce83f2acc33c33f38730529a6b73b950295e08a8698a8245116eb1c81fd0: Status 404 returned error can't find the container with id 3e4fce83f2acc33c33f38730529a6b73b950295e08a8698a8245116eb1c81fd0 Nov 26 00:36:25 crc kubenswrapper[4766]: I1126 00:36:25.479454 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" event={"ID":"b90f7164-a5e9-452f-8733-f549f15e17ec","Type":"ContainerStarted","Data":"3e4fce83f2acc33c33f38730529a6b73b950295e08a8698a8245116eb1c81fd0"} Nov 26 00:36:30 crc kubenswrapper[4766]: I1126 00:36:30.512844 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" event={"ID":"b90f7164-a5e9-452f-8733-f549f15e17ec","Type":"ContainerStarted","Data":"29a28daa257ad0dbe49f0e720598861e7a54f0846b11c442b98c25eeb4f4d57f"} Nov 26 00:36:32 crc kubenswrapper[4766]: I1126 00:36:32.819230 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9wnr8" Nov 26 00:36:32 crc kubenswrapper[4766]: I1126 00:36:32.871826 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9wnr8" Nov 26 00:36:35 crc kubenswrapper[4766]: I1126 00:36:35.229084 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9wnr8"] Nov 26 00:36:35 crc kubenswrapper[4766]: I1126 00:36:35.229664 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9wnr8" podUID="5afbc53d-2420-468c-b993-2ca253705495" containerName="registry-server" containerID="cri-o://2296ed958a42582968b336519b8f6156036edb45ce18d088faf8c38bf11b973b" gracePeriod=2 Nov 26 00:36:35 crc kubenswrapper[4766]: I1126 00:36:35.543187 4766 generic.go:334] "Generic (PLEG): container finished" podID="5afbc53d-2420-468c-b993-2ca253705495" containerID="2296ed958a42582968b336519b8f6156036edb45ce18d088faf8c38bf11b973b" exitCode=0 Nov 26 00:36:35 crc kubenswrapper[4766]: I1126 00:36:35.543239 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9wnr8" event={"ID":"5afbc53d-2420-468c-b993-2ca253705495","Type":"ContainerDied","Data":"2296ed958a42582968b336519b8f6156036edb45ce18d088faf8c38bf11b973b"} Nov 26 00:36:36 crc kubenswrapper[4766]: I1126 00:36:36.550603 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" event={"ID":"b90f7164-a5e9-452f-8733-f549f15e17ec","Type":"ContainerStarted","Data":"fe9b6677013f76e6fd3c1b223fd0342dce245d0790b5d103c621df97ac8b7c6c"} Nov 26 00:36:36 crc kubenswrapper[4766]: I1126 00:36:36.551392 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" Nov 26 00:36:36 crc kubenswrapper[4766]: I1126 00:36:36.556215 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" Nov 26 00:36:36 crc kubenswrapper[4766]: I1126 00:36:36.582747 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" podStartSLOduration=1.646177302 podStartE2EDuration="12.582728972s" podCreationTimestamp="2025-11-26 00:36:24 +0000 UTC" firstStartedPulling="2025-11-26 00:36:25.327429113 +0000 UTC m=+766.176199543" lastFinishedPulling="2025-11-26 00:36:36.263980783 +0000 UTC m=+777.112751213" observedRunningTime="2025-11-26 00:36:36.576197147 +0000 UTC m=+777.424967587" watchObservedRunningTime="2025-11-26 00:36:36.582728972 +0000 UTC m=+777.431499402" Nov 26 00:36:36 crc kubenswrapper[4766]: I1126 00:36:36.593278 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9wnr8" Nov 26 00:36:36 crc kubenswrapper[4766]: I1126 00:36:36.717936 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bphcz\" (UniqueName: \"kubernetes.io/projected/5afbc53d-2420-468c-b993-2ca253705495-kube-api-access-bphcz\") pod \"5afbc53d-2420-468c-b993-2ca253705495\" (UID: \"5afbc53d-2420-468c-b993-2ca253705495\") " Nov 26 00:36:36 crc kubenswrapper[4766]: I1126 00:36:36.718078 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5afbc53d-2420-468c-b993-2ca253705495-catalog-content\") pod \"5afbc53d-2420-468c-b993-2ca253705495\" (UID: \"5afbc53d-2420-468c-b993-2ca253705495\") " Nov 26 00:36:36 crc kubenswrapper[4766]: I1126 00:36:36.718156 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5afbc53d-2420-468c-b993-2ca253705495-utilities\") pod \"5afbc53d-2420-468c-b993-2ca253705495\" (UID: \"5afbc53d-2420-468c-b993-2ca253705495\") " Nov 26 00:36:36 crc kubenswrapper[4766]: I1126 00:36:36.718981 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5afbc53d-2420-468c-b993-2ca253705495-utilities" (OuterVolumeSpecName: "utilities") pod "5afbc53d-2420-468c-b993-2ca253705495" (UID: "5afbc53d-2420-468c-b993-2ca253705495"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:36:36 crc kubenswrapper[4766]: I1126 00:36:36.723353 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5afbc53d-2420-468c-b993-2ca253705495-kube-api-access-bphcz" (OuterVolumeSpecName: "kube-api-access-bphcz") pod "5afbc53d-2420-468c-b993-2ca253705495" (UID: "5afbc53d-2420-468c-b993-2ca253705495"). InnerVolumeSpecName "kube-api-access-bphcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:36:36 crc kubenswrapper[4766]: I1126 00:36:36.806567 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5afbc53d-2420-468c-b993-2ca253705495-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5afbc53d-2420-468c-b993-2ca253705495" (UID: "5afbc53d-2420-468c-b993-2ca253705495"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:36:36 crc kubenswrapper[4766]: I1126 00:36:36.820182 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5afbc53d-2420-468c-b993-2ca253705495-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:36:36 crc kubenswrapper[4766]: I1126 00:36:36.820231 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5afbc53d-2420-468c-b993-2ca253705495-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:36:36 crc kubenswrapper[4766]: I1126 00:36:36.820245 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bphcz\" (UniqueName: \"kubernetes.io/projected/5afbc53d-2420-468c-b993-2ca253705495-kube-api-access-bphcz\") on node \"crc\" DevicePath \"\"" Nov 26 00:36:37 crc kubenswrapper[4766]: I1126 00:36:37.558722 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9wnr8" event={"ID":"5afbc53d-2420-468c-b993-2ca253705495","Type":"ContainerDied","Data":"fd30c6f73e68262ce0aa773a9f2c1bba572f40cbc67f6e8989dca53f2a89cbfa"} Nov 26 00:36:37 crc kubenswrapper[4766]: I1126 00:36:37.558739 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9wnr8" Nov 26 00:36:37 crc kubenswrapper[4766]: I1126 00:36:37.559107 4766 scope.go:117] "RemoveContainer" containerID="2296ed958a42582968b336519b8f6156036edb45ce18d088faf8c38bf11b973b" Nov 26 00:36:37 crc kubenswrapper[4766]: I1126 00:36:37.575768 4766 scope.go:117] "RemoveContainer" containerID="e705713ce7a2d4cd62d4544a3d21cc829bdcd866b7d0b2c681b2ec62354a9e02" Nov 26 00:36:37 crc kubenswrapper[4766]: I1126 00:36:37.592950 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9wnr8"] Nov 26 00:36:37 crc kubenswrapper[4766]: I1126 00:36:37.594996 4766 scope.go:117] "RemoveContainer" containerID="1dc7216892c9d2e9031f41f09daa351b7076a9fe2dac1bb4bae49f552a7f90d9" Nov 26 00:36:37 crc kubenswrapper[4766]: I1126 00:36:37.597476 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9wnr8"] Nov 26 00:36:37 crc kubenswrapper[4766]: I1126 00:36:37.836226 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5afbc53d-2420-468c-b993-2ca253705495" path="/var/lib/kubelet/pods/5afbc53d-2420-468c-b993-2ca253705495/volumes" Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.273235 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Nov 26 00:36:41 crc kubenswrapper[4766]: E1126 00:36:41.274118 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5afbc53d-2420-468c-b993-2ca253705495" containerName="registry-server" Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.274134 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5afbc53d-2420-468c-b993-2ca253705495" containerName="registry-server" Nov 26 00:36:41 crc kubenswrapper[4766]: E1126 00:36:41.274153 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5afbc53d-2420-468c-b993-2ca253705495" containerName="extract-content" Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.274163 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5afbc53d-2420-468c-b993-2ca253705495" containerName="extract-content" Nov 26 00:36:41 crc kubenswrapper[4766]: E1126 00:36:41.274177 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5afbc53d-2420-468c-b993-2ca253705495" containerName="extract-utilities" Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.274186 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5afbc53d-2420-468c-b993-2ca253705495" containerName="extract-utilities" Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.274304 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="5afbc53d-2420-468c-b993-2ca253705495" containerName="registry-server" Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.275056 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.278163 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.278211 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.285774 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.378307 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e5ce3064-98df-48d1-9605-85d8aa455c3d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e5ce3064-98df-48d1-9605-85d8aa455c3d\") pod \"minio\" (UID: \"b5504c01-f067-4bdf-943a-e06bd67284f3\") " pod="minio-dev/minio" Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.378393 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gjm2\" (UniqueName: \"kubernetes.io/projected/b5504c01-f067-4bdf-943a-e06bd67284f3-kube-api-access-4gjm2\") pod \"minio\" (UID: \"b5504c01-f067-4bdf-943a-e06bd67284f3\") " pod="minio-dev/minio" Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.479785 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.479858 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.479914 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.480073 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e5ce3064-98df-48d1-9605-85d8aa455c3d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e5ce3064-98df-48d1-9605-85d8aa455c3d\") pod \"minio\" (UID: \"b5504c01-f067-4bdf-943a-e06bd67284f3\") " pod="minio-dev/minio" Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.480137 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gjm2\" (UniqueName: \"kubernetes.io/projected/b5504c01-f067-4bdf-943a-e06bd67284f3-kube-api-access-4gjm2\") pod \"minio\" (UID: \"b5504c01-f067-4bdf-943a-e06bd67284f3\") " pod="minio-dev/minio" Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.480541 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2f5bd86ee9ff3b132bc7816555a18731d6351ecdb67cc2c61c0c55d699b956fc"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.480598 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://2f5bd86ee9ff3b132bc7816555a18731d6351ecdb67cc2c61c0c55d699b956fc" gracePeriod=600 Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.498033 4766 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.498302 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e5ce3064-98df-48d1-9605-85d8aa455c3d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e5ce3064-98df-48d1-9605-85d8aa455c3d\") pod \"minio\" (UID: \"b5504c01-f067-4bdf-943a-e06bd67284f3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f437ecaa0a7a9669c4e0b0c835a301c0dbf587e483ef743b1585b49ea658c15e/globalmount\"" pod="minio-dev/minio" Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.500034 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gjm2\" (UniqueName: \"kubernetes.io/projected/b5504c01-f067-4bdf-943a-e06bd67284f3-kube-api-access-4gjm2\") pod \"minio\" (UID: \"b5504c01-f067-4bdf-943a-e06bd67284f3\") " pod="minio-dev/minio" Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.532940 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e5ce3064-98df-48d1-9605-85d8aa455c3d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e5ce3064-98df-48d1-9605-85d8aa455c3d\") pod \"minio\" (UID: \"b5504c01-f067-4bdf-943a-e06bd67284f3\") " pod="minio-dev/minio" Nov 26 00:36:41 crc kubenswrapper[4766]: I1126 00:36:41.626367 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Nov 26 00:36:42 crc kubenswrapper[4766]: I1126 00:36:42.056792 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Nov 26 00:36:42 crc kubenswrapper[4766]: W1126 00:36:42.067347 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5504c01_f067_4bdf_943a_e06bd67284f3.slice/crio-321bcb5c3403cf651887eb26bad8483dd397c45ddfd6ead04140a1e9b5e63fdc WatchSource:0}: Error finding container 321bcb5c3403cf651887eb26bad8483dd397c45ddfd6ead04140a1e9b5e63fdc: Status 404 returned error can't find the container with id 321bcb5c3403cf651887eb26bad8483dd397c45ddfd6ead04140a1e9b5e63fdc Nov 26 00:36:42 crc kubenswrapper[4766]: I1126 00:36:42.586634 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"b5504c01-f067-4bdf-943a-e06bd67284f3","Type":"ContainerStarted","Data":"321bcb5c3403cf651887eb26bad8483dd397c45ddfd6ead04140a1e9b5e63fdc"} Nov 26 00:36:42 crc kubenswrapper[4766]: I1126 00:36:42.589271 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="2f5bd86ee9ff3b132bc7816555a18731d6351ecdb67cc2c61c0c55d699b956fc" exitCode=0 Nov 26 00:36:42 crc kubenswrapper[4766]: I1126 00:36:42.589305 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"2f5bd86ee9ff3b132bc7816555a18731d6351ecdb67cc2c61c0c55d699b956fc"} Nov 26 00:36:42 crc kubenswrapper[4766]: I1126 00:36:42.589324 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"d5bd12ab3efcc84d754246df84a11f62a9d2d5fd80aa034846fa0733f2acc7ee"} Nov 26 00:36:42 crc kubenswrapper[4766]: I1126 00:36:42.589339 4766 scope.go:117] "RemoveContainer" containerID="ba0d2e8d44b0113663280906db232a40bb0eaded73c2a35e07448bc5254dc338" Nov 26 00:36:45 crc kubenswrapper[4766]: I1126 00:36:45.616882 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"b5504c01-f067-4bdf-943a-e06bd67284f3","Type":"ContainerStarted","Data":"c45c2b6ac7deefc00b9347f9529eb68771ec93db5e6fa6cfcf7822b21e3016ac"} Nov 26 00:36:45 crc kubenswrapper[4766]: I1126 00:36:45.636551 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=4.802390862 podStartE2EDuration="7.636528932s" podCreationTimestamp="2025-11-26 00:36:38 +0000 UTC" firstStartedPulling="2025-11-26 00:36:42.069446978 +0000 UTC m=+782.918217408" lastFinishedPulling="2025-11-26 00:36:44.903585048 +0000 UTC m=+785.752355478" observedRunningTime="2025-11-26 00:36:45.633782583 +0000 UTC m=+786.482553023" watchObservedRunningTime="2025-11-26 00:36:45.636528932 +0000 UTC m=+786.485299372" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.063028 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-7khhv"] Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.064499 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.080366 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-ca-bundle" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.080639 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-dockercfg-4qhbv" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.080818 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-http" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.081024 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-config" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.081195 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-grpc" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.087003 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-7khhv"] Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.179888 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/bfb11063-6ad6-463b-87e6-eaf9333e2d4a-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-7khhv\" (UID: \"bfb11063-6ad6-463b-87e6-eaf9333e2d4a\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.179945 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/bfb11063-6ad6-463b-87e6-eaf9333e2d4a-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-7khhv\" (UID: \"bfb11063-6ad6-463b-87e6-eaf9333e2d4a\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.180306 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsxgg\" (UniqueName: \"kubernetes.io/projected/bfb11063-6ad6-463b-87e6-eaf9333e2d4a-kube-api-access-tsxgg\") pod \"logging-loki-distributor-76cc67bf56-7khhv\" (UID: \"bfb11063-6ad6-463b-87e6-eaf9333e2d4a\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.180436 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfb11063-6ad6-463b-87e6-eaf9333e2d4a-config\") pod \"logging-loki-distributor-76cc67bf56-7khhv\" (UID: \"bfb11063-6ad6-463b-87e6-eaf9333e2d4a\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.180582 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bfb11063-6ad6-463b-87e6-eaf9333e2d4a-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-7khhv\" (UID: \"bfb11063-6ad6-463b-87e6-eaf9333e2d4a\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.199626 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-tv6v8"] Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.200483 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.203505 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-grpc" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.203860 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-http" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.204749 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-s3" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.214348 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-tv6v8"] Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.281600 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/bfb11063-6ad6-463b-87e6-eaf9333e2d4a-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-7khhv\" (UID: \"bfb11063-6ad6-463b-87e6-eaf9333e2d4a\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.281686 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/bfb11063-6ad6-463b-87e6-eaf9333e2d4a-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-7khhv\" (UID: \"bfb11063-6ad6-463b-87e6-eaf9333e2d4a\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.281728 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/036dedff-92b9-4ad9-89b4-15dfb60e73c2-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-tv6v8\" (UID: \"036dedff-92b9-4ad9-89b4-15dfb60e73c2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.281789 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/036dedff-92b9-4ad9-89b4-15dfb60e73c2-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-tv6v8\" (UID: \"036dedff-92b9-4ad9-89b4-15dfb60e73c2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.281838 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/036dedff-92b9-4ad9-89b4-15dfb60e73c2-config\") pod \"logging-loki-querier-5895d59bb8-tv6v8\" (UID: \"036dedff-92b9-4ad9-89b4-15dfb60e73c2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.281876 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsxgg\" (UniqueName: \"kubernetes.io/projected/bfb11063-6ad6-463b-87e6-eaf9333e2d4a-kube-api-access-tsxgg\") pod \"logging-loki-distributor-76cc67bf56-7khhv\" (UID: \"bfb11063-6ad6-463b-87e6-eaf9333e2d4a\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.281920 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/036dedff-92b9-4ad9-89b4-15dfb60e73c2-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-tv6v8\" (UID: \"036dedff-92b9-4ad9-89b4-15dfb60e73c2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.281936 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/036dedff-92b9-4ad9-89b4-15dfb60e73c2-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-tv6v8\" (UID: \"036dedff-92b9-4ad9-89b4-15dfb60e73c2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.281956 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfb11063-6ad6-463b-87e6-eaf9333e2d4a-config\") pod \"logging-loki-distributor-76cc67bf56-7khhv\" (UID: \"bfb11063-6ad6-463b-87e6-eaf9333e2d4a\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.281993 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjrjk\" (UniqueName: \"kubernetes.io/projected/036dedff-92b9-4ad9-89b4-15dfb60e73c2-kube-api-access-qjrjk\") pod \"logging-loki-querier-5895d59bb8-tv6v8\" (UID: \"036dedff-92b9-4ad9-89b4-15dfb60e73c2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.282029 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bfb11063-6ad6-463b-87e6-eaf9333e2d4a-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-7khhv\" (UID: \"bfb11063-6ad6-463b-87e6-eaf9333e2d4a\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.283230 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bfb11063-6ad6-463b-87e6-eaf9333e2d4a-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-7khhv\" (UID: \"bfb11063-6ad6-463b-87e6-eaf9333e2d4a\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.283422 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfb11063-6ad6-463b-87e6-eaf9333e2d4a-config\") pod \"logging-loki-distributor-76cc67bf56-7khhv\" (UID: \"bfb11063-6ad6-463b-87e6-eaf9333e2d4a\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.287936 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/bfb11063-6ad6-463b-87e6-eaf9333e2d4a-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-7khhv\" (UID: \"bfb11063-6ad6-463b-87e6-eaf9333e2d4a\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.291550 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/bfb11063-6ad6-463b-87e6-eaf9333e2d4a-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-7khhv\" (UID: \"bfb11063-6ad6-463b-87e6-eaf9333e2d4a\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.294375 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8"] Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.295586 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.298829 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-http" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.299047 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-grpc" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.302317 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8"] Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.303204 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsxgg\" (UniqueName: \"kubernetes.io/projected/bfb11063-6ad6-463b-87e6-eaf9333e2d4a-kube-api-access-tsxgg\") pod \"logging-loki-distributor-76cc67bf56-7khhv\" (UID: \"bfb11063-6ad6-463b-87e6-eaf9333e2d4a\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.381181 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.382911 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg58g\" (UniqueName: \"kubernetes.io/projected/7fb68151-644a-4d8b-9274-18bb7073cc41-kube-api-access-gg58g\") pod \"logging-loki-query-frontend-84558f7c9f-tltn8\" (UID: \"7fb68151-644a-4d8b-9274-18bb7073cc41\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.383077 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/036dedff-92b9-4ad9-89b4-15dfb60e73c2-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-tv6v8\" (UID: \"036dedff-92b9-4ad9-89b4-15dfb60e73c2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.383137 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/7fb68151-644a-4d8b-9274-18bb7073cc41-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-tltn8\" (UID: \"7fb68151-644a-4d8b-9274-18bb7073cc41\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.383180 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/7fb68151-644a-4d8b-9274-18bb7073cc41-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-tltn8\" (UID: \"7fb68151-644a-4d8b-9274-18bb7073cc41\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.383224 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/036dedff-92b9-4ad9-89b4-15dfb60e73c2-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-tv6v8\" (UID: \"036dedff-92b9-4ad9-89b4-15dfb60e73c2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.383265 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/036dedff-92b9-4ad9-89b4-15dfb60e73c2-config\") pod \"logging-loki-querier-5895d59bb8-tv6v8\" (UID: \"036dedff-92b9-4ad9-89b4-15dfb60e73c2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.383309 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/036dedff-92b9-4ad9-89b4-15dfb60e73c2-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-tv6v8\" (UID: \"036dedff-92b9-4ad9-89b4-15dfb60e73c2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.383326 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/036dedff-92b9-4ad9-89b4-15dfb60e73c2-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-tv6v8\" (UID: \"036dedff-92b9-4ad9-89b4-15dfb60e73c2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.383358 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjrjk\" (UniqueName: \"kubernetes.io/projected/036dedff-92b9-4ad9-89b4-15dfb60e73c2-kube-api-access-qjrjk\") pod \"logging-loki-querier-5895d59bb8-tv6v8\" (UID: \"036dedff-92b9-4ad9-89b4-15dfb60e73c2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.383380 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7fb68151-644a-4d8b-9274-18bb7073cc41-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-tltn8\" (UID: \"7fb68151-644a-4d8b-9274-18bb7073cc41\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.383419 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb68151-644a-4d8b-9274-18bb7073cc41-config\") pod \"logging-loki-query-frontend-84558f7c9f-tltn8\" (UID: \"7fb68151-644a-4d8b-9274-18bb7073cc41\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.384317 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/036dedff-92b9-4ad9-89b4-15dfb60e73c2-config\") pod \"logging-loki-querier-5895d59bb8-tv6v8\" (UID: \"036dedff-92b9-4ad9-89b4-15dfb60e73c2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.384392 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/036dedff-92b9-4ad9-89b4-15dfb60e73c2-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-tv6v8\" (UID: \"036dedff-92b9-4ad9-89b4-15dfb60e73c2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.399237 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/036dedff-92b9-4ad9-89b4-15dfb60e73c2-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-tv6v8\" (UID: \"036dedff-92b9-4ad9-89b4-15dfb60e73c2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.399264 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/036dedff-92b9-4ad9-89b4-15dfb60e73c2-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-tv6v8\" (UID: \"036dedff-92b9-4ad9-89b4-15dfb60e73c2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.399805 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-5d9974794d-bbcbk"] Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.402059 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.404529 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/036dedff-92b9-4ad9-89b4-15dfb60e73c2-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-tv6v8\" (UID: \"036dedff-92b9-4ad9-89b4-15dfb60e73c2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.404886 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-client-http" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.405380 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway-ca-bundle" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.405428 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.405480 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-http" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.408750 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.428963 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-5d9974794d-bbcbk"] Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.430107 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjrjk\" (UniqueName: \"kubernetes.io/projected/036dedff-92b9-4ad9-89b4-15dfb60e73c2-kube-api-access-qjrjk\") pod \"logging-loki-querier-5895d59bb8-tv6v8\" (UID: \"036dedff-92b9-4ad9-89b4-15dfb60e73c2\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.455962 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-5d9974794d-hgkj4"] Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.457241 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.460029 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-dockercfg-mw87j" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.464189 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-5d9974794d-hgkj4"] Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.484522 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/261b6830-cc71-4704-ab0d-fd3a9708a786-tls-secret\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.484575 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/261b6830-cc71-4704-ab0d-fd3a9708a786-logging-loki-ca-bundle\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.484615 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/7fb68151-644a-4d8b-9274-18bb7073cc41-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-tltn8\" (UID: \"7fb68151-644a-4d8b-9274-18bb7073cc41\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.484696 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/7fb68151-644a-4d8b-9274-18bb7073cc41-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-tltn8\" (UID: \"7fb68151-644a-4d8b-9274-18bb7073cc41\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.485415 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/261b6830-cc71-4704-ab0d-fd3a9708a786-rbac\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.485480 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/261b6830-cc71-4704-ab0d-fd3a9708a786-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.486786 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7fb68151-644a-4d8b-9274-18bb7073cc41-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-tltn8\" (UID: \"7fb68151-644a-4d8b-9274-18bb7073cc41\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.486913 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/261b6830-cc71-4704-ab0d-fd3a9708a786-tenants\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.487875 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7fb68151-644a-4d8b-9274-18bb7073cc41-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-tltn8\" (UID: \"7fb68151-644a-4d8b-9274-18bb7073cc41\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.487939 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb68151-644a-4d8b-9274-18bb7073cc41-config\") pod \"logging-loki-query-frontend-84558f7c9f-tltn8\" (UID: \"7fb68151-644a-4d8b-9274-18bb7073cc41\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.488674 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb68151-644a-4d8b-9274-18bb7073cc41-config\") pod \"logging-loki-query-frontend-84558f7c9f-tltn8\" (UID: \"7fb68151-644a-4d8b-9274-18bb7073cc41\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.488722 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg58g\" (UniqueName: \"kubernetes.io/projected/7fb68151-644a-4d8b-9274-18bb7073cc41-kube-api-access-gg58g\") pod \"logging-loki-query-frontend-84558f7c9f-tltn8\" (UID: \"7fb68151-644a-4d8b-9274-18bb7073cc41\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.488808 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnvk7\" (UniqueName: \"kubernetes.io/projected/261b6830-cc71-4704-ab0d-fd3a9708a786-kube-api-access-gnvk7\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.488857 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/261b6830-cc71-4704-ab0d-fd3a9708a786-lokistack-gateway\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.489014 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/261b6830-cc71-4704-ab0d-fd3a9708a786-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.493370 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/7fb68151-644a-4d8b-9274-18bb7073cc41-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-tltn8\" (UID: \"7fb68151-644a-4d8b-9274-18bb7073cc41\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.493598 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/7fb68151-644a-4d8b-9274-18bb7073cc41-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-tltn8\" (UID: \"7fb68151-644a-4d8b-9274-18bb7073cc41\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.511849 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg58g\" (UniqueName: \"kubernetes.io/projected/7fb68151-644a-4d8b-9274-18bb7073cc41-kube-api-access-gg58g\") pod \"logging-loki-query-frontend-84558f7c9f-tltn8\" (UID: \"7fb68151-644a-4d8b-9274-18bb7073cc41\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.520094 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.590587 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/261b6830-cc71-4704-ab0d-fd3a9708a786-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.590640 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-tls-secret\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.590720 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-lokistack-gateway\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.590747 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-rbac\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.590768 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9brx\" (UniqueName: \"kubernetes.io/projected/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-kube-api-access-p9brx\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.590796 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/261b6830-cc71-4704-ab0d-fd3a9708a786-tenants\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.590827 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.590853 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnvk7\" (UniqueName: \"kubernetes.io/projected/261b6830-cc71-4704-ab0d-fd3a9708a786-kube-api-access-gnvk7\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.590875 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-tenants\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.590897 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/261b6830-cc71-4704-ab0d-fd3a9708a786-lokistack-gateway\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.590915 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-logging-loki-ca-bundle\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.590932 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/261b6830-cc71-4704-ab0d-fd3a9708a786-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.590949 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.590976 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/261b6830-cc71-4704-ab0d-fd3a9708a786-tls-secret\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.590994 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/261b6830-cc71-4704-ab0d-fd3a9708a786-logging-loki-ca-bundle\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.591018 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/261b6830-cc71-4704-ab0d-fd3a9708a786-rbac\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: E1126 00:36:49.594187 4766 secret.go:188] Couldn't get secret openshift-logging/logging-loki-gateway-http: secret "logging-loki-gateway-http" not found Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.594411 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/261b6830-cc71-4704-ab0d-fd3a9708a786-rbac\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: E1126 00:36:49.594562 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/261b6830-cc71-4704-ab0d-fd3a9708a786-tls-secret podName:261b6830-cc71-4704-ab0d-fd3a9708a786 nodeName:}" failed. No retries permitted until 2025-11-26 00:36:50.094534244 +0000 UTC m=+790.943304674 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/261b6830-cc71-4704-ab0d-fd3a9708a786-tls-secret") pod "logging-loki-gateway-5d9974794d-bbcbk" (UID: "261b6830-cc71-4704-ab0d-fd3a9708a786") : secret "logging-loki-gateway-http" not found Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.594931 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/261b6830-cc71-4704-ab0d-fd3a9708a786-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.595033 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/261b6830-cc71-4704-ab0d-fd3a9708a786-lokistack-gateway\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.595209 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/261b6830-cc71-4704-ab0d-fd3a9708a786-logging-loki-ca-bundle\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.595907 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/261b6830-cc71-4704-ab0d-fd3a9708a786-tenants\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.600763 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/261b6830-cc71-4704-ab0d-fd3a9708a786-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.611636 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnvk7\" (UniqueName: \"kubernetes.io/projected/261b6830-cc71-4704-ab0d-fd3a9708a786-kube-api-access-gnvk7\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.651029 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.692244 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-tenants\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.692583 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-logging-loki-ca-bundle\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.692603 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.692731 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-tls-secret\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.692768 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-lokistack-gateway\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.693906 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.695542 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-logging-loki-ca-bundle\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.692796 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-rbac\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.695631 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9brx\" (UniqueName: \"kubernetes.io/projected/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-kube-api-access-p9brx\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.695792 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-lokistack-gateway\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.696713 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-tenants\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.697243 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-rbac\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.697428 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-tls-secret\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.698362 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.703830 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.720011 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9brx\" (UniqueName: \"kubernetes.io/projected/bc3b1ab0-39c8-42a1-b0d1-c45baa43df87-kube-api-access-p9brx\") pod \"logging-loki-gateway-5d9974794d-hgkj4\" (UID: \"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.794639 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.865743 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-7khhv"] Nov 26 00:36:49 crc kubenswrapper[4766]: I1126 00:36:49.974355 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-tv6v8"] Nov 26 00:36:49 crc kubenswrapper[4766]: W1126 00:36:49.984170 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod036dedff_92b9_4ad9_89b4_15dfb60e73c2.slice/crio-c3f60e0e0ebb6adb4c2844d0047e86596624690cb2b030cdb2da42c3af3ae230 WatchSource:0}: Error finding container c3f60e0e0ebb6adb4c2844d0047e86596624690cb2b030cdb2da42c3af3ae230: Status 404 returned error can't find the container with id c3f60e0e0ebb6adb4c2844d0047e86596624690cb2b030cdb2da42c3af3ae230 Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.097090 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8"] Nov 26 00:36:50 crc kubenswrapper[4766]: W1126 00:36:50.102859 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fb68151_644a_4d8b_9274_18bb7073cc41.slice/crio-8d75f0bf16ee3b66e62dd43b5ad373ba1850bc9411f60e23301c8f57584d9add WatchSource:0}: Error finding container 8d75f0bf16ee3b66e62dd43b5ad373ba1850bc9411f60e23301c8f57584d9add: Status 404 returned error can't find the container with id 8d75f0bf16ee3b66e62dd43b5ad373ba1850bc9411f60e23301c8f57584d9add Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.113337 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/261b6830-cc71-4704-ab0d-fd3a9708a786-tls-secret\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.118776 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/261b6830-cc71-4704-ab0d-fd3a9708a786-tls-secret\") pod \"logging-loki-gateway-5d9974794d-bbcbk\" (UID: \"261b6830-cc71-4704-ab0d-fd3a9708a786\") " pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.213679 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-5d9974794d-hgkj4"] Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.222550 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.223486 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.227422 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-http" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.231397 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-grpc" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.234009 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.263034 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.264726 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.271562 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-grpc" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.283877 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-http" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.287374 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.323450 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b4c0046a-4a99-4483-8e41-bfa6361fc5e6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b4c0046a-4a99-4483-8e41-bfa6361fc5e6\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.323521 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/aae58301-788d-430a-b020-a5f530d70c5a-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.323681 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-02a17d3c-dbc5-4595-8fd7-595a5542b402\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-02a17d3c-dbc5-4595-8fd7-595a5542b402\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.324069 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/aae58301-788d-430a-b020-a5f530d70c5a-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.324113 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aae58301-788d-430a-b020-a5f530d70c5a-config\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.324163 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/aae58301-788d-430a-b020-a5f530d70c5a-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.324208 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2crxn\" (UniqueName: \"kubernetes.io/projected/aae58301-788d-430a-b020-a5f530d70c5a-kube-api-access-2crxn\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.324254 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aae58301-788d-430a-b020-a5f530d70c5a-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.373032 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.374117 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.379133 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-http" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.379453 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-grpc" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.384120 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.396878 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.426555 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2crxn\" (UniqueName: \"kubernetes.io/projected/aae58301-788d-430a-b020-a5f530d70c5a-kube-api-access-2crxn\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.426635 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aae58301-788d-430a-b020-a5f530d70c5a-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.426688 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj6xt\" (UniqueName: \"kubernetes.io/projected/4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e-kube-api-access-vj6xt\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") " pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.426734 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") " pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.426773 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e-config\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") " pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.426810 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b4c0046a-4a99-4483-8e41-bfa6361fc5e6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b4c0046a-4a99-4483-8e41-bfa6361fc5e6\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.426838 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/aae58301-788d-430a-b020-a5f530d70c5a-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.426873 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-02a17d3c-dbc5-4595-8fd7-595a5542b402\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-02a17d3c-dbc5-4595-8fd7-595a5542b402\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.426901 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") " pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.426932 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") " pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.426962 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") " pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.426998 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7aefef02-a7e9-4edc-a373-3032c1ae6935\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7aefef02-a7e9-4edc-a373-3032c1ae6935\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") " pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.427029 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/aae58301-788d-430a-b020-a5f530d70c5a-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.427057 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aae58301-788d-430a-b020-a5f530d70c5a-config\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.427105 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/aae58301-788d-430a-b020-a5f530d70c5a-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.432853 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aae58301-788d-430a-b020-a5f530d70c5a-config\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.436930 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aae58301-788d-430a-b020-a5f530d70c5a-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.441915 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/aae58301-788d-430a-b020-a5f530d70c5a-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.450815 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/aae58301-788d-430a-b020-a5f530d70c5a-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.459206 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/aae58301-788d-430a-b020-a5f530d70c5a-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.465690 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2crxn\" (UniqueName: \"kubernetes.io/projected/aae58301-788d-430a-b020-a5f530d70c5a-kube-api-access-2crxn\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.472463 4766 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.472536 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b4c0046a-4a99-4483-8e41-bfa6361fc5e6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b4c0046a-4a99-4483-8e41-bfa6361fc5e6\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2533cb7d0c7f926b72872f9c4d80b7913a2a3ced664c1b1052820ca3a2f7c996/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.472668 4766 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.472735 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-02a17d3c-dbc5-4595-8fd7-595a5542b402\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-02a17d3c-dbc5-4595-8fd7-595a5542b402\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8678a28ec324e40a7c3776b6f21be79f0c9213fbf5419462f49468407ebf7171/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.507636 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b4c0046a-4a99-4483-8e41-bfa6361fc5e6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b4c0046a-4a99-4483-8e41-bfa6361fc5e6\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.510937 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-02a17d3c-dbc5-4595-8fd7-595a5542b402\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-02a17d3c-dbc5-4595-8fd7-595a5542b402\") pod \"logging-loki-ingester-0\" (UID: \"aae58301-788d-430a-b020-a5f530d70c5a\") " pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.531123 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/87bee09b-03e7-4b6c-a28f-7e12aab1eb45-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.531210 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/87bee09b-03e7-4b6c-a28f-7e12aab1eb45-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.531256 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj6xt\" (UniqueName: \"kubernetes.io/projected/4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e-kube-api-access-vj6xt\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") " pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.531290 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87bee09b-03e7-4b6c-a28f-7e12aab1eb45-config\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.531332 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1df9e0de-4b66-416b-8c4a-81294543cb3b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1df9e0de-4b66-416b-8c4a-81294543cb3b\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.531369 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") " pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.531413 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e-config\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") " pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.531456 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knvs6\" (UniqueName: \"kubernetes.io/projected/87bee09b-03e7-4b6c-a28f-7e12aab1eb45-kube-api-access-knvs6\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.531501 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") " pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.531544 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") " pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.531586 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") " pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.531640 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7aefef02-a7e9-4edc-a373-3032c1ae6935\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7aefef02-a7e9-4edc-a373-3032c1ae6935\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") " pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.531691 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/87bee09b-03e7-4b6c-a28f-7e12aab1eb45-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.531739 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/87bee09b-03e7-4b6c-a28f-7e12aab1eb45-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.536742 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") " pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.537546 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e-config\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") " pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.542375 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") " pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.543044 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.547259 4766 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.547296 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7aefef02-a7e9-4edc-a373-3032c1ae6935\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7aefef02-a7e9-4edc-a373-3032c1ae6935\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f90fa1931e04a506b659831363abe248bdbb7d9ab6a38cecc12cd10c779cb1a2/globalmount\"" pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.549387 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") " pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.552692 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj6xt\" (UniqueName: \"kubernetes.io/projected/4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e-kube-api-access-vj6xt\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") " pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.554848 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") " pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.581603 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7aefef02-a7e9-4edc-a373-3032c1ae6935\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7aefef02-a7e9-4edc-a373-3032c1ae6935\") pod \"logging-loki-compactor-0\" (UID: \"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e\") " pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.634324 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/87bee09b-03e7-4b6c-a28f-7e12aab1eb45-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.634400 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/87bee09b-03e7-4b6c-a28f-7e12aab1eb45-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.634431 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/87bee09b-03e7-4b6c-a28f-7e12aab1eb45-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.634458 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87bee09b-03e7-4b6c-a28f-7e12aab1eb45-config\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.634489 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1df9e0de-4b66-416b-8c4a-81294543cb3b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1df9e0de-4b66-416b-8c4a-81294543cb3b\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.634539 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knvs6\" (UniqueName: \"kubernetes.io/projected/87bee09b-03e7-4b6c-a28f-7e12aab1eb45-kube-api-access-knvs6\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.634600 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/87bee09b-03e7-4b6c-a28f-7e12aab1eb45-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.635246 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/87bee09b-03e7-4b6c-a28f-7e12aab1eb45-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.636597 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87bee09b-03e7-4b6c-a28f-7e12aab1eb45-config\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.638338 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/87bee09b-03e7-4b6c-a28f-7e12aab1eb45-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.640121 4766 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.640150 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1df9e0de-4b66-416b-8c4a-81294543cb3b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1df9e0de-4b66-416b-8c4a-81294543cb3b\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/821062be4f47c1be465c8b3696e7d81a91f8849f1fda18aa000c494ecd136bc1/globalmount\"" pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.641332 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/87bee09b-03e7-4b6c-a28f-7e12aab1eb45-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.643302 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/87bee09b-03e7-4b6c-a28f-7e12aab1eb45-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.653123 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" event={"ID":"7fb68151-644a-4d8b-9274-18bb7073cc41","Type":"ContainerStarted","Data":"8d75f0bf16ee3b66e62dd43b5ad373ba1850bc9411f60e23301c8f57584d9add"} Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.654471 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" event={"ID":"036dedff-92b9-4ad9-89b4-15dfb60e73c2","Type":"ContainerStarted","Data":"c3f60e0e0ebb6adb4c2844d0047e86596624690cb2b030cdb2da42c3af3ae230"} Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.657034 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" event={"ID":"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87","Type":"ContainerStarted","Data":"a4dc480d19ce4e7266360bd87cf6fa6c3d6018928485933dd1d44af07026ae58"} Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.658366 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" event={"ID":"bfb11063-6ad6-463b-87e6-eaf9333e2d4a","Type":"ContainerStarted","Data":"c8467ef87f1a11c7ae84ea677b2cfda154307a8928f5f94cb25360151ca071d7"} Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.659495 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knvs6\" (UniqueName: \"kubernetes.io/projected/87bee09b-03e7-4b6c-a28f-7e12aab1eb45-kube-api-access-knvs6\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.673706 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1df9e0de-4b66-416b-8c4a-81294543cb3b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1df9e0de-4b66-416b-8c4a-81294543cb3b\") pod \"logging-loki-index-gateway-0\" (UID: \"87bee09b-03e7-4b6c-a28f-7e12aab1eb45\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.697221 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.859972 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-5d9974794d-bbcbk"] Nov 26 00:36:50 crc kubenswrapper[4766]: W1126 00:36:50.863869 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod261b6830_cc71_4704_ab0d_fd3a9708a786.slice/crio-4c791fc7eacafe2b9b7f0af13a6bcdc171d38b183303a3f6f2913b1ebbef905e WatchSource:0}: Error finding container 4c791fc7eacafe2b9b7f0af13a6bcdc171d38b183303a3f6f2913b1ebbef905e: Status 404 returned error can't find the container with id 4c791fc7eacafe2b9b7f0af13a6bcdc171d38b183303a3f6f2913b1ebbef905e Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.879262 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:50 crc kubenswrapper[4766]: I1126 00:36:50.970570 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Nov 26 00:36:51 crc kubenswrapper[4766]: I1126 00:36:51.117333 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Nov 26 00:36:51 crc kubenswrapper[4766]: W1126 00:36:51.118560 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87bee09b_03e7_4b6c_a28f_7e12aab1eb45.slice/crio-41e2788fb4f0754ee72c057ae233b39d369bf938ae3828f0d334d082ef9d1168 WatchSource:0}: Error finding container 41e2788fb4f0754ee72c057ae233b39d369bf938ae3828f0d334d082ef9d1168: Status 404 returned error can't find the container with id 41e2788fb4f0754ee72c057ae233b39d369bf938ae3828f0d334d082ef9d1168 Nov 26 00:36:51 crc kubenswrapper[4766]: I1126 00:36:51.307958 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Nov 26 00:36:51 crc kubenswrapper[4766]: W1126 00:36:51.352867 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a6f0b48_aeb6_4df1_ba96_6fcc9c31b08e.slice/crio-caa734172e598cffeb62b8eb7814d133fb495ac2af877ea4358108b310cd56c5 WatchSource:0}: Error finding container caa734172e598cffeb62b8eb7814d133fb495ac2af877ea4358108b310cd56c5: Status 404 returned error can't find the container with id caa734172e598cffeb62b8eb7814d133fb495ac2af877ea4358108b310cd56c5 Nov 26 00:36:51 crc kubenswrapper[4766]: I1126 00:36:51.675545 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"87bee09b-03e7-4b6c-a28f-7e12aab1eb45","Type":"ContainerStarted","Data":"41e2788fb4f0754ee72c057ae233b39d369bf938ae3828f0d334d082ef9d1168"} Nov 26 00:36:51 crc kubenswrapper[4766]: I1126 00:36:51.677814 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"aae58301-788d-430a-b020-a5f530d70c5a","Type":"ContainerStarted","Data":"932166d819d97f86ab0b4b75fedf89022970a73585c77d9f4b802e5622951ab7"} Nov 26 00:36:51 crc kubenswrapper[4766]: I1126 00:36:51.681871 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e","Type":"ContainerStarted","Data":"caa734172e598cffeb62b8eb7814d133fb495ac2af877ea4358108b310cd56c5"} Nov 26 00:36:51 crc kubenswrapper[4766]: I1126 00:36:51.683319 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" event={"ID":"261b6830-cc71-4704-ab0d-fd3a9708a786","Type":"ContainerStarted","Data":"4c791fc7eacafe2b9b7f0af13a6bcdc171d38b183303a3f6f2913b1ebbef905e"} Nov 26 00:36:54 crc kubenswrapper[4766]: I1126 00:36:54.702340 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" event={"ID":"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87","Type":"ContainerStarted","Data":"069fba2de375d7331edd64e522acd51de0a863ceeda84af745ce3ff179a23c93"} Nov 26 00:36:54 crc kubenswrapper[4766]: I1126 00:36:54.703678 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" event={"ID":"bfb11063-6ad6-463b-87e6-eaf9333e2d4a","Type":"ContainerStarted","Data":"84a58e3f12339b3e7c30f95daa199f2a1fc87d2d548cb3824eba026ee8c524d0"} Nov 26 00:36:54 crc kubenswrapper[4766]: I1126 00:36:54.703813 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" Nov 26 00:36:54 crc kubenswrapper[4766]: I1126 00:36:54.705007 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"87bee09b-03e7-4b6c-a28f-7e12aab1eb45","Type":"ContainerStarted","Data":"7348f64a8efd038db48b137f45dfd09d5f8bf966927a92e064efeecb540d9a85"} Nov 26 00:36:54 crc kubenswrapper[4766]: I1126 00:36:54.705273 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:36:54 crc kubenswrapper[4766]: I1126 00:36:54.707831 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"aae58301-788d-430a-b020-a5f530d70c5a","Type":"ContainerStarted","Data":"c4ac145d063d46e687394ac35de954db415992ab49498e026a7c634bddfd8bcf"} Nov 26 00:36:54 crc kubenswrapper[4766]: I1126 00:36:54.707940 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:36:54 crc kubenswrapper[4766]: I1126 00:36:54.709111 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" event={"ID":"7fb68151-644a-4d8b-9274-18bb7073cc41","Type":"ContainerStarted","Data":"bb175ecbdfeee6e27043b8b4185a7f3fe4b9370e55b344ad3e4b736dc62b4475"} Nov 26 00:36:54 crc kubenswrapper[4766]: I1126 00:36:54.709644 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" Nov 26 00:36:54 crc kubenswrapper[4766]: I1126 00:36:54.710981 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e","Type":"ContainerStarted","Data":"9802ab49008772bbd46ce6ee132eeb86d6716d8dd04e80dfdaabcb0b9db8e55e"} Nov 26 00:36:54 crc kubenswrapper[4766]: I1126 00:36:54.711422 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:36:54 crc kubenswrapper[4766]: I1126 00:36:54.712523 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" event={"ID":"261b6830-cc71-4704-ab0d-fd3a9708a786","Type":"ContainerStarted","Data":"218ac982cd2937ac26de9aae8cc8b3281b72f2d6362c8a1935ec1a080c7951ae"} Nov 26 00:36:54 crc kubenswrapper[4766]: I1126 00:36:54.713643 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" event={"ID":"036dedff-92b9-4ad9-89b4-15dfb60e73c2","Type":"ContainerStarted","Data":"ba11ff2e97990d3dd59f2b88129d1267b75d1eb30a8b7106adff8abb9e6b1c7b"} Nov 26 00:36:54 crc kubenswrapper[4766]: I1126 00:36:54.713807 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:36:54 crc kubenswrapper[4766]: I1126 00:36:54.724017 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" podStartSLOduration=1.796579022 podStartE2EDuration="5.723999173s" podCreationTimestamp="2025-11-26 00:36:49 +0000 UTC" firstStartedPulling="2025-11-26 00:36:49.877853257 +0000 UTC m=+790.726623687" lastFinishedPulling="2025-11-26 00:36:53.805273378 +0000 UTC m=+794.654043838" observedRunningTime="2025-11-26 00:36:54.719014917 +0000 UTC m=+795.567785347" watchObservedRunningTime="2025-11-26 00:36:54.723999173 +0000 UTC m=+795.572769603" Nov 26 00:36:54 crc kubenswrapper[4766]: I1126 00:36:54.744853 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" podStartSLOduration=1.985218803 podStartE2EDuration="5.744828518s" podCreationTimestamp="2025-11-26 00:36:49 +0000 UTC" firstStartedPulling="2025-11-26 00:36:50.105268458 +0000 UTC m=+790.954038888" lastFinishedPulling="2025-11-26 00:36:53.864878173 +0000 UTC m=+794.713648603" observedRunningTime="2025-11-26 00:36:54.73814922 +0000 UTC m=+795.586919650" watchObservedRunningTime="2025-11-26 00:36:54.744828518 +0000 UTC m=+795.593598948" Nov 26 00:36:54 crc kubenswrapper[4766]: I1126 00:36:54.762507 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-compactor-0" podStartSLOduration=3.277789436 podStartE2EDuration="5.762484314s" podCreationTimestamp="2025-11-26 00:36:49 +0000 UTC" firstStartedPulling="2025-11-26 00:36:51.358563759 +0000 UTC m=+792.207334189" lastFinishedPulling="2025-11-26 00:36:53.843258647 +0000 UTC m=+794.692029067" observedRunningTime="2025-11-26 00:36:54.759037357 +0000 UTC m=+795.607807797" watchObservedRunningTime="2025-11-26 00:36:54.762484314 +0000 UTC m=+795.611254744" Nov 26 00:36:54 crc kubenswrapper[4766]: I1126 00:36:54.779915 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-ingester-0" podStartSLOduration=2.938402268 podStartE2EDuration="5.779895644s" podCreationTimestamp="2025-11-26 00:36:49 +0000 UTC" firstStartedPulling="2025-11-26 00:36:50.981728686 +0000 UTC m=+791.830499116" lastFinishedPulling="2025-11-26 00:36:53.823222062 +0000 UTC m=+794.671992492" observedRunningTime="2025-11-26 00:36:54.775480502 +0000 UTC m=+795.624250942" watchObservedRunningTime="2025-11-26 00:36:54.779895644 +0000 UTC m=+795.628666084" Nov 26 00:36:54 crc kubenswrapper[4766]: I1126 00:36:54.796170 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" podStartSLOduration=1.9392289919999999 podStartE2EDuration="5.796151514s" podCreationTimestamp="2025-11-26 00:36:49 +0000 UTC" firstStartedPulling="2025-11-26 00:36:49.98889077 +0000 UTC m=+790.837661210" lastFinishedPulling="2025-11-26 00:36:53.845813302 +0000 UTC m=+794.694583732" observedRunningTime="2025-11-26 00:36:54.792982234 +0000 UTC m=+795.641752684" watchObservedRunningTime="2025-11-26 00:36:54.796151514 +0000 UTC m=+795.644921944" Nov 26 00:36:54 crc kubenswrapper[4766]: I1126 00:36:54.819600 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-index-gateway-0" podStartSLOduration=3.108967705 podStartE2EDuration="5.819580076s" podCreationTimestamp="2025-11-26 00:36:49 +0000 UTC" firstStartedPulling="2025-11-26 00:36:51.120763106 +0000 UTC m=+791.969533536" lastFinishedPulling="2025-11-26 00:36:53.831375477 +0000 UTC m=+794.680145907" observedRunningTime="2025-11-26 00:36:54.815566144 +0000 UTC m=+795.664336594" watchObservedRunningTime="2025-11-26 00:36:54.819580076 +0000 UTC m=+795.668350526" Nov 26 00:36:56 crc kubenswrapper[4766]: I1126 00:36:56.728986 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" event={"ID":"bc3b1ab0-39c8-42a1-b0d1-c45baa43df87","Type":"ContainerStarted","Data":"18c6035d08d3f9edfddd822a04e66117d670600a4c2e39a97dd1981dd61c127f"} Nov 26 00:36:56 crc kubenswrapper[4766]: I1126 00:36:56.729550 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:56 crc kubenswrapper[4766]: I1126 00:36:56.729565 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:56 crc kubenswrapper[4766]: I1126 00:36:56.732573 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" event={"ID":"261b6830-cc71-4704-ab0d-fd3a9708a786","Type":"ContainerStarted","Data":"be8d63b0ef21613c4232f374b19b58bbd7f680c60e40972cf094bc495005d883"} Nov 26 00:36:56 crc kubenswrapper[4766]: I1126 00:36:56.733282 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:56 crc kubenswrapper[4766]: I1126 00:36:56.733314 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:56 crc kubenswrapper[4766]: I1126 00:36:56.741347 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:56 crc kubenswrapper[4766]: I1126 00:36:56.745007 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:56 crc kubenswrapper[4766]: I1126 00:36:56.749623 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" Nov 26 00:36:56 crc kubenswrapper[4766]: I1126 00:36:56.750386 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" podStartSLOduration=1.444091152 podStartE2EDuration="7.75036646s" podCreationTimestamp="2025-11-26 00:36:49 +0000 UTC" firstStartedPulling="2025-11-26 00:36:50.224693123 +0000 UTC m=+791.073463553" lastFinishedPulling="2025-11-26 00:36:56.530968431 +0000 UTC m=+797.379738861" observedRunningTime="2025-11-26 00:36:56.749171179 +0000 UTC m=+797.597941609" watchObservedRunningTime="2025-11-26 00:36:56.75036646 +0000 UTC m=+797.599136890" Nov 26 00:36:56 crc kubenswrapper[4766]: I1126 00:36:56.761524 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-5d9974794d-hgkj4" Nov 26 00:36:56 crc kubenswrapper[4766]: I1126 00:36:56.772607 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-5d9974794d-bbcbk" podStartSLOduration=2.103570962 podStartE2EDuration="7.772586341s" podCreationTimestamp="2025-11-26 00:36:49 +0000 UTC" firstStartedPulling="2025-11-26 00:36:50.865903951 +0000 UTC m=+791.714674381" lastFinishedPulling="2025-11-26 00:36:56.53491933 +0000 UTC m=+797.383689760" observedRunningTime="2025-11-26 00:36:56.770387995 +0000 UTC m=+797.619158425" watchObservedRunningTime="2025-11-26 00:36:56.772586341 +0000 UTC m=+797.621356761" Nov 26 00:37:05 crc kubenswrapper[4766]: I1126 00:37:05.638391 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qxfsx"] Nov 26 00:37:05 crc kubenswrapper[4766]: I1126 00:37:05.640474 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qxfsx" Nov 26 00:37:05 crc kubenswrapper[4766]: I1126 00:37:05.649578 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qxfsx"] Nov 26 00:37:05 crc kubenswrapper[4766]: I1126 00:37:05.753231 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6807f6c2-7753-4dd5-a273-e173de1ca49f-utilities\") pod \"community-operators-qxfsx\" (UID: \"6807f6c2-7753-4dd5-a273-e173de1ca49f\") " pod="openshift-marketplace/community-operators-qxfsx" Nov 26 00:37:05 crc kubenswrapper[4766]: I1126 00:37:05.753295 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6807f6c2-7753-4dd5-a273-e173de1ca49f-catalog-content\") pod \"community-operators-qxfsx\" (UID: \"6807f6c2-7753-4dd5-a273-e173de1ca49f\") " pod="openshift-marketplace/community-operators-qxfsx" Nov 26 00:37:05 crc kubenswrapper[4766]: I1126 00:37:05.753668 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24klq\" (UniqueName: \"kubernetes.io/projected/6807f6c2-7753-4dd5-a273-e173de1ca49f-kube-api-access-24klq\") pod \"community-operators-qxfsx\" (UID: \"6807f6c2-7753-4dd5-a273-e173de1ca49f\") " pod="openshift-marketplace/community-operators-qxfsx" Nov 26 00:37:05 crc kubenswrapper[4766]: I1126 00:37:05.855288 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24klq\" (UniqueName: \"kubernetes.io/projected/6807f6c2-7753-4dd5-a273-e173de1ca49f-kube-api-access-24klq\") pod \"community-operators-qxfsx\" (UID: \"6807f6c2-7753-4dd5-a273-e173de1ca49f\") " pod="openshift-marketplace/community-operators-qxfsx" Nov 26 00:37:05 crc kubenswrapper[4766]: I1126 00:37:05.855364 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6807f6c2-7753-4dd5-a273-e173de1ca49f-utilities\") pod \"community-operators-qxfsx\" (UID: \"6807f6c2-7753-4dd5-a273-e173de1ca49f\") " pod="openshift-marketplace/community-operators-qxfsx" Nov 26 00:37:05 crc kubenswrapper[4766]: I1126 00:37:05.855396 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6807f6c2-7753-4dd5-a273-e173de1ca49f-catalog-content\") pod \"community-operators-qxfsx\" (UID: \"6807f6c2-7753-4dd5-a273-e173de1ca49f\") " pod="openshift-marketplace/community-operators-qxfsx" Nov 26 00:37:05 crc kubenswrapper[4766]: I1126 00:37:05.856225 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6807f6c2-7753-4dd5-a273-e173de1ca49f-catalog-content\") pod \"community-operators-qxfsx\" (UID: \"6807f6c2-7753-4dd5-a273-e173de1ca49f\") " pod="openshift-marketplace/community-operators-qxfsx" Nov 26 00:37:05 crc kubenswrapper[4766]: I1126 00:37:05.856552 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6807f6c2-7753-4dd5-a273-e173de1ca49f-utilities\") pod \"community-operators-qxfsx\" (UID: \"6807f6c2-7753-4dd5-a273-e173de1ca49f\") " pod="openshift-marketplace/community-operators-qxfsx" Nov 26 00:37:05 crc kubenswrapper[4766]: I1126 00:37:05.881555 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24klq\" (UniqueName: \"kubernetes.io/projected/6807f6c2-7753-4dd5-a273-e173de1ca49f-kube-api-access-24klq\") pod \"community-operators-qxfsx\" (UID: \"6807f6c2-7753-4dd5-a273-e173de1ca49f\") " pod="openshift-marketplace/community-operators-qxfsx" Nov 26 00:37:05 crc kubenswrapper[4766]: I1126 00:37:05.956506 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qxfsx" Nov 26 00:37:06 crc kubenswrapper[4766]: I1126 00:37:06.776117 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qxfsx"] Nov 26 00:37:06 crc kubenswrapper[4766]: W1126 00:37:06.784825 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6807f6c2_7753_4dd5_a273_e173de1ca49f.slice/crio-32920355fd7a12f17fba7cc00972db05675546a363d1d79c8e16fd2887822ed6 WatchSource:0}: Error finding container 32920355fd7a12f17fba7cc00972db05675546a363d1d79c8e16fd2887822ed6: Status 404 returned error can't find the container with id 32920355fd7a12f17fba7cc00972db05675546a363d1d79c8e16fd2887822ed6 Nov 26 00:37:07 crc kubenswrapper[4766]: I1126 00:37:07.798535 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxfsx" event={"ID":"6807f6c2-7753-4dd5-a273-e173de1ca49f","Type":"ContainerStarted","Data":"32920355fd7a12f17fba7cc00972db05675546a363d1d79c8e16fd2887822ed6"} Nov 26 00:37:09 crc kubenswrapper[4766]: I1126 00:37:09.389539 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-76cc67bf56-7khhv" Nov 26 00:37:09 crc kubenswrapper[4766]: I1126 00:37:09.526348 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-5895d59bb8-tv6v8" Nov 26 00:37:09 crc kubenswrapper[4766]: I1126 00:37:09.657307 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-tltn8" Nov 26 00:37:09 crc kubenswrapper[4766]: I1126 00:37:09.811854 4766 generic.go:334] "Generic (PLEG): container finished" podID="6807f6c2-7753-4dd5-a273-e173de1ca49f" containerID="6d5a8ca8e32aed732b731d442655db61148e6035737dbe9ff91dbddb2c8ae52c" exitCode=0 Nov 26 00:37:09 crc kubenswrapper[4766]: I1126 00:37:09.811893 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxfsx" event={"ID":"6807f6c2-7753-4dd5-a273-e173de1ca49f","Type":"ContainerDied","Data":"6d5a8ca8e32aed732b731d442655db61148e6035737dbe9ff91dbddb2c8ae52c"} Nov 26 00:37:10 crc kubenswrapper[4766]: I1126 00:37:10.597668 4766 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Nov 26 00:37:10 crc kubenswrapper[4766]: I1126 00:37:10.598630 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="aae58301-788d-430a-b020-a5f530d70c5a" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 26 00:37:10 crc kubenswrapper[4766]: I1126 00:37:10.703189 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Nov 26 00:37:10 crc kubenswrapper[4766]: I1126 00:37:10.819125 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxfsx" event={"ID":"6807f6c2-7753-4dd5-a273-e173de1ca49f","Type":"ContainerStarted","Data":"c5a55ca712f70c12c3ebda5816b2f47564c981539b360d75476871028e651881"} Nov 26 00:37:10 crc kubenswrapper[4766]: I1126 00:37:10.884062 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Nov 26 00:37:11 crc kubenswrapper[4766]: I1126 00:37:11.827120 4766 generic.go:334] "Generic (PLEG): container finished" podID="6807f6c2-7753-4dd5-a273-e173de1ca49f" containerID="c5a55ca712f70c12c3ebda5816b2f47564c981539b360d75476871028e651881" exitCode=0 Nov 26 00:37:11 crc kubenswrapper[4766]: I1126 00:37:11.838296 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxfsx" event={"ID":"6807f6c2-7753-4dd5-a273-e173de1ca49f","Type":"ContainerDied","Data":"c5a55ca712f70c12c3ebda5816b2f47564c981539b360d75476871028e651881"} Nov 26 00:37:12 crc kubenswrapper[4766]: I1126 00:37:12.853240 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxfsx" event={"ID":"6807f6c2-7753-4dd5-a273-e173de1ca49f","Type":"ContainerStarted","Data":"d660717ee5aacb1cc18ed7339b44375f045055deae8337669ec46e326014b5bb"} Nov 26 00:37:12 crc kubenswrapper[4766]: I1126 00:37:12.888148 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qxfsx" podStartSLOduration=5.454883413 podStartE2EDuration="7.888132653s" podCreationTimestamp="2025-11-26 00:37:05 +0000 UTC" firstStartedPulling="2025-11-26 00:37:09.813123341 +0000 UTC m=+810.661893771" lastFinishedPulling="2025-11-26 00:37:12.246372581 +0000 UTC m=+813.095143011" observedRunningTime="2025-11-26 00:37:12.885898307 +0000 UTC m=+813.734668747" watchObservedRunningTime="2025-11-26 00:37:12.888132653 +0000 UTC m=+813.736903083" Nov 26 00:37:15 crc kubenswrapper[4766]: I1126 00:37:15.956759 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qxfsx" Nov 26 00:37:15 crc kubenswrapper[4766]: I1126 00:37:15.958138 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qxfsx" Nov 26 00:37:16 crc kubenswrapper[4766]: I1126 00:37:16.011897 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qxfsx" Nov 26 00:37:20 crc kubenswrapper[4766]: I1126 00:37:20.549208 4766 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Nov 26 00:37:20 crc kubenswrapper[4766]: I1126 00:37:20.549532 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="aae58301-788d-430a-b020-a5f530d70c5a" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 26 00:37:26 crc kubenswrapper[4766]: I1126 00:37:26.006967 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qxfsx" Nov 26 00:37:26 crc kubenswrapper[4766]: I1126 00:37:26.056428 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qxfsx"] Nov 26 00:37:26 crc kubenswrapper[4766]: I1126 00:37:26.950903 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qxfsx" podUID="6807f6c2-7753-4dd5-a273-e173de1ca49f" containerName="registry-server" containerID="cri-o://d660717ee5aacb1cc18ed7339b44375f045055deae8337669ec46e326014b5bb" gracePeriod=2 Nov 26 00:37:27 crc kubenswrapper[4766]: I1126 00:37:27.339947 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qxfsx" Nov 26 00:37:27 crc kubenswrapper[4766]: I1126 00:37:27.456798 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6807f6c2-7753-4dd5-a273-e173de1ca49f-catalog-content\") pod \"6807f6c2-7753-4dd5-a273-e173de1ca49f\" (UID: \"6807f6c2-7753-4dd5-a273-e173de1ca49f\") " Nov 26 00:37:27 crc kubenswrapper[4766]: I1126 00:37:27.456917 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24klq\" (UniqueName: \"kubernetes.io/projected/6807f6c2-7753-4dd5-a273-e173de1ca49f-kube-api-access-24klq\") pod \"6807f6c2-7753-4dd5-a273-e173de1ca49f\" (UID: \"6807f6c2-7753-4dd5-a273-e173de1ca49f\") " Nov 26 00:37:27 crc kubenswrapper[4766]: I1126 00:37:27.458093 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6807f6c2-7753-4dd5-a273-e173de1ca49f-utilities\") pod \"6807f6c2-7753-4dd5-a273-e173de1ca49f\" (UID: \"6807f6c2-7753-4dd5-a273-e173de1ca49f\") " Nov 26 00:37:27 crc kubenswrapper[4766]: I1126 00:37:27.458737 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6807f6c2-7753-4dd5-a273-e173de1ca49f-utilities" (OuterVolumeSpecName: "utilities") pod "6807f6c2-7753-4dd5-a273-e173de1ca49f" (UID: "6807f6c2-7753-4dd5-a273-e173de1ca49f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:37:27 crc kubenswrapper[4766]: I1126 00:37:27.462413 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6807f6c2-7753-4dd5-a273-e173de1ca49f-kube-api-access-24klq" (OuterVolumeSpecName: "kube-api-access-24klq") pod "6807f6c2-7753-4dd5-a273-e173de1ca49f" (UID: "6807f6c2-7753-4dd5-a273-e173de1ca49f"). InnerVolumeSpecName "kube-api-access-24klq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:37:27 crc kubenswrapper[4766]: I1126 00:37:27.506276 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6807f6c2-7753-4dd5-a273-e173de1ca49f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6807f6c2-7753-4dd5-a273-e173de1ca49f" (UID: "6807f6c2-7753-4dd5-a273-e173de1ca49f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:37:27 crc kubenswrapper[4766]: I1126 00:37:27.559351 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6807f6c2-7753-4dd5-a273-e173de1ca49f-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:37:27 crc kubenswrapper[4766]: I1126 00:37:27.559385 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6807f6c2-7753-4dd5-a273-e173de1ca49f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:37:27 crc kubenswrapper[4766]: I1126 00:37:27.559397 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24klq\" (UniqueName: \"kubernetes.io/projected/6807f6c2-7753-4dd5-a273-e173de1ca49f-kube-api-access-24klq\") on node \"crc\" DevicePath \"\"" Nov 26 00:37:27 crc kubenswrapper[4766]: I1126 00:37:27.958697 4766 generic.go:334] "Generic (PLEG): container finished" podID="6807f6c2-7753-4dd5-a273-e173de1ca49f" containerID="d660717ee5aacb1cc18ed7339b44375f045055deae8337669ec46e326014b5bb" exitCode=0 Nov 26 00:37:27 crc kubenswrapper[4766]: I1126 00:37:27.958735 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxfsx" event={"ID":"6807f6c2-7753-4dd5-a273-e173de1ca49f","Type":"ContainerDied","Data":"d660717ee5aacb1cc18ed7339b44375f045055deae8337669ec46e326014b5bb"} Nov 26 00:37:27 crc kubenswrapper[4766]: I1126 00:37:27.958769 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qxfsx" event={"ID":"6807f6c2-7753-4dd5-a273-e173de1ca49f","Type":"ContainerDied","Data":"32920355fd7a12f17fba7cc00972db05675546a363d1d79c8e16fd2887822ed6"} Nov 26 00:37:27 crc kubenswrapper[4766]: I1126 00:37:27.958787 4766 scope.go:117] "RemoveContainer" containerID="d660717ee5aacb1cc18ed7339b44375f045055deae8337669ec46e326014b5bb" Nov 26 00:37:27 crc kubenswrapper[4766]: I1126 00:37:27.958781 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qxfsx" Nov 26 00:37:27 crc kubenswrapper[4766]: I1126 00:37:27.975899 4766 scope.go:117] "RemoveContainer" containerID="c5a55ca712f70c12c3ebda5816b2f47564c981539b360d75476871028e651881" Nov 26 00:37:27 crc kubenswrapper[4766]: I1126 00:37:27.980934 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qxfsx"] Nov 26 00:37:27 crc kubenswrapper[4766]: I1126 00:37:27.990240 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qxfsx"] Nov 26 00:37:27 crc kubenswrapper[4766]: I1126 00:37:27.993958 4766 scope.go:117] "RemoveContainer" containerID="6d5a8ca8e32aed732b731d442655db61148e6035737dbe9ff91dbddb2c8ae52c" Nov 26 00:37:28 crc kubenswrapper[4766]: I1126 00:37:28.015784 4766 scope.go:117] "RemoveContainer" containerID="d660717ee5aacb1cc18ed7339b44375f045055deae8337669ec46e326014b5bb" Nov 26 00:37:28 crc kubenswrapper[4766]: E1126 00:37:28.016244 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d660717ee5aacb1cc18ed7339b44375f045055deae8337669ec46e326014b5bb\": container with ID starting with d660717ee5aacb1cc18ed7339b44375f045055deae8337669ec46e326014b5bb not found: ID does not exist" containerID="d660717ee5aacb1cc18ed7339b44375f045055deae8337669ec46e326014b5bb" Nov 26 00:37:28 crc kubenswrapper[4766]: I1126 00:37:28.016274 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d660717ee5aacb1cc18ed7339b44375f045055deae8337669ec46e326014b5bb"} err="failed to get container status \"d660717ee5aacb1cc18ed7339b44375f045055deae8337669ec46e326014b5bb\": rpc error: code = NotFound desc = could not find container \"d660717ee5aacb1cc18ed7339b44375f045055deae8337669ec46e326014b5bb\": container with ID starting with d660717ee5aacb1cc18ed7339b44375f045055deae8337669ec46e326014b5bb not found: ID does not exist" Nov 26 00:37:28 crc kubenswrapper[4766]: I1126 00:37:28.016293 4766 scope.go:117] "RemoveContainer" containerID="c5a55ca712f70c12c3ebda5816b2f47564c981539b360d75476871028e651881" Nov 26 00:37:28 crc kubenswrapper[4766]: E1126 00:37:28.016596 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5a55ca712f70c12c3ebda5816b2f47564c981539b360d75476871028e651881\": container with ID starting with c5a55ca712f70c12c3ebda5816b2f47564c981539b360d75476871028e651881 not found: ID does not exist" containerID="c5a55ca712f70c12c3ebda5816b2f47564c981539b360d75476871028e651881" Nov 26 00:37:28 crc kubenswrapper[4766]: I1126 00:37:28.016641 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5a55ca712f70c12c3ebda5816b2f47564c981539b360d75476871028e651881"} err="failed to get container status \"c5a55ca712f70c12c3ebda5816b2f47564c981539b360d75476871028e651881\": rpc error: code = NotFound desc = could not find container \"c5a55ca712f70c12c3ebda5816b2f47564c981539b360d75476871028e651881\": container with ID starting with c5a55ca712f70c12c3ebda5816b2f47564c981539b360d75476871028e651881 not found: ID does not exist" Nov 26 00:37:28 crc kubenswrapper[4766]: I1126 00:37:28.016682 4766 scope.go:117] "RemoveContainer" containerID="6d5a8ca8e32aed732b731d442655db61148e6035737dbe9ff91dbddb2c8ae52c" Nov 26 00:37:28 crc kubenswrapper[4766]: E1126 00:37:28.016978 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d5a8ca8e32aed732b731d442655db61148e6035737dbe9ff91dbddb2c8ae52c\": container with ID starting with 6d5a8ca8e32aed732b731d442655db61148e6035737dbe9ff91dbddb2c8ae52c not found: ID does not exist" containerID="6d5a8ca8e32aed732b731d442655db61148e6035737dbe9ff91dbddb2c8ae52c" Nov 26 00:37:28 crc kubenswrapper[4766]: I1126 00:37:28.017000 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d5a8ca8e32aed732b731d442655db61148e6035737dbe9ff91dbddb2c8ae52c"} err="failed to get container status \"6d5a8ca8e32aed732b731d442655db61148e6035737dbe9ff91dbddb2c8ae52c\": rpc error: code = NotFound desc = could not find container \"6d5a8ca8e32aed732b731d442655db61148e6035737dbe9ff91dbddb2c8ae52c\": container with ID starting with 6d5a8ca8e32aed732b731d442655db61148e6035737dbe9ff91dbddb2c8ae52c not found: ID does not exist" Nov 26 00:37:29 crc kubenswrapper[4766]: I1126 00:37:29.835942 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6807f6c2-7753-4dd5-a273-e173de1ca49f" path="/var/lib/kubelet/pods/6807f6c2-7753-4dd5-a273-e173de1ca49f/volumes" Nov 26 00:37:30 crc kubenswrapper[4766]: I1126 00:37:30.548931 4766 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Nov 26 00:37:30 crc kubenswrapper[4766]: I1126 00:37:30.548991 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="aae58301-788d-430a-b020-a5f530d70c5a" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 26 00:37:40 crc kubenswrapper[4766]: I1126 00:37:40.549862 4766 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Nov 26 00:37:40 crc kubenswrapper[4766]: I1126 00:37:40.550435 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="aae58301-788d-430a-b020-a5f530d70c5a" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 26 00:37:44 crc kubenswrapper[4766]: I1126 00:37:44.858959 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vd4pl"] Nov 26 00:37:44 crc kubenswrapper[4766]: E1126 00:37:44.859569 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6807f6c2-7753-4dd5-a273-e173de1ca49f" containerName="extract-content" Nov 26 00:37:44 crc kubenswrapper[4766]: I1126 00:37:44.859584 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6807f6c2-7753-4dd5-a273-e173de1ca49f" containerName="extract-content" Nov 26 00:37:44 crc kubenswrapper[4766]: E1126 00:37:44.859596 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6807f6c2-7753-4dd5-a273-e173de1ca49f" containerName="registry-server" Nov 26 00:37:44 crc kubenswrapper[4766]: I1126 00:37:44.859604 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6807f6c2-7753-4dd5-a273-e173de1ca49f" containerName="registry-server" Nov 26 00:37:44 crc kubenswrapper[4766]: E1126 00:37:44.859615 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6807f6c2-7753-4dd5-a273-e173de1ca49f" containerName="extract-utilities" Nov 26 00:37:44 crc kubenswrapper[4766]: I1126 00:37:44.859623 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6807f6c2-7753-4dd5-a273-e173de1ca49f" containerName="extract-utilities" Nov 26 00:37:44 crc kubenswrapper[4766]: I1126 00:37:44.859766 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="6807f6c2-7753-4dd5-a273-e173de1ca49f" containerName="registry-server" Nov 26 00:37:44 crc kubenswrapper[4766]: I1126 00:37:44.860976 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vd4pl" Nov 26 00:37:44 crc kubenswrapper[4766]: I1126 00:37:44.871288 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vd4pl"] Nov 26 00:37:44 crc kubenswrapper[4766]: I1126 00:37:44.905266 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2475b25d-66de-43ec-bd12-44d22a1b9ebb-utilities\") pod \"redhat-marketplace-vd4pl\" (UID: \"2475b25d-66de-43ec-bd12-44d22a1b9ebb\") " pod="openshift-marketplace/redhat-marketplace-vd4pl" Nov 26 00:37:44 crc kubenswrapper[4766]: I1126 00:37:44.905366 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n99b\" (UniqueName: \"kubernetes.io/projected/2475b25d-66de-43ec-bd12-44d22a1b9ebb-kube-api-access-9n99b\") pod \"redhat-marketplace-vd4pl\" (UID: \"2475b25d-66de-43ec-bd12-44d22a1b9ebb\") " pod="openshift-marketplace/redhat-marketplace-vd4pl" Nov 26 00:37:44 crc kubenswrapper[4766]: I1126 00:37:44.905474 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2475b25d-66de-43ec-bd12-44d22a1b9ebb-catalog-content\") pod \"redhat-marketplace-vd4pl\" (UID: \"2475b25d-66de-43ec-bd12-44d22a1b9ebb\") " pod="openshift-marketplace/redhat-marketplace-vd4pl" Nov 26 00:37:45 crc kubenswrapper[4766]: I1126 00:37:45.006981 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n99b\" (UniqueName: \"kubernetes.io/projected/2475b25d-66de-43ec-bd12-44d22a1b9ebb-kube-api-access-9n99b\") pod \"redhat-marketplace-vd4pl\" (UID: \"2475b25d-66de-43ec-bd12-44d22a1b9ebb\") " pod="openshift-marketplace/redhat-marketplace-vd4pl" Nov 26 00:37:45 crc kubenswrapper[4766]: I1126 00:37:45.007101 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2475b25d-66de-43ec-bd12-44d22a1b9ebb-catalog-content\") pod \"redhat-marketplace-vd4pl\" (UID: \"2475b25d-66de-43ec-bd12-44d22a1b9ebb\") " pod="openshift-marketplace/redhat-marketplace-vd4pl" Nov 26 00:37:45 crc kubenswrapper[4766]: I1126 00:37:45.007146 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2475b25d-66de-43ec-bd12-44d22a1b9ebb-utilities\") pod \"redhat-marketplace-vd4pl\" (UID: \"2475b25d-66de-43ec-bd12-44d22a1b9ebb\") " pod="openshift-marketplace/redhat-marketplace-vd4pl" Nov 26 00:37:45 crc kubenswrapper[4766]: I1126 00:37:45.007761 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2475b25d-66de-43ec-bd12-44d22a1b9ebb-utilities\") pod \"redhat-marketplace-vd4pl\" (UID: \"2475b25d-66de-43ec-bd12-44d22a1b9ebb\") " pod="openshift-marketplace/redhat-marketplace-vd4pl" Nov 26 00:37:45 crc kubenswrapper[4766]: I1126 00:37:45.007795 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2475b25d-66de-43ec-bd12-44d22a1b9ebb-catalog-content\") pod \"redhat-marketplace-vd4pl\" (UID: \"2475b25d-66de-43ec-bd12-44d22a1b9ebb\") " pod="openshift-marketplace/redhat-marketplace-vd4pl" Nov 26 00:37:45 crc kubenswrapper[4766]: I1126 00:37:45.028951 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n99b\" (UniqueName: \"kubernetes.io/projected/2475b25d-66de-43ec-bd12-44d22a1b9ebb-kube-api-access-9n99b\") pod \"redhat-marketplace-vd4pl\" (UID: \"2475b25d-66de-43ec-bd12-44d22a1b9ebb\") " pod="openshift-marketplace/redhat-marketplace-vd4pl" Nov 26 00:37:45 crc kubenswrapper[4766]: I1126 00:37:45.177509 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vd4pl" Nov 26 00:37:45 crc kubenswrapper[4766]: I1126 00:37:45.448717 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vd4pl"] Nov 26 00:37:46 crc kubenswrapper[4766]: I1126 00:37:46.077663 4766 generic.go:334] "Generic (PLEG): container finished" podID="2475b25d-66de-43ec-bd12-44d22a1b9ebb" containerID="3129475353aa0297f16a7949dbfe2e97b10e131be07524d377c239c924768e81" exitCode=0 Nov 26 00:37:46 crc kubenswrapper[4766]: I1126 00:37:46.077718 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vd4pl" event={"ID":"2475b25d-66de-43ec-bd12-44d22a1b9ebb","Type":"ContainerDied","Data":"3129475353aa0297f16a7949dbfe2e97b10e131be07524d377c239c924768e81"} Nov 26 00:37:46 crc kubenswrapper[4766]: I1126 00:37:46.077759 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vd4pl" event={"ID":"2475b25d-66de-43ec-bd12-44d22a1b9ebb","Type":"ContainerStarted","Data":"2447e7903d4f5a40bafe277c3e178ff5200c08047723f7ab8359fa5bc0d11817"} Nov 26 00:37:47 crc kubenswrapper[4766]: I1126 00:37:47.103295 4766 generic.go:334] "Generic (PLEG): container finished" podID="2475b25d-66de-43ec-bd12-44d22a1b9ebb" containerID="56f7bdd954b952d8888152e2297acbf595b67c375cd0a7b41556cc07f82c3a03" exitCode=0 Nov 26 00:37:47 crc kubenswrapper[4766]: I1126 00:37:47.103348 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vd4pl" event={"ID":"2475b25d-66de-43ec-bd12-44d22a1b9ebb","Type":"ContainerDied","Data":"56f7bdd954b952d8888152e2297acbf595b67c375cd0a7b41556cc07f82c3a03"} Nov 26 00:37:48 crc kubenswrapper[4766]: I1126 00:37:48.110682 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vd4pl" event={"ID":"2475b25d-66de-43ec-bd12-44d22a1b9ebb","Type":"ContainerStarted","Data":"95257da13b3ac0d1558182522c70d671eeb7ce97bb6519db4ba38864131cc623"} Nov 26 00:37:48 crc kubenswrapper[4766]: I1126 00:37:48.130843 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vd4pl" podStartSLOduration=2.7002391919999997 podStartE2EDuration="4.130825799s" podCreationTimestamp="2025-11-26 00:37:44 +0000 UTC" firstStartedPulling="2025-11-26 00:37:46.080153018 +0000 UTC m=+846.928923468" lastFinishedPulling="2025-11-26 00:37:47.510739645 +0000 UTC m=+848.359510075" observedRunningTime="2025-11-26 00:37:48.126135131 +0000 UTC m=+848.974905561" watchObservedRunningTime="2025-11-26 00:37:48.130825799 +0000 UTC m=+848.979596229" Nov 26 00:37:50 crc kubenswrapper[4766]: I1126 00:37:50.548591 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Nov 26 00:37:55 crc kubenswrapper[4766]: I1126 00:37:55.177819 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vd4pl" Nov 26 00:37:55 crc kubenswrapper[4766]: I1126 00:37:55.178290 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vd4pl" Nov 26 00:37:55 crc kubenswrapper[4766]: I1126 00:37:55.234931 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vd4pl" Nov 26 00:37:56 crc kubenswrapper[4766]: I1126 00:37:56.203491 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vd4pl" Nov 26 00:37:56 crc kubenswrapper[4766]: I1126 00:37:56.246191 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vd4pl"] Nov 26 00:37:58 crc kubenswrapper[4766]: I1126 00:37:58.171080 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vd4pl" podUID="2475b25d-66de-43ec-bd12-44d22a1b9ebb" containerName="registry-server" containerID="cri-o://95257da13b3ac0d1558182522c70d671eeb7ce97bb6519db4ba38864131cc623" gracePeriod=2 Nov 26 00:37:58 crc kubenswrapper[4766]: I1126 00:37:58.643182 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vd4pl" Nov 26 00:37:58 crc kubenswrapper[4766]: I1126 00:37:58.757216 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2475b25d-66de-43ec-bd12-44d22a1b9ebb-utilities\") pod \"2475b25d-66de-43ec-bd12-44d22a1b9ebb\" (UID: \"2475b25d-66de-43ec-bd12-44d22a1b9ebb\") " Nov 26 00:37:58 crc kubenswrapper[4766]: I1126 00:37:58.757280 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2475b25d-66de-43ec-bd12-44d22a1b9ebb-catalog-content\") pod \"2475b25d-66de-43ec-bd12-44d22a1b9ebb\" (UID: \"2475b25d-66de-43ec-bd12-44d22a1b9ebb\") " Nov 26 00:37:58 crc kubenswrapper[4766]: I1126 00:37:58.757310 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n99b\" (UniqueName: \"kubernetes.io/projected/2475b25d-66de-43ec-bd12-44d22a1b9ebb-kube-api-access-9n99b\") pod \"2475b25d-66de-43ec-bd12-44d22a1b9ebb\" (UID: \"2475b25d-66de-43ec-bd12-44d22a1b9ebb\") " Nov 26 00:37:58 crc kubenswrapper[4766]: I1126 00:37:58.758866 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2475b25d-66de-43ec-bd12-44d22a1b9ebb-utilities" (OuterVolumeSpecName: "utilities") pod "2475b25d-66de-43ec-bd12-44d22a1b9ebb" (UID: "2475b25d-66de-43ec-bd12-44d22a1b9ebb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:37:58 crc kubenswrapper[4766]: I1126 00:37:58.762270 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2475b25d-66de-43ec-bd12-44d22a1b9ebb-kube-api-access-9n99b" (OuterVolumeSpecName: "kube-api-access-9n99b") pod "2475b25d-66de-43ec-bd12-44d22a1b9ebb" (UID: "2475b25d-66de-43ec-bd12-44d22a1b9ebb"). InnerVolumeSpecName "kube-api-access-9n99b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:37:58 crc kubenswrapper[4766]: I1126 00:37:58.773316 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2475b25d-66de-43ec-bd12-44d22a1b9ebb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2475b25d-66de-43ec-bd12-44d22a1b9ebb" (UID: "2475b25d-66de-43ec-bd12-44d22a1b9ebb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:37:58 crc kubenswrapper[4766]: I1126 00:37:58.858408 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2475b25d-66de-43ec-bd12-44d22a1b9ebb-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:37:58 crc kubenswrapper[4766]: I1126 00:37:58.858442 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2475b25d-66de-43ec-bd12-44d22a1b9ebb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:37:58 crc kubenswrapper[4766]: I1126 00:37:58.858455 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n99b\" (UniqueName: \"kubernetes.io/projected/2475b25d-66de-43ec-bd12-44d22a1b9ebb-kube-api-access-9n99b\") on node \"crc\" DevicePath \"\"" Nov 26 00:37:59 crc kubenswrapper[4766]: I1126 00:37:59.180188 4766 generic.go:334] "Generic (PLEG): container finished" podID="2475b25d-66de-43ec-bd12-44d22a1b9ebb" containerID="95257da13b3ac0d1558182522c70d671eeb7ce97bb6519db4ba38864131cc623" exitCode=0 Nov 26 00:37:59 crc kubenswrapper[4766]: I1126 00:37:59.180238 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vd4pl" event={"ID":"2475b25d-66de-43ec-bd12-44d22a1b9ebb","Type":"ContainerDied","Data":"95257da13b3ac0d1558182522c70d671eeb7ce97bb6519db4ba38864131cc623"} Nov 26 00:37:59 crc kubenswrapper[4766]: I1126 00:37:59.180269 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vd4pl" event={"ID":"2475b25d-66de-43ec-bd12-44d22a1b9ebb","Type":"ContainerDied","Data":"2447e7903d4f5a40bafe277c3e178ff5200c08047723f7ab8359fa5bc0d11817"} Nov 26 00:37:59 crc kubenswrapper[4766]: I1126 00:37:59.180288 4766 scope.go:117] "RemoveContainer" containerID="95257da13b3ac0d1558182522c70d671eeb7ce97bb6519db4ba38864131cc623" Nov 26 00:37:59 crc kubenswrapper[4766]: I1126 00:37:59.180443 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vd4pl" Nov 26 00:37:59 crc kubenswrapper[4766]: I1126 00:37:59.208208 4766 scope.go:117] "RemoveContainer" containerID="56f7bdd954b952d8888152e2297acbf595b67c375cd0a7b41556cc07f82c3a03" Nov 26 00:37:59 crc kubenswrapper[4766]: I1126 00:37:59.220818 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vd4pl"] Nov 26 00:37:59 crc kubenswrapper[4766]: I1126 00:37:59.230931 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vd4pl"] Nov 26 00:37:59 crc kubenswrapper[4766]: I1126 00:37:59.233757 4766 scope.go:117] "RemoveContainer" containerID="3129475353aa0297f16a7949dbfe2e97b10e131be07524d377c239c924768e81" Nov 26 00:37:59 crc kubenswrapper[4766]: I1126 00:37:59.265945 4766 scope.go:117] "RemoveContainer" containerID="95257da13b3ac0d1558182522c70d671eeb7ce97bb6519db4ba38864131cc623" Nov 26 00:37:59 crc kubenswrapper[4766]: E1126 00:37:59.279784 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95257da13b3ac0d1558182522c70d671eeb7ce97bb6519db4ba38864131cc623\": container with ID starting with 95257da13b3ac0d1558182522c70d671eeb7ce97bb6519db4ba38864131cc623 not found: ID does not exist" containerID="95257da13b3ac0d1558182522c70d671eeb7ce97bb6519db4ba38864131cc623" Nov 26 00:37:59 crc kubenswrapper[4766]: I1126 00:37:59.279829 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95257da13b3ac0d1558182522c70d671eeb7ce97bb6519db4ba38864131cc623"} err="failed to get container status \"95257da13b3ac0d1558182522c70d671eeb7ce97bb6519db4ba38864131cc623\": rpc error: code = NotFound desc = could not find container \"95257da13b3ac0d1558182522c70d671eeb7ce97bb6519db4ba38864131cc623\": container with ID starting with 95257da13b3ac0d1558182522c70d671eeb7ce97bb6519db4ba38864131cc623 not found: ID does not exist" Nov 26 00:37:59 crc kubenswrapper[4766]: I1126 00:37:59.279853 4766 scope.go:117] "RemoveContainer" containerID="56f7bdd954b952d8888152e2297acbf595b67c375cd0a7b41556cc07f82c3a03" Nov 26 00:37:59 crc kubenswrapper[4766]: E1126 00:37:59.281762 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56f7bdd954b952d8888152e2297acbf595b67c375cd0a7b41556cc07f82c3a03\": container with ID starting with 56f7bdd954b952d8888152e2297acbf595b67c375cd0a7b41556cc07f82c3a03 not found: ID does not exist" containerID="56f7bdd954b952d8888152e2297acbf595b67c375cd0a7b41556cc07f82c3a03" Nov 26 00:37:59 crc kubenswrapper[4766]: I1126 00:37:59.281784 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56f7bdd954b952d8888152e2297acbf595b67c375cd0a7b41556cc07f82c3a03"} err="failed to get container status \"56f7bdd954b952d8888152e2297acbf595b67c375cd0a7b41556cc07f82c3a03\": rpc error: code = NotFound desc = could not find container \"56f7bdd954b952d8888152e2297acbf595b67c375cd0a7b41556cc07f82c3a03\": container with ID starting with 56f7bdd954b952d8888152e2297acbf595b67c375cd0a7b41556cc07f82c3a03 not found: ID does not exist" Nov 26 00:37:59 crc kubenswrapper[4766]: I1126 00:37:59.281798 4766 scope.go:117] "RemoveContainer" containerID="3129475353aa0297f16a7949dbfe2e97b10e131be07524d377c239c924768e81" Nov 26 00:37:59 crc kubenswrapper[4766]: E1126 00:37:59.286831 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3129475353aa0297f16a7949dbfe2e97b10e131be07524d377c239c924768e81\": container with ID starting with 3129475353aa0297f16a7949dbfe2e97b10e131be07524d377c239c924768e81 not found: ID does not exist" containerID="3129475353aa0297f16a7949dbfe2e97b10e131be07524d377c239c924768e81" Nov 26 00:37:59 crc kubenswrapper[4766]: I1126 00:37:59.286881 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3129475353aa0297f16a7949dbfe2e97b10e131be07524d377c239c924768e81"} err="failed to get container status \"3129475353aa0297f16a7949dbfe2e97b10e131be07524d377c239c924768e81\": rpc error: code = NotFound desc = could not find container \"3129475353aa0297f16a7949dbfe2e97b10e131be07524d377c239c924768e81\": container with ID starting with 3129475353aa0297f16a7949dbfe2e97b10e131be07524d377c239c924768e81 not found: ID does not exist" Nov 26 00:37:59 crc kubenswrapper[4766]: I1126 00:37:59.855284 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2475b25d-66de-43ec-bd12-44d22a1b9ebb" path="/var/lib/kubelet/pods/2475b25d-66de-43ec-bd12-44d22a1b9ebb/volumes" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.522693 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-tcpsm"] Nov 26 00:38:10 crc kubenswrapper[4766]: E1126 00:38:10.523390 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2475b25d-66de-43ec-bd12-44d22a1b9ebb" containerName="extract-content" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.523402 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="2475b25d-66de-43ec-bd12-44d22a1b9ebb" containerName="extract-content" Nov 26 00:38:10 crc kubenswrapper[4766]: E1126 00:38:10.523415 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2475b25d-66de-43ec-bd12-44d22a1b9ebb" containerName="registry-server" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.523422 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="2475b25d-66de-43ec-bd12-44d22a1b9ebb" containerName="registry-server" Nov 26 00:38:10 crc kubenswrapper[4766]: E1126 00:38:10.523446 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2475b25d-66de-43ec-bd12-44d22a1b9ebb" containerName="extract-utilities" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.523459 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="2475b25d-66de-43ec-bd12-44d22a1b9ebb" containerName="extract-utilities" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.523578 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="2475b25d-66de-43ec-bd12-44d22a1b9ebb" containerName="registry-server" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.524175 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.526248 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.526906 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.527338 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-bfcr8" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.527500 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.528446 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.531750 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-collector-token\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.531813 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-trusted-ca\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.531846 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-metrics\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.531874 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c7f23a28-529a-4df5-9b8a-2a21f9300b11-tmp\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.531905 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-collector-syslog-receiver\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.531935 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-config-openshift-service-cacrt\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.531970 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-entrypoint\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.531994 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-config\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.532019 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrpjc\" (UniqueName: \"kubernetes.io/projected/c7f23a28-529a-4df5-9b8a-2a21f9300b11-kube-api-access-qrpjc\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.532062 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/c7f23a28-529a-4df5-9b8a-2a21f9300b11-sa-token\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.532107 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/c7f23a28-529a-4df5-9b8a-2a21f9300b11-datadir\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.532921 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.540296 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-tcpsm"] Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.633250 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/c7f23a28-529a-4df5-9b8a-2a21f9300b11-datadir\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.633319 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-collector-token\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.633352 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-trusted-ca\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.633375 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-metrics\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.633391 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c7f23a28-529a-4df5-9b8a-2a21f9300b11-tmp\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.633411 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-collector-syslog-receiver\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.633419 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/c7f23a28-529a-4df5-9b8a-2a21f9300b11-datadir\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.633434 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-config-openshift-service-cacrt\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.633552 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-entrypoint\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.633580 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-config\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.633617 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrpjc\" (UniqueName: \"kubernetes.io/projected/c7f23a28-529a-4df5-9b8a-2a21f9300b11-kube-api-access-qrpjc\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.633712 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/c7f23a28-529a-4df5-9b8a-2a21f9300b11-sa-token\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.634075 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-config-openshift-service-cacrt\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.634886 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-entrypoint\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.634943 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-trusted-ca\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: E1126 00:38:10.635020 4766 secret.go:188] Couldn't get secret openshift-logging/collector-syslog-receiver: secret "collector-syslog-receiver" not found Nov 26 00:38:10 crc kubenswrapper[4766]: E1126 00:38:10.635063 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-collector-syslog-receiver podName:c7f23a28-529a-4df5-9b8a-2a21f9300b11 nodeName:}" failed. No retries permitted until 2025-11-26 00:38:11.135048208 +0000 UTC m=+871.983818638 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "collector-syslog-receiver" (UniqueName: "kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-collector-syslog-receiver") pod "collector-tcpsm" (UID: "c7f23a28-529a-4df5-9b8a-2a21f9300b11") : secret "collector-syslog-receiver" not found Nov 26 00:38:10 crc kubenswrapper[4766]: E1126 00:38:10.635320 4766 secret.go:188] Couldn't get secret openshift-logging/collector-metrics: secret "collector-metrics" not found Nov 26 00:38:10 crc kubenswrapper[4766]: E1126 00:38:10.635604 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-metrics podName:c7f23a28-529a-4df5-9b8a-2a21f9300b11 nodeName:}" failed. No retries permitted until 2025-11-26 00:38:11.135593991 +0000 UTC m=+871.984364491 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics" (UniqueName: "kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-metrics") pod "collector-tcpsm" (UID: "c7f23a28-529a-4df5-9b8a-2a21f9300b11") : secret "collector-metrics" not found Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.635511 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-config\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.639992 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-collector-token\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.649917 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c7f23a28-529a-4df5-9b8a-2a21f9300b11-tmp\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.666166 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/c7f23a28-529a-4df5-9b8a-2a21f9300b11-sa-token\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.671359 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrpjc\" (UniqueName: \"kubernetes.io/projected/c7f23a28-529a-4df5-9b8a-2a21f9300b11-kube-api-access-qrpjc\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:10 crc kubenswrapper[4766]: I1126 00:38:10.698016 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-tcpsm"] Nov 26 00:38:10 crc kubenswrapper[4766]: E1126 00:38:10.698526 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[collector-syslog-receiver metrics], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-logging/collector-tcpsm" podUID="c7f23a28-529a-4df5-9b8a-2a21f9300b11" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.139683 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-metrics\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.139735 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-collector-syslog-receiver\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.142924 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-collector-syslog-receiver\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.148121 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-metrics\") pod \"collector-tcpsm\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " pod="openshift-logging/collector-tcpsm" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.262755 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-tcpsm" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.272342 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-tcpsm" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.341945 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/c7f23a28-529a-4df5-9b8a-2a21f9300b11-datadir\") pod \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.342010 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-config\") pod \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.342058 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-config-openshift-service-cacrt\") pod \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.342084 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-entrypoint\") pod \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.342064 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c7f23a28-529a-4df5-9b8a-2a21f9300b11-datadir" (OuterVolumeSpecName: "datadir") pod "c7f23a28-529a-4df5-9b8a-2a21f9300b11" (UID: "c7f23a28-529a-4df5-9b8a-2a21f9300b11"). InnerVolumeSpecName "datadir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.342131 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-collector-syslog-receiver\") pod \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.342166 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c7f23a28-529a-4df5-9b8a-2a21f9300b11-tmp\") pod \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.342193 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrpjc\" (UniqueName: \"kubernetes.io/projected/c7f23a28-529a-4df5-9b8a-2a21f9300b11-kube-api-access-qrpjc\") pod \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.342226 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-trusted-ca\") pod \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.342248 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/c7f23a28-529a-4df5-9b8a-2a21f9300b11-sa-token\") pod \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.342279 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-collector-token\") pod \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.342307 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-metrics\") pod \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\" (UID: \"c7f23a28-529a-4df5-9b8a-2a21f9300b11\") " Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.342528 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-config-openshift-service-cacrt" (OuterVolumeSpecName: "config-openshift-service-cacrt") pod "c7f23a28-529a-4df5-9b8a-2a21f9300b11" (UID: "c7f23a28-529a-4df5-9b8a-2a21f9300b11"). InnerVolumeSpecName "config-openshift-service-cacrt". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.342540 4766 reconciler_common.go:293] "Volume detached for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/c7f23a28-529a-4df5-9b8a-2a21f9300b11-datadir\") on node \"crc\" DevicePath \"\"" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.342855 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-config" (OuterVolumeSpecName: "config") pod "c7f23a28-529a-4df5-9b8a-2a21f9300b11" (UID: "c7f23a28-529a-4df5-9b8a-2a21f9300b11"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.343731 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-entrypoint" (OuterVolumeSpecName: "entrypoint") pod "c7f23a28-529a-4df5-9b8a-2a21f9300b11" (UID: "c7f23a28-529a-4df5-9b8a-2a21f9300b11"). InnerVolumeSpecName "entrypoint". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.343796 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "c7f23a28-529a-4df5-9b8a-2a21f9300b11" (UID: "c7f23a28-529a-4df5-9b8a-2a21f9300b11"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.345947 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-collector-syslog-receiver" (OuterVolumeSpecName: "collector-syslog-receiver") pod "c7f23a28-529a-4df5-9b8a-2a21f9300b11" (UID: "c7f23a28-529a-4df5-9b8a-2a21f9300b11"). InnerVolumeSpecName "collector-syslog-receiver". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.346060 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-metrics" (OuterVolumeSpecName: "metrics") pod "c7f23a28-529a-4df5-9b8a-2a21f9300b11" (UID: "c7f23a28-529a-4df5-9b8a-2a21f9300b11"). InnerVolumeSpecName "metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.346897 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7f23a28-529a-4df5-9b8a-2a21f9300b11-kube-api-access-qrpjc" (OuterVolumeSpecName: "kube-api-access-qrpjc") pod "c7f23a28-529a-4df5-9b8a-2a21f9300b11" (UID: "c7f23a28-529a-4df5-9b8a-2a21f9300b11"). InnerVolumeSpecName "kube-api-access-qrpjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.347574 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7f23a28-529a-4df5-9b8a-2a21f9300b11-sa-token" (OuterVolumeSpecName: "sa-token") pod "c7f23a28-529a-4df5-9b8a-2a21f9300b11" (UID: "c7f23a28-529a-4df5-9b8a-2a21f9300b11"). InnerVolumeSpecName "sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.347850 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7f23a28-529a-4df5-9b8a-2a21f9300b11-tmp" (OuterVolumeSpecName: "tmp") pod "c7f23a28-529a-4df5-9b8a-2a21f9300b11" (UID: "c7f23a28-529a-4df5-9b8a-2a21f9300b11"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.349937 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-collector-token" (OuterVolumeSpecName: "collector-token") pod "c7f23a28-529a-4df5-9b8a-2a21f9300b11" (UID: "c7f23a28-529a-4df5-9b8a-2a21f9300b11"). InnerVolumeSpecName "collector-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.443539 4766 reconciler_common.go:293] "Volume detached for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-collector-syslog-receiver\") on node \"crc\" DevicePath \"\"" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.443586 4766 reconciler_common.go:293] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c7f23a28-529a-4df5-9b8a-2a21f9300b11-tmp\") on node \"crc\" DevicePath \"\"" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.443599 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrpjc\" (UniqueName: \"kubernetes.io/projected/c7f23a28-529a-4df5-9b8a-2a21f9300b11-kube-api-access-qrpjc\") on node \"crc\" DevicePath \"\"" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.443609 4766 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.443619 4766 reconciler_common.go:293] "Volume detached for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/c7f23a28-529a-4df5-9b8a-2a21f9300b11-sa-token\") on node \"crc\" DevicePath \"\"" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.443634 4766 reconciler_common.go:293] "Volume detached for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-collector-token\") on node \"crc\" DevicePath \"\"" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.443861 4766 reconciler_common.go:293] "Volume detached for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/c7f23a28-529a-4df5-9b8a-2a21f9300b11-metrics\") on node \"crc\" DevicePath \"\"" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.443884 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.443899 4766 reconciler_common.go:293] "Volume detached for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-config-openshift-service-cacrt\") on node \"crc\" DevicePath \"\"" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.443914 4766 reconciler_common.go:293] "Volume detached for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/c7f23a28-529a-4df5-9b8a-2a21f9300b11-entrypoint\") on node \"crc\" DevicePath \"\"" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.570735 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6vvl2"] Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.572223 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6vvl2" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.583429 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6vvl2"] Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.747684 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05ffb7d1-00de-436c-b9d6-b4d517479ad8-catalog-content\") pod \"certified-operators-6vvl2\" (UID: \"05ffb7d1-00de-436c-b9d6-b4d517479ad8\") " pod="openshift-marketplace/certified-operators-6vvl2" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.748040 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05ffb7d1-00de-436c-b9d6-b4d517479ad8-utilities\") pod \"certified-operators-6vvl2\" (UID: \"05ffb7d1-00de-436c-b9d6-b4d517479ad8\") " pod="openshift-marketplace/certified-operators-6vvl2" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.748090 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqsxx\" (UniqueName: \"kubernetes.io/projected/05ffb7d1-00de-436c-b9d6-b4d517479ad8-kube-api-access-dqsxx\") pod \"certified-operators-6vvl2\" (UID: \"05ffb7d1-00de-436c-b9d6-b4d517479ad8\") " pod="openshift-marketplace/certified-operators-6vvl2" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.849066 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05ffb7d1-00de-436c-b9d6-b4d517479ad8-catalog-content\") pod \"certified-operators-6vvl2\" (UID: \"05ffb7d1-00de-436c-b9d6-b4d517479ad8\") " pod="openshift-marketplace/certified-operators-6vvl2" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.849145 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05ffb7d1-00de-436c-b9d6-b4d517479ad8-utilities\") pod \"certified-operators-6vvl2\" (UID: \"05ffb7d1-00de-436c-b9d6-b4d517479ad8\") " pod="openshift-marketplace/certified-operators-6vvl2" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.849195 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqsxx\" (UniqueName: \"kubernetes.io/projected/05ffb7d1-00de-436c-b9d6-b4d517479ad8-kube-api-access-dqsxx\") pod \"certified-operators-6vvl2\" (UID: \"05ffb7d1-00de-436c-b9d6-b4d517479ad8\") " pod="openshift-marketplace/certified-operators-6vvl2" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.849670 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05ffb7d1-00de-436c-b9d6-b4d517479ad8-catalog-content\") pod \"certified-operators-6vvl2\" (UID: \"05ffb7d1-00de-436c-b9d6-b4d517479ad8\") " pod="openshift-marketplace/certified-operators-6vvl2" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.849771 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05ffb7d1-00de-436c-b9d6-b4d517479ad8-utilities\") pod \"certified-operators-6vvl2\" (UID: \"05ffb7d1-00de-436c-b9d6-b4d517479ad8\") " pod="openshift-marketplace/certified-operators-6vvl2" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.866892 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqsxx\" (UniqueName: \"kubernetes.io/projected/05ffb7d1-00de-436c-b9d6-b4d517479ad8-kube-api-access-dqsxx\") pod \"certified-operators-6vvl2\" (UID: \"05ffb7d1-00de-436c-b9d6-b4d517479ad8\") " pod="openshift-marketplace/certified-operators-6vvl2" Nov 26 00:38:11 crc kubenswrapper[4766]: I1126 00:38:11.896516 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6vvl2" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.268596 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-tcpsm" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.305298 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-tcpsm"] Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.324175 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-logging/collector-tcpsm"] Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.332794 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-zzrrp"] Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.333858 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.337599 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-zzrrp"] Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.338927 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-bfcr8" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.339836 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.341830 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.342043 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.345156 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.347264 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.365868 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-trusted-ca\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.366161 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-collector-syslog-receiver\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.366200 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-config\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.366236 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-entrypoint\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.366297 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-tmp\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.366382 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-metrics\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.366447 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-collector-token\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.366480 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-datadir\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.366519 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llzvq\" (UniqueName: \"kubernetes.io/projected/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-kube-api-access-llzvq\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.366548 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-config-openshift-service-cacrt\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.366578 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-sa-token\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.372149 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6vvl2"] Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.467576 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-collector-token\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.467683 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-datadir\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.467717 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llzvq\" (UniqueName: \"kubernetes.io/projected/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-kube-api-access-llzvq\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.467770 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-config-openshift-service-cacrt\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.467799 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-sa-token\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.467860 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-trusted-ca\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.467882 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-collector-syslog-receiver\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.467916 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-config\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.467939 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-entrypoint\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.467957 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-tmp\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.467996 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-metrics\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.468932 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-datadir\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.469517 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-config-openshift-service-cacrt\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.469784 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-entrypoint\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.470544 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-config\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.470965 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-trusted-ca\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.474585 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-tmp\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.475026 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-metrics\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.476240 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-collector-syslog-receiver\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.478243 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-collector-token\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.487508 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llzvq\" (UniqueName: \"kubernetes.io/projected/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-kube-api-access-llzvq\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.487589 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/bed1e93e-177c-4e72-99b7-e35f6d8f6cf9-sa-token\") pod \"collector-zzrrp\" (UID: \"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9\") " pod="openshift-logging/collector-zzrrp" Nov 26 00:38:12 crc kubenswrapper[4766]: I1126 00:38:12.673496 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-zzrrp" Nov 26 00:38:13 crc kubenswrapper[4766]: I1126 00:38:13.157558 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-zzrrp"] Nov 26 00:38:13 crc kubenswrapper[4766]: W1126 00:38:13.165775 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbed1e93e_177c_4e72_99b7_e35f6d8f6cf9.slice/crio-66ad01ac8289f0230e43fb2c180b78efcdba35f818ce5a31c4e1979a1bd63971 WatchSource:0}: Error finding container 66ad01ac8289f0230e43fb2c180b78efcdba35f818ce5a31c4e1979a1bd63971: Status 404 returned error can't find the container with id 66ad01ac8289f0230e43fb2c180b78efcdba35f818ce5a31c4e1979a1bd63971 Nov 26 00:38:13 crc kubenswrapper[4766]: I1126 00:38:13.275757 4766 generic.go:334] "Generic (PLEG): container finished" podID="05ffb7d1-00de-436c-b9d6-b4d517479ad8" containerID="0d510036621ca419e6d33165a85ebc295de739aff535789980aa87eca8374d7a" exitCode=0 Nov 26 00:38:13 crc kubenswrapper[4766]: I1126 00:38:13.275843 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vvl2" event={"ID":"05ffb7d1-00de-436c-b9d6-b4d517479ad8","Type":"ContainerDied","Data":"0d510036621ca419e6d33165a85ebc295de739aff535789980aa87eca8374d7a"} Nov 26 00:38:13 crc kubenswrapper[4766]: I1126 00:38:13.275874 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vvl2" event={"ID":"05ffb7d1-00de-436c-b9d6-b4d517479ad8","Type":"ContainerStarted","Data":"ba3258047d665d2674423f705d30bfd9bd4e1be91fd797b5094bb3fac718982f"} Nov 26 00:38:13 crc kubenswrapper[4766]: I1126 00:38:13.276806 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-zzrrp" event={"ID":"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9","Type":"ContainerStarted","Data":"66ad01ac8289f0230e43fb2c180b78efcdba35f818ce5a31c4e1979a1bd63971"} Nov 26 00:38:13 crc kubenswrapper[4766]: I1126 00:38:13.840960 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7f23a28-529a-4df5-9b8a-2a21f9300b11" path="/var/lib/kubelet/pods/c7f23a28-529a-4df5-9b8a-2a21f9300b11/volumes" Nov 26 00:38:14 crc kubenswrapper[4766]: I1126 00:38:14.285185 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vvl2" event={"ID":"05ffb7d1-00de-436c-b9d6-b4d517479ad8","Type":"ContainerStarted","Data":"ee7e76c0751d9936743c1876c5e7e321195f5d63d6fbff4f240b99a3c5333a74"} Nov 26 00:38:15 crc kubenswrapper[4766]: I1126 00:38:15.294914 4766 generic.go:334] "Generic (PLEG): container finished" podID="05ffb7d1-00de-436c-b9d6-b4d517479ad8" containerID="ee7e76c0751d9936743c1876c5e7e321195f5d63d6fbff4f240b99a3c5333a74" exitCode=0 Nov 26 00:38:15 crc kubenswrapper[4766]: I1126 00:38:15.294990 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vvl2" event={"ID":"05ffb7d1-00de-436c-b9d6-b4d517479ad8","Type":"ContainerDied","Data":"ee7e76c0751d9936743c1876c5e7e321195f5d63d6fbff4f240b99a3c5333a74"} Nov 26 00:38:16 crc kubenswrapper[4766]: I1126 00:38:16.305013 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vvl2" event={"ID":"05ffb7d1-00de-436c-b9d6-b4d517479ad8","Type":"ContainerStarted","Data":"ca63a3089127a2d6718d812658f81c8a2fa2e7d40d45261cb92513e4e5479b5e"} Nov 26 00:38:16 crc kubenswrapper[4766]: I1126 00:38:16.329420 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6vvl2" podStartSLOduration=2.652737251 podStartE2EDuration="5.329395476s" podCreationTimestamp="2025-11-26 00:38:11 +0000 UTC" firstStartedPulling="2025-11-26 00:38:13.27757226 +0000 UTC m=+874.126342690" lastFinishedPulling="2025-11-26 00:38:15.954230485 +0000 UTC m=+876.803000915" observedRunningTime="2025-11-26 00:38:16.325257342 +0000 UTC m=+877.174027772" watchObservedRunningTime="2025-11-26 00:38:16.329395476 +0000 UTC m=+877.178165906" Nov 26 00:38:21 crc kubenswrapper[4766]: I1126 00:38:21.897390 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6vvl2" Nov 26 00:38:21 crc kubenswrapper[4766]: I1126 00:38:21.897745 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6vvl2" Nov 26 00:38:21 crc kubenswrapper[4766]: I1126 00:38:21.935414 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6vvl2" Nov 26 00:38:22 crc kubenswrapper[4766]: I1126 00:38:22.350897 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-zzrrp" event={"ID":"bed1e93e-177c-4e72-99b7-e35f6d8f6cf9","Type":"ContainerStarted","Data":"0eabb7de45f8f914a378a0451c11517c01659d14aa14999330b4ecab7201723a"} Nov 26 00:38:22 crc kubenswrapper[4766]: I1126 00:38:22.373532 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/collector-zzrrp" podStartSLOduration=2.227155123 podStartE2EDuration="10.373513855s" podCreationTimestamp="2025-11-26 00:38:12 +0000 UTC" firstStartedPulling="2025-11-26 00:38:13.16744876 +0000 UTC m=+874.016219180" lastFinishedPulling="2025-11-26 00:38:21.313807482 +0000 UTC m=+882.162577912" observedRunningTime="2025-11-26 00:38:22.372709845 +0000 UTC m=+883.221480275" watchObservedRunningTime="2025-11-26 00:38:22.373513855 +0000 UTC m=+883.222284285" Nov 26 00:38:22 crc kubenswrapper[4766]: I1126 00:38:22.405298 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6vvl2" Nov 26 00:38:22 crc kubenswrapper[4766]: I1126 00:38:22.449329 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6vvl2"] Nov 26 00:38:24 crc kubenswrapper[4766]: I1126 00:38:24.363229 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6vvl2" podUID="05ffb7d1-00de-436c-b9d6-b4d517479ad8" containerName="registry-server" containerID="cri-o://ca63a3089127a2d6718d812658f81c8a2fa2e7d40d45261cb92513e4e5479b5e" gracePeriod=2 Nov 26 00:38:24 crc kubenswrapper[4766]: I1126 00:38:24.749341 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6vvl2" Nov 26 00:38:24 crc kubenswrapper[4766]: I1126 00:38:24.801115 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqsxx\" (UniqueName: \"kubernetes.io/projected/05ffb7d1-00de-436c-b9d6-b4d517479ad8-kube-api-access-dqsxx\") pod \"05ffb7d1-00de-436c-b9d6-b4d517479ad8\" (UID: \"05ffb7d1-00de-436c-b9d6-b4d517479ad8\") " Nov 26 00:38:24 crc kubenswrapper[4766]: I1126 00:38:24.801189 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05ffb7d1-00de-436c-b9d6-b4d517479ad8-catalog-content\") pod \"05ffb7d1-00de-436c-b9d6-b4d517479ad8\" (UID: \"05ffb7d1-00de-436c-b9d6-b4d517479ad8\") " Nov 26 00:38:24 crc kubenswrapper[4766]: I1126 00:38:24.801214 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05ffb7d1-00de-436c-b9d6-b4d517479ad8-utilities\") pod \"05ffb7d1-00de-436c-b9d6-b4d517479ad8\" (UID: \"05ffb7d1-00de-436c-b9d6-b4d517479ad8\") " Nov 26 00:38:24 crc kubenswrapper[4766]: I1126 00:38:24.802323 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05ffb7d1-00de-436c-b9d6-b4d517479ad8-utilities" (OuterVolumeSpecName: "utilities") pod "05ffb7d1-00de-436c-b9d6-b4d517479ad8" (UID: "05ffb7d1-00de-436c-b9d6-b4d517479ad8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:38:24 crc kubenswrapper[4766]: I1126 00:38:24.807318 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05ffb7d1-00de-436c-b9d6-b4d517479ad8-kube-api-access-dqsxx" (OuterVolumeSpecName: "kube-api-access-dqsxx") pod "05ffb7d1-00de-436c-b9d6-b4d517479ad8" (UID: "05ffb7d1-00de-436c-b9d6-b4d517479ad8"). InnerVolumeSpecName "kube-api-access-dqsxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:38:24 crc kubenswrapper[4766]: I1126 00:38:24.853903 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05ffb7d1-00de-436c-b9d6-b4d517479ad8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05ffb7d1-00de-436c-b9d6-b4d517479ad8" (UID: "05ffb7d1-00de-436c-b9d6-b4d517479ad8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:38:24 crc kubenswrapper[4766]: I1126 00:38:24.902542 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqsxx\" (UniqueName: \"kubernetes.io/projected/05ffb7d1-00de-436c-b9d6-b4d517479ad8-kube-api-access-dqsxx\") on node \"crc\" DevicePath \"\"" Nov 26 00:38:24 crc kubenswrapper[4766]: I1126 00:38:24.902573 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05ffb7d1-00de-436c-b9d6-b4d517479ad8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:38:24 crc kubenswrapper[4766]: I1126 00:38:24.902583 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05ffb7d1-00de-436c-b9d6-b4d517479ad8-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:38:25 crc kubenswrapper[4766]: I1126 00:38:25.371268 4766 generic.go:334] "Generic (PLEG): container finished" podID="05ffb7d1-00de-436c-b9d6-b4d517479ad8" containerID="ca63a3089127a2d6718d812658f81c8a2fa2e7d40d45261cb92513e4e5479b5e" exitCode=0 Nov 26 00:38:25 crc kubenswrapper[4766]: I1126 00:38:25.371307 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vvl2" event={"ID":"05ffb7d1-00de-436c-b9d6-b4d517479ad8","Type":"ContainerDied","Data":"ca63a3089127a2d6718d812658f81c8a2fa2e7d40d45261cb92513e4e5479b5e"} Nov 26 00:38:25 crc kubenswrapper[4766]: I1126 00:38:25.371344 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vvl2" event={"ID":"05ffb7d1-00de-436c-b9d6-b4d517479ad8","Type":"ContainerDied","Data":"ba3258047d665d2674423f705d30bfd9bd4e1be91fd797b5094bb3fac718982f"} Nov 26 00:38:25 crc kubenswrapper[4766]: I1126 00:38:25.371361 4766 scope.go:117] "RemoveContainer" containerID="ca63a3089127a2d6718d812658f81c8a2fa2e7d40d45261cb92513e4e5479b5e" Nov 26 00:38:25 crc kubenswrapper[4766]: I1126 00:38:25.371478 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6vvl2" Nov 26 00:38:25 crc kubenswrapper[4766]: I1126 00:38:25.389970 4766 scope.go:117] "RemoveContainer" containerID="ee7e76c0751d9936743c1876c5e7e321195f5d63d6fbff4f240b99a3c5333a74" Nov 26 00:38:25 crc kubenswrapper[4766]: I1126 00:38:25.401448 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6vvl2"] Nov 26 00:38:25 crc kubenswrapper[4766]: I1126 00:38:25.406066 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6vvl2"] Nov 26 00:38:25 crc kubenswrapper[4766]: I1126 00:38:25.419715 4766 scope.go:117] "RemoveContainer" containerID="0d510036621ca419e6d33165a85ebc295de739aff535789980aa87eca8374d7a" Nov 26 00:38:25 crc kubenswrapper[4766]: I1126 00:38:25.435733 4766 scope.go:117] "RemoveContainer" containerID="ca63a3089127a2d6718d812658f81c8a2fa2e7d40d45261cb92513e4e5479b5e" Nov 26 00:38:25 crc kubenswrapper[4766]: E1126 00:38:25.436366 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca63a3089127a2d6718d812658f81c8a2fa2e7d40d45261cb92513e4e5479b5e\": container with ID starting with ca63a3089127a2d6718d812658f81c8a2fa2e7d40d45261cb92513e4e5479b5e not found: ID does not exist" containerID="ca63a3089127a2d6718d812658f81c8a2fa2e7d40d45261cb92513e4e5479b5e" Nov 26 00:38:25 crc kubenswrapper[4766]: I1126 00:38:25.436413 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca63a3089127a2d6718d812658f81c8a2fa2e7d40d45261cb92513e4e5479b5e"} err="failed to get container status \"ca63a3089127a2d6718d812658f81c8a2fa2e7d40d45261cb92513e4e5479b5e\": rpc error: code = NotFound desc = could not find container \"ca63a3089127a2d6718d812658f81c8a2fa2e7d40d45261cb92513e4e5479b5e\": container with ID starting with ca63a3089127a2d6718d812658f81c8a2fa2e7d40d45261cb92513e4e5479b5e not found: ID does not exist" Nov 26 00:38:25 crc kubenswrapper[4766]: I1126 00:38:25.436443 4766 scope.go:117] "RemoveContainer" containerID="ee7e76c0751d9936743c1876c5e7e321195f5d63d6fbff4f240b99a3c5333a74" Nov 26 00:38:25 crc kubenswrapper[4766]: E1126 00:38:25.437020 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee7e76c0751d9936743c1876c5e7e321195f5d63d6fbff4f240b99a3c5333a74\": container with ID starting with ee7e76c0751d9936743c1876c5e7e321195f5d63d6fbff4f240b99a3c5333a74 not found: ID does not exist" containerID="ee7e76c0751d9936743c1876c5e7e321195f5d63d6fbff4f240b99a3c5333a74" Nov 26 00:38:25 crc kubenswrapper[4766]: I1126 00:38:25.437049 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee7e76c0751d9936743c1876c5e7e321195f5d63d6fbff4f240b99a3c5333a74"} err="failed to get container status \"ee7e76c0751d9936743c1876c5e7e321195f5d63d6fbff4f240b99a3c5333a74\": rpc error: code = NotFound desc = could not find container \"ee7e76c0751d9936743c1876c5e7e321195f5d63d6fbff4f240b99a3c5333a74\": container with ID starting with ee7e76c0751d9936743c1876c5e7e321195f5d63d6fbff4f240b99a3c5333a74 not found: ID does not exist" Nov 26 00:38:25 crc kubenswrapper[4766]: I1126 00:38:25.437072 4766 scope.go:117] "RemoveContainer" containerID="0d510036621ca419e6d33165a85ebc295de739aff535789980aa87eca8374d7a" Nov 26 00:38:25 crc kubenswrapper[4766]: E1126 00:38:25.437845 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d510036621ca419e6d33165a85ebc295de739aff535789980aa87eca8374d7a\": container with ID starting with 0d510036621ca419e6d33165a85ebc295de739aff535789980aa87eca8374d7a not found: ID does not exist" containerID="0d510036621ca419e6d33165a85ebc295de739aff535789980aa87eca8374d7a" Nov 26 00:38:25 crc kubenswrapper[4766]: I1126 00:38:25.437881 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d510036621ca419e6d33165a85ebc295de739aff535789980aa87eca8374d7a"} err="failed to get container status \"0d510036621ca419e6d33165a85ebc295de739aff535789980aa87eca8374d7a\": rpc error: code = NotFound desc = could not find container \"0d510036621ca419e6d33165a85ebc295de739aff535789980aa87eca8374d7a\": container with ID starting with 0d510036621ca419e6d33165a85ebc295de739aff535789980aa87eca8374d7a not found: ID does not exist" Nov 26 00:38:25 crc kubenswrapper[4766]: I1126 00:38:25.836797 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05ffb7d1-00de-436c-b9d6-b4d517479ad8" path="/var/lib/kubelet/pods/05ffb7d1-00de-436c-b9d6-b4d517479ad8/volumes" Nov 26 00:38:41 crc kubenswrapper[4766]: I1126 00:38:41.479537 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:38:41 crc kubenswrapper[4766]: I1126 00:38:41.480084 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:38:44 crc kubenswrapper[4766]: I1126 00:38:44.366337 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm"] Nov 26 00:38:44 crc kubenswrapper[4766]: E1126 00:38:44.366941 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05ffb7d1-00de-436c-b9d6-b4d517479ad8" containerName="registry-server" Nov 26 00:38:44 crc kubenswrapper[4766]: I1126 00:38:44.366953 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="05ffb7d1-00de-436c-b9d6-b4d517479ad8" containerName="registry-server" Nov 26 00:38:44 crc kubenswrapper[4766]: E1126 00:38:44.366969 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05ffb7d1-00de-436c-b9d6-b4d517479ad8" containerName="extract-utilities" Nov 26 00:38:44 crc kubenswrapper[4766]: I1126 00:38:44.366977 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="05ffb7d1-00de-436c-b9d6-b4d517479ad8" containerName="extract-utilities" Nov 26 00:38:44 crc kubenswrapper[4766]: E1126 00:38:44.366990 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05ffb7d1-00de-436c-b9d6-b4d517479ad8" containerName="extract-content" Nov 26 00:38:44 crc kubenswrapper[4766]: I1126 00:38:44.366997 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="05ffb7d1-00de-436c-b9d6-b4d517479ad8" containerName="extract-content" Nov 26 00:38:44 crc kubenswrapper[4766]: I1126 00:38:44.367110 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="05ffb7d1-00de-436c-b9d6-b4d517479ad8" containerName="registry-server" Nov 26 00:38:44 crc kubenswrapper[4766]: I1126 00:38:44.368387 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm" Nov 26 00:38:44 crc kubenswrapper[4766]: I1126 00:38:44.370704 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 26 00:38:44 crc kubenswrapper[4766]: I1126 00:38:44.378763 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm"] Nov 26 00:38:44 crc kubenswrapper[4766]: I1126 00:38:44.419677 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9e6e8bab-7232-4619-a55d-95125dfaf6d0-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm\" (UID: \"9e6e8bab-7232-4619-a55d-95125dfaf6d0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm" Nov 26 00:38:44 crc kubenswrapper[4766]: I1126 00:38:44.419993 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9e6e8bab-7232-4619-a55d-95125dfaf6d0-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm\" (UID: \"9e6e8bab-7232-4619-a55d-95125dfaf6d0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm" Nov 26 00:38:44 crc kubenswrapper[4766]: I1126 00:38:44.420146 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t62nk\" (UniqueName: \"kubernetes.io/projected/9e6e8bab-7232-4619-a55d-95125dfaf6d0-kube-api-access-t62nk\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm\" (UID: \"9e6e8bab-7232-4619-a55d-95125dfaf6d0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm" Nov 26 00:38:44 crc kubenswrapper[4766]: I1126 00:38:44.522108 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t62nk\" (UniqueName: \"kubernetes.io/projected/9e6e8bab-7232-4619-a55d-95125dfaf6d0-kube-api-access-t62nk\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm\" (UID: \"9e6e8bab-7232-4619-a55d-95125dfaf6d0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm" Nov 26 00:38:44 crc kubenswrapper[4766]: I1126 00:38:44.522213 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9e6e8bab-7232-4619-a55d-95125dfaf6d0-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm\" (UID: \"9e6e8bab-7232-4619-a55d-95125dfaf6d0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm" Nov 26 00:38:44 crc kubenswrapper[4766]: I1126 00:38:44.522254 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9e6e8bab-7232-4619-a55d-95125dfaf6d0-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm\" (UID: \"9e6e8bab-7232-4619-a55d-95125dfaf6d0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm" Nov 26 00:38:44 crc kubenswrapper[4766]: I1126 00:38:44.522788 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9e6e8bab-7232-4619-a55d-95125dfaf6d0-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm\" (UID: \"9e6e8bab-7232-4619-a55d-95125dfaf6d0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm" Nov 26 00:38:44 crc kubenswrapper[4766]: I1126 00:38:44.522876 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9e6e8bab-7232-4619-a55d-95125dfaf6d0-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm\" (UID: \"9e6e8bab-7232-4619-a55d-95125dfaf6d0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm" Nov 26 00:38:44 crc kubenswrapper[4766]: I1126 00:38:44.542732 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t62nk\" (UniqueName: \"kubernetes.io/projected/9e6e8bab-7232-4619-a55d-95125dfaf6d0-kube-api-access-t62nk\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm\" (UID: \"9e6e8bab-7232-4619-a55d-95125dfaf6d0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm" Nov 26 00:38:44 crc kubenswrapper[4766]: I1126 00:38:44.708892 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm" Nov 26 00:38:45 crc kubenswrapper[4766]: I1126 00:38:45.178897 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm"] Nov 26 00:38:45 crc kubenswrapper[4766]: I1126 00:38:45.508658 4766 generic.go:334] "Generic (PLEG): container finished" podID="9e6e8bab-7232-4619-a55d-95125dfaf6d0" containerID="fc98bf06b05c3c7d74f274a0f82040b72b4279f9ead72dc610ac95b6c16ab04f" exitCode=0 Nov 26 00:38:45 crc kubenswrapper[4766]: I1126 00:38:45.508840 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm" event={"ID":"9e6e8bab-7232-4619-a55d-95125dfaf6d0","Type":"ContainerDied","Data":"fc98bf06b05c3c7d74f274a0f82040b72b4279f9ead72dc610ac95b6c16ab04f"} Nov 26 00:38:45 crc kubenswrapper[4766]: I1126 00:38:45.508996 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm" event={"ID":"9e6e8bab-7232-4619-a55d-95125dfaf6d0","Type":"ContainerStarted","Data":"7b2bd835b236a71461311c800546245bb71991eb4cd5d3cddc7ddb4ce15fac31"} Nov 26 00:38:45 crc kubenswrapper[4766]: I1126 00:38:45.510871 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 00:38:50 crc kubenswrapper[4766]: I1126 00:38:50.536133 4766 generic.go:334] "Generic (PLEG): container finished" podID="9e6e8bab-7232-4619-a55d-95125dfaf6d0" containerID="63e2e48982071f94167b9da77f96e2c3794c530a926a1b46a8742630955de2df" exitCode=0 Nov 26 00:38:50 crc kubenswrapper[4766]: I1126 00:38:50.536223 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm" event={"ID":"9e6e8bab-7232-4619-a55d-95125dfaf6d0","Type":"ContainerDied","Data":"63e2e48982071f94167b9da77f96e2c3794c530a926a1b46a8742630955de2df"} Nov 26 00:38:51 crc kubenswrapper[4766]: I1126 00:38:51.543347 4766 generic.go:334] "Generic (PLEG): container finished" podID="9e6e8bab-7232-4619-a55d-95125dfaf6d0" containerID="47471ca88040542d77f14fa301095ec182195725c8fcb6019fc363b2e0577413" exitCode=0 Nov 26 00:38:51 crc kubenswrapper[4766]: I1126 00:38:51.543537 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm" event={"ID":"9e6e8bab-7232-4619-a55d-95125dfaf6d0","Type":"ContainerDied","Data":"47471ca88040542d77f14fa301095ec182195725c8fcb6019fc363b2e0577413"} Nov 26 00:38:52 crc kubenswrapper[4766]: I1126 00:38:52.824115 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm" Nov 26 00:38:52 crc kubenswrapper[4766]: I1126 00:38:52.939644 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9e6e8bab-7232-4619-a55d-95125dfaf6d0-bundle\") pod \"9e6e8bab-7232-4619-a55d-95125dfaf6d0\" (UID: \"9e6e8bab-7232-4619-a55d-95125dfaf6d0\") " Nov 26 00:38:52 crc kubenswrapper[4766]: I1126 00:38:52.939771 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9e6e8bab-7232-4619-a55d-95125dfaf6d0-util\") pod \"9e6e8bab-7232-4619-a55d-95125dfaf6d0\" (UID: \"9e6e8bab-7232-4619-a55d-95125dfaf6d0\") " Nov 26 00:38:52 crc kubenswrapper[4766]: I1126 00:38:52.939817 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t62nk\" (UniqueName: \"kubernetes.io/projected/9e6e8bab-7232-4619-a55d-95125dfaf6d0-kube-api-access-t62nk\") pod \"9e6e8bab-7232-4619-a55d-95125dfaf6d0\" (UID: \"9e6e8bab-7232-4619-a55d-95125dfaf6d0\") " Nov 26 00:38:52 crc kubenswrapper[4766]: I1126 00:38:52.941066 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e6e8bab-7232-4619-a55d-95125dfaf6d0-bundle" (OuterVolumeSpecName: "bundle") pod "9e6e8bab-7232-4619-a55d-95125dfaf6d0" (UID: "9e6e8bab-7232-4619-a55d-95125dfaf6d0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:38:52 crc kubenswrapper[4766]: I1126 00:38:52.945825 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e6e8bab-7232-4619-a55d-95125dfaf6d0-kube-api-access-t62nk" (OuterVolumeSpecName: "kube-api-access-t62nk") pod "9e6e8bab-7232-4619-a55d-95125dfaf6d0" (UID: "9e6e8bab-7232-4619-a55d-95125dfaf6d0"). InnerVolumeSpecName "kube-api-access-t62nk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:38:52 crc kubenswrapper[4766]: I1126 00:38:52.950188 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e6e8bab-7232-4619-a55d-95125dfaf6d0-util" (OuterVolumeSpecName: "util") pod "9e6e8bab-7232-4619-a55d-95125dfaf6d0" (UID: "9e6e8bab-7232-4619-a55d-95125dfaf6d0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:38:53 crc kubenswrapper[4766]: I1126 00:38:53.042012 4766 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9e6e8bab-7232-4619-a55d-95125dfaf6d0-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:38:53 crc kubenswrapper[4766]: I1126 00:38:53.042043 4766 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9e6e8bab-7232-4619-a55d-95125dfaf6d0-util\") on node \"crc\" DevicePath \"\"" Nov 26 00:38:53 crc kubenswrapper[4766]: I1126 00:38:53.042053 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t62nk\" (UniqueName: \"kubernetes.io/projected/9e6e8bab-7232-4619-a55d-95125dfaf6d0-kube-api-access-t62nk\") on node \"crc\" DevicePath \"\"" Nov 26 00:38:53 crc kubenswrapper[4766]: I1126 00:38:53.558192 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm" event={"ID":"9e6e8bab-7232-4619-a55d-95125dfaf6d0","Type":"ContainerDied","Data":"7b2bd835b236a71461311c800546245bb71991eb4cd5d3cddc7ddb4ce15fac31"} Nov 26 00:38:53 crc kubenswrapper[4766]: I1126 00:38:53.558243 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b2bd835b236a71461311c800546245bb71991eb4cd5d3cddc7ddb4ce15fac31" Nov 26 00:38:53 crc kubenswrapper[4766]: I1126 00:38:53.558245 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm" Nov 26 00:38:56 crc kubenswrapper[4766]: I1126 00:38:56.232744 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-ztjbb"] Nov 26 00:38:56 crc kubenswrapper[4766]: E1126 00:38:56.233362 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6e8bab-7232-4619-a55d-95125dfaf6d0" containerName="extract" Nov 26 00:38:56 crc kubenswrapper[4766]: I1126 00:38:56.233381 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6e8bab-7232-4619-a55d-95125dfaf6d0" containerName="extract" Nov 26 00:38:56 crc kubenswrapper[4766]: E1126 00:38:56.233399 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6e8bab-7232-4619-a55d-95125dfaf6d0" containerName="pull" Nov 26 00:38:56 crc kubenswrapper[4766]: I1126 00:38:56.233407 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6e8bab-7232-4619-a55d-95125dfaf6d0" containerName="pull" Nov 26 00:38:56 crc kubenswrapper[4766]: E1126 00:38:56.233430 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6e8bab-7232-4619-a55d-95125dfaf6d0" containerName="util" Nov 26 00:38:56 crc kubenswrapper[4766]: I1126 00:38:56.233438 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6e8bab-7232-4619-a55d-95125dfaf6d0" containerName="util" Nov 26 00:38:56 crc kubenswrapper[4766]: I1126 00:38:56.233591 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e6e8bab-7232-4619-a55d-95125dfaf6d0" containerName="extract" Nov 26 00:38:56 crc kubenswrapper[4766]: I1126 00:38:56.234224 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-ztjbb" Nov 26 00:38:56 crc kubenswrapper[4766]: I1126 00:38:56.238413 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 26 00:38:56 crc kubenswrapper[4766]: I1126 00:38:56.238897 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 26 00:38:56 crc kubenswrapper[4766]: I1126 00:38:56.239154 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-rd2kk" Nov 26 00:38:56 crc kubenswrapper[4766]: I1126 00:38:56.252955 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-ztjbb"] Nov 26 00:38:56 crc kubenswrapper[4766]: I1126 00:38:56.287773 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdjbg\" (UniqueName: \"kubernetes.io/projected/61f41116-4bc2-47c5-b805-d0a0990c6024-kube-api-access-xdjbg\") pod \"nmstate-operator-557fdffb88-ztjbb\" (UID: \"61f41116-4bc2-47c5-b805-d0a0990c6024\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-ztjbb" Nov 26 00:38:56 crc kubenswrapper[4766]: I1126 00:38:56.389099 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdjbg\" (UniqueName: \"kubernetes.io/projected/61f41116-4bc2-47c5-b805-d0a0990c6024-kube-api-access-xdjbg\") pod \"nmstate-operator-557fdffb88-ztjbb\" (UID: \"61f41116-4bc2-47c5-b805-d0a0990c6024\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-ztjbb" Nov 26 00:38:56 crc kubenswrapper[4766]: I1126 00:38:56.405940 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdjbg\" (UniqueName: \"kubernetes.io/projected/61f41116-4bc2-47c5-b805-d0a0990c6024-kube-api-access-xdjbg\") pod \"nmstate-operator-557fdffb88-ztjbb\" (UID: \"61f41116-4bc2-47c5-b805-d0a0990c6024\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-ztjbb" Nov 26 00:38:56 crc kubenswrapper[4766]: I1126 00:38:56.552717 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-ztjbb" Nov 26 00:38:57 crc kubenswrapper[4766]: I1126 00:38:57.081111 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-ztjbb"] Nov 26 00:38:57 crc kubenswrapper[4766]: I1126 00:38:57.584769 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-ztjbb" event={"ID":"61f41116-4bc2-47c5-b805-d0a0990c6024","Type":"ContainerStarted","Data":"ad5654a884a20ff548935bf0ff2bac61381de282a874addf61a2e8757b454a28"} Nov 26 00:38:59 crc kubenswrapper[4766]: I1126 00:38:59.598800 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-ztjbb" event={"ID":"61f41116-4bc2-47c5-b805-d0a0990c6024","Type":"ContainerStarted","Data":"4e4172b8310a60aefb54e4ceca60fff9e988a23940af2bd527ec6c6775b9ab05"} Nov 26 00:38:59 crc kubenswrapper[4766]: I1126 00:38:59.616282 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-ztjbb" podStartSLOduration=1.478200251 podStartE2EDuration="3.616255924s" podCreationTimestamp="2025-11-26 00:38:56 +0000 UTC" firstStartedPulling="2025-11-26 00:38:57.089705797 +0000 UTC m=+917.938476227" lastFinishedPulling="2025-11-26 00:38:59.22776147 +0000 UTC m=+920.076531900" observedRunningTime="2025-11-26 00:38:59.613585637 +0000 UTC m=+920.462356077" watchObservedRunningTime="2025-11-26 00:38:59.616255924 +0000 UTC m=+920.465026354" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.455773 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-f58l6"] Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.457057 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-f58l6" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.463312 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-mn64d"] Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.464332 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-6xjwq" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.464459 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-mn64d" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.467318 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.480038 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-f58l6"] Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.488361 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-mn64d"] Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.494012 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-q4sb9"] Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.495047 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-q4sb9" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.550097 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/58a75814-134a-4a52-aae0-5c21109f658f-dbus-socket\") pod \"nmstate-handler-q4sb9\" (UID: \"58a75814-134a-4a52-aae0-5c21109f658f\") " pod="openshift-nmstate/nmstate-handler-q4sb9" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.550466 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/58a75814-134a-4a52-aae0-5c21109f658f-ovs-socket\") pod \"nmstate-handler-q4sb9\" (UID: \"58a75814-134a-4a52-aae0-5c21109f658f\") " pod="openshift-nmstate/nmstate-handler-q4sb9" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.550557 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmhm5\" (UniqueName: \"kubernetes.io/projected/58a75814-134a-4a52-aae0-5c21109f658f-kube-api-access-wmhm5\") pod \"nmstate-handler-q4sb9\" (UID: \"58a75814-134a-4a52-aae0-5c21109f658f\") " pod="openshift-nmstate/nmstate-handler-q4sb9" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.550602 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/d874f3cf-9f30-431e-81e1-6334a0cf4351-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-mn64d\" (UID: \"d874f3cf-9f30-431e-81e1-6334a0cf4351\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-mn64d" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.550643 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/58a75814-134a-4a52-aae0-5c21109f658f-nmstate-lock\") pod \"nmstate-handler-q4sb9\" (UID: \"58a75814-134a-4a52-aae0-5c21109f658f\") " pod="openshift-nmstate/nmstate-handler-q4sb9" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.550685 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tj8z7\" (UniqueName: \"kubernetes.io/projected/d874f3cf-9f30-431e-81e1-6334a0cf4351-kube-api-access-tj8z7\") pod \"nmstate-webhook-6b89b748d8-mn64d\" (UID: \"d874f3cf-9f30-431e-81e1-6334a0cf4351\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-mn64d" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.550704 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhvpf\" (UniqueName: \"kubernetes.io/projected/8f6f1e2d-6ebd-4f98-bc08-8cba715f33ae-kube-api-access-rhvpf\") pod \"nmstate-metrics-5dcf9c57c5-f58l6\" (UID: \"8f6f1e2d-6ebd-4f98-bc08-8cba715f33ae\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-f58l6" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.627380 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nzhrs"] Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.628620 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nzhrs" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.636124 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.636330 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.636496 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-tf6mh" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.651673 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-nzhrs\" (UID: \"d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nzhrs" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.651713 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-nzhrs\" (UID: \"d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nzhrs" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.651733 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkmp7\" (UniqueName: \"kubernetes.io/projected/d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130-kube-api-access-fkmp7\") pod \"nmstate-console-plugin-5874bd7bc5-nzhrs\" (UID: \"d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nzhrs" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.651776 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmhm5\" (UniqueName: \"kubernetes.io/projected/58a75814-134a-4a52-aae0-5c21109f658f-kube-api-access-wmhm5\") pod \"nmstate-handler-q4sb9\" (UID: \"58a75814-134a-4a52-aae0-5c21109f658f\") " pod="openshift-nmstate/nmstate-handler-q4sb9" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.651807 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/d874f3cf-9f30-431e-81e1-6334a0cf4351-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-mn64d\" (UID: \"d874f3cf-9f30-431e-81e1-6334a0cf4351\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-mn64d" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.651833 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/58a75814-134a-4a52-aae0-5c21109f658f-nmstate-lock\") pod \"nmstate-handler-q4sb9\" (UID: \"58a75814-134a-4a52-aae0-5c21109f658f\") " pod="openshift-nmstate/nmstate-handler-q4sb9" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.651848 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tj8z7\" (UniqueName: \"kubernetes.io/projected/d874f3cf-9f30-431e-81e1-6334a0cf4351-kube-api-access-tj8z7\") pod \"nmstate-webhook-6b89b748d8-mn64d\" (UID: \"d874f3cf-9f30-431e-81e1-6334a0cf4351\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-mn64d" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.651863 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhvpf\" (UniqueName: \"kubernetes.io/projected/8f6f1e2d-6ebd-4f98-bc08-8cba715f33ae-kube-api-access-rhvpf\") pod \"nmstate-metrics-5dcf9c57c5-f58l6\" (UID: \"8f6f1e2d-6ebd-4f98-bc08-8cba715f33ae\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-f58l6" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.651928 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/58a75814-134a-4a52-aae0-5c21109f658f-dbus-socket\") pod \"nmstate-handler-q4sb9\" (UID: \"58a75814-134a-4a52-aae0-5c21109f658f\") " pod="openshift-nmstate/nmstate-handler-q4sb9" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.651947 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/58a75814-134a-4a52-aae0-5c21109f658f-ovs-socket\") pod \"nmstate-handler-q4sb9\" (UID: \"58a75814-134a-4a52-aae0-5c21109f658f\") " pod="openshift-nmstate/nmstate-handler-q4sb9" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.652010 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/58a75814-134a-4a52-aae0-5c21109f658f-ovs-socket\") pod \"nmstate-handler-q4sb9\" (UID: \"58a75814-134a-4a52-aae0-5c21109f658f\") " pod="openshift-nmstate/nmstate-handler-q4sb9" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.652140 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nzhrs"] Nov 26 00:39:04 crc kubenswrapper[4766]: E1126 00:39:04.652478 4766 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.652560 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/58a75814-134a-4a52-aae0-5c21109f658f-nmstate-lock\") pod \"nmstate-handler-q4sb9\" (UID: \"58a75814-134a-4a52-aae0-5c21109f658f\") " pod="openshift-nmstate/nmstate-handler-q4sb9" Nov 26 00:39:04 crc kubenswrapper[4766]: E1126 00:39:04.652610 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d874f3cf-9f30-431e-81e1-6334a0cf4351-tls-key-pair podName:d874f3cf-9f30-431e-81e1-6334a0cf4351 nodeName:}" failed. No retries permitted until 2025-11-26 00:39:05.152505061 +0000 UTC m=+926.001275491 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/d874f3cf-9f30-431e-81e1-6334a0cf4351-tls-key-pair") pod "nmstate-webhook-6b89b748d8-mn64d" (UID: "d874f3cf-9f30-431e-81e1-6334a0cf4351") : secret "openshift-nmstate-webhook" not found Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.652668 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/58a75814-134a-4a52-aae0-5c21109f658f-dbus-socket\") pod \"nmstate-handler-q4sb9\" (UID: \"58a75814-134a-4a52-aae0-5c21109f658f\") " pod="openshift-nmstate/nmstate-handler-q4sb9" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.675983 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmhm5\" (UniqueName: \"kubernetes.io/projected/58a75814-134a-4a52-aae0-5c21109f658f-kube-api-access-wmhm5\") pod \"nmstate-handler-q4sb9\" (UID: \"58a75814-134a-4a52-aae0-5c21109f658f\") " pod="openshift-nmstate/nmstate-handler-q4sb9" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.682641 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhvpf\" (UniqueName: \"kubernetes.io/projected/8f6f1e2d-6ebd-4f98-bc08-8cba715f33ae-kube-api-access-rhvpf\") pod \"nmstate-metrics-5dcf9c57c5-f58l6\" (UID: \"8f6f1e2d-6ebd-4f98-bc08-8cba715f33ae\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-f58l6" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.684154 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tj8z7\" (UniqueName: \"kubernetes.io/projected/d874f3cf-9f30-431e-81e1-6334a0cf4351-kube-api-access-tj8z7\") pod \"nmstate-webhook-6b89b748d8-mn64d\" (UID: \"d874f3cf-9f30-431e-81e1-6334a0cf4351\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-mn64d" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.754069 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-nzhrs\" (UID: \"d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nzhrs" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.754110 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-nzhrs\" (UID: \"d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nzhrs" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.754133 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkmp7\" (UniqueName: \"kubernetes.io/projected/d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130-kube-api-access-fkmp7\") pod \"nmstate-console-plugin-5874bd7bc5-nzhrs\" (UID: \"d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nzhrs" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.754887 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-nzhrs\" (UID: \"d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nzhrs" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.757967 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-nzhrs\" (UID: \"d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nzhrs" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.778846 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkmp7\" (UniqueName: \"kubernetes.io/projected/d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130-kube-api-access-fkmp7\") pod \"nmstate-console-plugin-5874bd7bc5-nzhrs\" (UID: \"d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nzhrs" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.799539 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-f58l6" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.825345 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5bc7576dcd-sqbpf"] Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.826401 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.837952 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-q4sb9" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.848422 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5bc7576dcd-sqbpf"] Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.856774 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbh7t\" (UniqueName: \"kubernetes.io/projected/e1ca8391-338e-410b-a7ce-83da6f733c87-kube-api-access-sbh7t\") pod \"console-5bc7576dcd-sqbpf\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.856829 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-trusted-ca-bundle\") pod \"console-5bc7576dcd-sqbpf\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.856872 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-oauth-serving-cert\") pod \"console-5bc7576dcd-sqbpf\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.856911 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e1ca8391-338e-410b-a7ce-83da6f733c87-console-oauth-config\") pod \"console-5bc7576dcd-sqbpf\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.857801 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-console-config\") pod \"console-5bc7576dcd-sqbpf\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.857832 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e1ca8391-338e-410b-a7ce-83da6f733c87-console-serving-cert\") pod \"console-5bc7576dcd-sqbpf\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.857899 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-service-ca\") pod \"console-5bc7576dcd-sqbpf\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.959569 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbh7t\" (UniqueName: \"kubernetes.io/projected/e1ca8391-338e-410b-a7ce-83da6f733c87-kube-api-access-sbh7t\") pod \"console-5bc7576dcd-sqbpf\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.959947 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-trusted-ca-bundle\") pod \"console-5bc7576dcd-sqbpf\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.959996 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-oauth-serving-cert\") pod \"console-5bc7576dcd-sqbpf\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.960032 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e1ca8391-338e-410b-a7ce-83da6f733c87-console-oauth-config\") pod \"console-5bc7576dcd-sqbpf\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.960077 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-console-config\") pod \"console-5bc7576dcd-sqbpf\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.960104 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e1ca8391-338e-410b-a7ce-83da6f733c87-console-serving-cert\") pod \"console-5bc7576dcd-sqbpf\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.960136 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-service-ca\") pod \"console-5bc7576dcd-sqbpf\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.961297 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-service-ca\") pod \"console-5bc7576dcd-sqbpf\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.962477 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-trusted-ca-bundle\") pod \"console-5bc7576dcd-sqbpf\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.964465 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-console-config\") pod \"console-5bc7576dcd-sqbpf\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.965284 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-oauth-serving-cert\") pod \"console-5bc7576dcd-sqbpf\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.966191 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nzhrs" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.973028 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e1ca8391-338e-410b-a7ce-83da6f733c87-console-serving-cert\") pod \"console-5bc7576dcd-sqbpf\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.974959 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e1ca8391-338e-410b-a7ce-83da6f733c87-console-oauth-config\") pod \"console-5bc7576dcd-sqbpf\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:04 crc kubenswrapper[4766]: I1126 00:39:04.984621 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbh7t\" (UniqueName: \"kubernetes.io/projected/e1ca8391-338e-410b-a7ce-83da6f733c87-kube-api-access-sbh7t\") pod \"console-5bc7576dcd-sqbpf\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:05 crc kubenswrapper[4766]: I1126 00:39:05.165437 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/d874f3cf-9f30-431e-81e1-6334a0cf4351-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-mn64d\" (UID: \"d874f3cf-9f30-431e-81e1-6334a0cf4351\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-mn64d" Nov 26 00:39:05 crc kubenswrapper[4766]: I1126 00:39:05.175350 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/d874f3cf-9f30-431e-81e1-6334a0cf4351-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-mn64d\" (UID: \"d874f3cf-9f30-431e-81e1-6334a0cf4351\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-mn64d" Nov 26 00:39:05 crc kubenswrapper[4766]: I1126 00:39:05.178064 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:05 crc kubenswrapper[4766]: I1126 00:39:05.295156 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-f58l6"] Nov 26 00:39:05 crc kubenswrapper[4766]: W1126 00:39:05.326257 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f6f1e2d_6ebd_4f98_bc08_8cba715f33ae.slice/crio-eb0c5d2925534bb5a2a8df4c238033eb02857857791f55cd59db412a6d112fcc WatchSource:0}: Error finding container eb0c5d2925534bb5a2a8df4c238033eb02857857791f55cd59db412a6d112fcc: Status 404 returned error can't find the container with id eb0c5d2925534bb5a2a8df4c238033eb02857857791f55cd59db412a6d112fcc Nov 26 00:39:05 crc kubenswrapper[4766]: I1126 00:39:05.396608 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nzhrs"] Nov 26 00:39:05 crc kubenswrapper[4766]: I1126 00:39:05.420852 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-mn64d" Nov 26 00:39:05 crc kubenswrapper[4766]: I1126 00:39:05.641796 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nzhrs" event={"ID":"d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130","Type":"ContainerStarted","Data":"61f584fe0d8f0bbbfd16f7fa83ba720b6316d34cb9e268c121a17e785ccd0a7c"} Nov 26 00:39:05 crc kubenswrapper[4766]: I1126 00:39:05.642895 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-q4sb9" event={"ID":"58a75814-134a-4a52-aae0-5c21109f658f","Type":"ContainerStarted","Data":"cd80ab686aeaba0353e14adf8978365d72d3135c222dfbe62a282afbe944de72"} Nov 26 00:39:05 crc kubenswrapper[4766]: I1126 00:39:05.643937 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-f58l6" event={"ID":"8f6f1e2d-6ebd-4f98-bc08-8cba715f33ae","Type":"ContainerStarted","Data":"eb0c5d2925534bb5a2a8df4c238033eb02857857791f55cd59db412a6d112fcc"} Nov 26 00:39:05 crc kubenswrapper[4766]: I1126 00:39:05.677059 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5bc7576dcd-sqbpf"] Nov 26 00:39:05 crc kubenswrapper[4766]: W1126 00:39:05.685216 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1ca8391_338e_410b_a7ce_83da6f733c87.slice/crio-b0ed3c9a2e9e49fb3d62aaee91a75aabfe7d60df162036f05febc5d2766942f4 WatchSource:0}: Error finding container b0ed3c9a2e9e49fb3d62aaee91a75aabfe7d60df162036f05febc5d2766942f4: Status 404 returned error can't find the container with id b0ed3c9a2e9e49fb3d62aaee91a75aabfe7d60df162036f05febc5d2766942f4 Nov 26 00:39:05 crc kubenswrapper[4766]: I1126 00:39:05.835234 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-mn64d"] Nov 26 00:39:05 crc kubenswrapper[4766]: W1126 00:39:05.836269 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd874f3cf_9f30_431e_81e1_6334a0cf4351.slice/crio-abbe87cfab7553093c9ca060a9e165f09bcf63460a64bd66c3721cf4f5074401 WatchSource:0}: Error finding container abbe87cfab7553093c9ca060a9e165f09bcf63460a64bd66c3721cf4f5074401: Status 404 returned error can't find the container with id abbe87cfab7553093c9ca060a9e165f09bcf63460a64bd66c3721cf4f5074401 Nov 26 00:39:06 crc kubenswrapper[4766]: I1126 00:39:06.653754 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5bc7576dcd-sqbpf" event={"ID":"e1ca8391-338e-410b-a7ce-83da6f733c87","Type":"ContainerStarted","Data":"ece9fdfedc40fbd881d5f91cfbe1c0398a1f91bdecc08911f99a8d8e6be1e380"} Nov 26 00:39:06 crc kubenswrapper[4766]: I1126 00:39:06.654061 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5bc7576dcd-sqbpf" event={"ID":"e1ca8391-338e-410b-a7ce-83da6f733c87","Type":"ContainerStarted","Data":"b0ed3c9a2e9e49fb3d62aaee91a75aabfe7d60df162036f05febc5d2766942f4"} Nov 26 00:39:06 crc kubenswrapper[4766]: I1126 00:39:06.655353 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-mn64d" event={"ID":"d874f3cf-9f30-431e-81e1-6334a0cf4351","Type":"ContainerStarted","Data":"abbe87cfab7553093c9ca060a9e165f09bcf63460a64bd66c3721cf4f5074401"} Nov 26 00:39:06 crc kubenswrapper[4766]: I1126 00:39:06.678466 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5bc7576dcd-sqbpf" podStartSLOduration=2.67844849 podStartE2EDuration="2.67844849s" podCreationTimestamp="2025-11-26 00:39:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:39:06.67247162 +0000 UTC m=+927.521242060" watchObservedRunningTime="2025-11-26 00:39:06.67844849 +0000 UTC m=+927.527218920" Nov 26 00:39:11 crc kubenswrapper[4766]: I1126 00:39:11.479458 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:39:11 crc kubenswrapper[4766]: I1126 00:39:11.479955 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:39:11 crc kubenswrapper[4766]: I1126 00:39:11.694264 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-q4sb9" event={"ID":"58a75814-134a-4a52-aae0-5c21109f658f","Type":"ContainerStarted","Data":"fc2432bee7b311c825efca38c70924cae700306a69554eb940992a219c55bbde"} Nov 26 00:39:11 crc kubenswrapper[4766]: I1126 00:39:11.694596 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-q4sb9" Nov 26 00:39:11 crc kubenswrapper[4766]: I1126 00:39:11.695767 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-f58l6" event={"ID":"8f6f1e2d-6ebd-4f98-bc08-8cba715f33ae","Type":"ContainerStarted","Data":"01c587f7733bc59ad26abe490a6c9ac8507a0924d97d8a4f574481b1f98ee712"} Nov 26 00:39:11 crc kubenswrapper[4766]: I1126 00:39:11.697245 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nzhrs" event={"ID":"d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130","Type":"ContainerStarted","Data":"d80f0a012c397f9f6fa85bc22637b0ed6a689fa5056ee4ebcc115650028ca042"} Nov 26 00:39:11 crc kubenswrapper[4766]: I1126 00:39:11.698932 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-mn64d" event={"ID":"d874f3cf-9f30-431e-81e1-6334a0cf4351","Type":"ContainerStarted","Data":"22107c1e98f58e88b4f38fc262f9c992e23d8b3c770a9def84e82c32203e9b3f"} Nov 26 00:39:11 crc kubenswrapper[4766]: I1126 00:39:11.699590 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-mn64d" Nov 26 00:39:11 crc kubenswrapper[4766]: I1126 00:39:11.709382 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-q4sb9" podStartSLOduration=1.686542521 podStartE2EDuration="7.709363953s" podCreationTimestamp="2025-11-26 00:39:04 +0000 UTC" firstStartedPulling="2025-11-26 00:39:04.886903494 +0000 UTC m=+925.735673924" lastFinishedPulling="2025-11-26 00:39:10.909724906 +0000 UTC m=+931.758495356" observedRunningTime="2025-11-26 00:39:11.708139012 +0000 UTC m=+932.556909442" watchObservedRunningTime="2025-11-26 00:39:11.709363953 +0000 UTC m=+932.558134383" Nov 26 00:39:11 crc kubenswrapper[4766]: I1126 00:39:11.724227 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-mn64d" podStartSLOduration=2.612154931 podStartE2EDuration="7.724210577s" podCreationTimestamp="2025-11-26 00:39:04 +0000 UTC" firstStartedPulling="2025-11-26 00:39:05.838728224 +0000 UTC m=+926.687498654" lastFinishedPulling="2025-11-26 00:39:10.95078387 +0000 UTC m=+931.799554300" observedRunningTime="2025-11-26 00:39:11.722008881 +0000 UTC m=+932.570779311" watchObservedRunningTime="2025-11-26 00:39:11.724210577 +0000 UTC m=+932.572981007" Nov 26 00:39:11 crc kubenswrapper[4766]: I1126 00:39:11.737919 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nzhrs" podStartSLOduration=2.259994581 podStartE2EDuration="7.737896011s" podCreationTimestamp="2025-11-26 00:39:04 +0000 UTC" firstStartedPulling="2025-11-26 00:39:05.409046803 +0000 UTC m=+926.257817233" lastFinishedPulling="2025-11-26 00:39:10.886948243 +0000 UTC m=+931.735718663" observedRunningTime="2025-11-26 00:39:11.735948402 +0000 UTC m=+932.584718832" watchObservedRunningTime="2025-11-26 00:39:11.737896011 +0000 UTC m=+932.586666441" Nov 26 00:39:13 crc kubenswrapper[4766]: I1126 00:39:13.714197 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-f58l6" event={"ID":"8f6f1e2d-6ebd-4f98-bc08-8cba715f33ae","Type":"ContainerStarted","Data":"e8ec5008aebd974b5c8c8425b32160feae237709de46b32eb19f3e47eba4ab65"} Nov 26 00:39:13 crc kubenswrapper[4766]: I1126 00:39:13.731980 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-f58l6" podStartSLOduration=1.620511928 podStartE2EDuration="9.731962708s" podCreationTimestamp="2025-11-26 00:39:04 +0000 UTC" firstStartedPulling="2025-11-26 00:39:05.345718028 +0000 UTC m=+926.194488458" lastFinishedPulling="2025-11-26 00:39:13.457168808 +0000 UTC m=+934.305939238" observedRunningTime="2025-11-26 00:39:13.729797784 +0000 UTC m=+934.578568204" watchObservedRunningTime="2025-11-26 00:39:13.731962708 +0000 UTC m=+934.580733128" Nov 26 00:39:15 crc kubenswrapper[4766]: I1126 00:39:15.179546 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:15 crc kubenswrapper[4766]: I1126 00:39:15.179606 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:15 crc kubenswrapper[4766]: I1126 00:39:15.184017 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:15 crc kubenswrapper[4766]: I1126 00:39:15.732432 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:39:15 crc kubenswrapper[4766]: I1126 00:39:15.790191 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-hqj86"] Nov 26 00:39:19 crc kubenswrapper[4766]: I1126 00:39:19.860279 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-q4sb9" Nov 26 00:39:25 crc kubenswrapper[4766]: I1126 00:39:25.428265 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-mn64d" Nov 26 00:39:39 crc kubenswrapper[4766]: I1126 00:39:39.104240 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk"] Nov 26 00:39:39 crc kubenswrapper[4766]: I1126 00:39:39.107597 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk" Nov 26 00:39:39 crc kubenswrapper[4766]: I1126 00:39:39.112198 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 26 00:39:39 crc kubenswrapper[4766]: I1126 00:39:39.114774 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk"] Nov 26 00:39:39 crc kubenswrapper[4766]: I1126 00:39:39.272290 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzq76\" (UniqueName: \"kubernetes.io/projected/631bfaa9-4706-44f7-9417-500418a18c73-kube-api-access-nzq76\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk\" (UID: \"631bfaa9-4706-44f7-9417-500418a18c73\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk" Nov 26 00:39:39 crc kubenswrapper[4766]: I1126 00:39:39.272714 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/631bfaa9-4706-44f7-9417-500418a18c73-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk\" (UID: \"631bfaa9-4706-44f7-9417-500418a18c73\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk" Nov 26 00:39:39 crc kubenswrapper[4766]: I1126 00:39:39.272810 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/631bfaa9-4706-44f7-9417-500418a18c73-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk\" (UID: \"631bfaa9-4706-44f7-9417-500418a18c73\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk" Nov 26 00:39:39 crc kubenswrapper[4766]: I1126 00:39:39.373695 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzq76\" (UniqueName: \"kubernetes.io/projected/631bfaa9-4706-44f7-9417-500418a18c73-kube-api-access-nzq76\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk\" (UID: \"631bfaa9-4706-44f7-9417-500418a18c73\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk" Nov 26 00:39:39 crc kubenswrapper[4766]: I1126 00:39:39.373779 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/631bfaa9-4706-44f7-9417-500418a18c73-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk\" (UID: \"631bfaa9-4706-44f7-9417-500418a18c73\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk" Nov 26 00:39:39 crc kubenswrapper[4766]: I1126 00:39:39.373843 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/631bfaa9-4706-44f7-9417-500418a18c73-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk\" (UID: \"631bfaa9-4706-44f7-9417-500418a18c73\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk" Nov 26 00:39:39 crc kubenswrapper[4766]: I1126 00:39:39.374353 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/631bfaa9-4706-44f7-9417-500418a18c73-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk\" (UID: \"631bfaa9-4706-44f7-9417-500418a18c73\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk" Nov 26 00:39:39 crc kubenswrapper[4766]: I1126 00:39:39.374349 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/631bfaa9-4706-44f7-9417-500418a18c73-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk\" (UID: \"631bfaa9-4706-44f7-9417-500418a18c73\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk" Nov 26 00:39:39 crc kubenswrapper[4766]: I1126 00:39:39.404921 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzq76\" (UniqueName: \"kubernetes.io/projected/631bfaa9-4706-44f7-9417-500418a18c73-kube-api-access-nzq76\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk\" (UID: \"631bfaa9-4706-44f7-9417-500418a18c73\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk" Nov 26 00:39:39 crc kubenswrapper[4766]: I1126 00:39:39.430281 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk" Nov 26 00:39:39 crc kubenswrapper[4766]: I1126 00:39:39.862896 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk"] Nov 26 00:39:39 crc kubenswrapper[4766]: I1126 00:39:39.941456 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk" event={"ID":"631bfaa9-4706-44f7-9417-500418a18c73","Type":"ContainerStarted","Data":"535ff060216bed91d283920447aa8c3b473ab28edadf41ee7905bf30aa4fb520"} Nov 26 00:39:40 crc kubenswrapper[4766]: I1126 00:39:40.824030 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-hqj86" podUID="e4d6938a-1f6b-4b90-8c85-4616c4def2e5" containerName="console" containerID="cri-o://ad32a7197fd193373afa7a36a53334c699a74cad5295cb9b7674cfaa44b97768" gracePeriod=15 Nov 26 00:39:40 crc kubenswrapper[4766]: I1126 00:39:40.948710 4766 generic.go:334] "Generic (PLEG): container finished" podID="631bfaa9-4706-44f7-9417-500418a18c73" containerID="8820e7a148f0fdb76d672fdb7c6941cce2507338a915f258b8eee3216dfe37aa" exitCode=0 Nov 26 00:39:40 crc kubenswrapper[4766]: I1126 00:39:40.948751 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk" event={"ID":"631bfaa9-4706-44f7-9417-500418a18c73","Type":"ContainerDied","Data":"8820e7a148f0fdb76d672fdb7c6941cce2507338a915f258b8eee3216dfe37aa"} Nov 26 00:39:40 crc kubenswrapper[4766]: I1126 00:39:40.957455 4766 patch_prober.go:28] interesting pod/console-f9d7485db-hqj86 container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Nov 26 00:39:40 crc kubenswrapper[4766]: I1126 00:39:40.957499 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-f9d7485db-hqj86" podUID="e4d6938a-1f6b-4b90-8c85-4616c4def2e5" containerName="console" probeResult="failure" output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.196692 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-hqj86_e4d6938a-1f6b-4b90-8c85-4616c4def2e5/console/0.log" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.196973 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.207050 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-console-oauth-config\") pod \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.207111 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-console-serving-cert\") pod \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.207145 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-trusted-ca-bundle\") pod \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.207173 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-oauth-serving-cert\") pod \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.207201 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-js6sh\" (UniqueName: \"kubernetes.io/projected/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-kube-api-access-js6sh\") pod \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.207223 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-service-ca\") pod \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.207239 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-console-config\") pod \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\" (UID: \"e4d6938a-1f6b-4b90-8c85-4616c4def2e5\") " Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.208228 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "e4d6938a-1f6b-4b90-8c85-4616c4def2e5" (UID: "e4d6938a-1f6b-4b90-8c85-4616c4def2e5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.208269 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-console-config" (OuterVolumeSpecName: "console-config") pod "e4d6938a-1f6b-4b90-8c85-4616c4def2e5" (UID: "e4d6938a-1f6b-4b90-8c85-4616c4def2e5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.208633 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "e4d6938a-1f6b-4b90-8c85-4616c4def2e5" (UID: "e4d6938a-1f6b-4b90-8c85-4616c4def2e5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.209954 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-service-ca" (OuterVolumeSpecName: "service-ca") pod "e4d6938a-1f6b-4b90-8c85-4616c4def2e5" (UID: "e4d6938a-1f6b-4b90-8c85-4616c4def2e5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.224257 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "e4d6938a-1f6b-4b90-8c85-4616c4def2e5" (UID: "e4d6938a-1f6b-4b90-8c85-4616c4def2e5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.239493 4766 scope.go:117] "RemoveContainer" containerID="ad32a7197fd193373afa7a36a53334c699a74cad5295cb9b7674cfaa44b97768" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.239877 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "e4d6938a-1f6b-4b90-8c85-4616c4def2e5" (UID: "e4d6938a-1f6b-4b90-8c85-4616c4def2e5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.246268 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-kube-api-access-js6sh" (OuterVolumeSpecName: "kube-api-access-js6sh") pod "e4d6938a-1f6b-4b90-8c85-4616c4def2e5" (UID: "e4d6938a-1f6b-4b90-8c85-4616c4def2e5"). InnerVolumeSpecName "kube-api-access-js6sh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.309692 4766 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.309721 4766 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.309730 4766 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.309738 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-js6sh\" (UniqueName: \"kubernetes.io/projected/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-kube-api-access-js6sh\") on node \"crc\" DevicePath \"\"" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.309750 4766 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.309906 4766 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-console-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.309914 4766 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e4d6938a-1f6b-4b90-8c85-4616c4def2e5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.479716 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.479773 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.479818 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.480400 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d5bd12ab3efcc84d754246df84a11f62a9d2d5fd80aa034846fa0733f2acc7ee"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.480450 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://d5bd12ab3efcc84d754246df84a11f62a9d2d5fd80aa034846fa0733f2acc7ee" gracePeriod=600 Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.957479 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-hqj86" event={"ID":"e4d6938a-1f6b-4b90-8c85-4616c4def2e5","Type":"ContainerDied","Data":"ad32a7197fd193373afa7a36a53334c699a74cad5295cb9b7674cfaa44b97768"} Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.957771 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-hqj86" event={"ID":"e4d6938a-1f6b-4b90-8c85-4616c4def2e5","Type":"ContainerDied","Data":"015c21f4b5a4534086c0aacb2fae8d63deaa0912687654f8ff65c424ecc17f3a"} Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.960836 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="d5bd12ab3efcc84d754246df84a11f62a9d2d5fd80aa034846fa0733f2acc7ee" exitCode=0 Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.960893 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-hqj86" Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.960922 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"d5bd12ab3efcc84d754246df84a11f62a9d2d5fd80aa034846fa0733f2acc7ee"} Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.961004 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"cef0ca8e4338b5dbd39bc68daed18f6f5044b46711899f9b65da90146983e184"} Nov 26 00:39:41 crc kubenswrapper[4766]: I1126 00:39:41.961047 4766 scope.go:117] "RemoveContainer" containerID="2f5bd86ee9ff3b132bc7816555a18731d6351ecdb67cc2c61c0c55d699b956fc" Nov 26 00:39:42 crc kubenswrapper[4766]: I1126 00:39:42.021300 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-hqj86"] Nov 26 00:39:42 crc kubenswrapper[4766]: I1126 00:39:42.026767 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-hqj86"] Nov 26 00:39:42 crc kubenswrapper[4766]: I1126 00:39:42.970393 4766 generic.go:334] "Generic (PLEG): container finished" podID="631bfaa9-4706-44f7-9417-500418a18c73" containerID="6b45f167cab6ddedaa564960a57524cb29bdd9553187b293bdc122f1c9987a3a" exitCode=0 Nov 26 00:39:42 crc kubenswrapper[4766]: I1126 00:39:42.970470 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk" event={"ID":"631bfaa9-4706-44f7-9417-500418a18c73","Type":"ContainerDied","Data":"6b45f167cab6ddedaa564960a57524cb29bdd9553187b293bdc122f1c9987a3a"} Nov 26 00:39:43 crc kubenswrapper[4766]: I1126 00:39:43.837755 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4d6938a-1f6b-4b90-8c85-4616c4def2e5" path="/var/lib/kubelet/pods/e4d6938a-1f6b-4b90-8c85-4616c4def2e5/volumes" Nov 26 00:39:43 crc kubenswrapper[4766]: I1126 00:39:43.989396 4766 generic.go:334] "Generic (PLEG): container finished" podID="631bfaa9-4706-44f7-9417-500418a18c73" containerID="f15f5a05e1628fb072a1a56d8cf33dbd7cb467e6467a12a2a2966e9452689f4e" exitCode=0 Nov 26 00:39:43 crc kubenswrapper[4766]: I1126 00:39:43.990056 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk" event={"ID":"631bfaa9-4706-44f7-9417-500418a18c73","Type":"ContainerDied","Data":"f15f5a05e1628fb072a1a56d8cf33dbd7cb467e6467a12a2a2966e9452689f4e"} Nov 26 00:39:45 crc kubenswrapper[4766]: I1126 00:39:45.304006 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk" Nov 26 00:39:45 crc kubenswrapper[4766]: I1126 00:39:45.378989 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/631bfaa9-4706-44f7-9417-500418a18c73-util\") pod \"631bfaa9-4706-44f7-9417-500418a18c73\" (UID: \"631bfaa9-4706-44f7-9417-500418a18c73\") " Nov 26 00:39:45 crc kubenswrapper[4766]: I1126 00:39:45.379042 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/631bfaa9-4706-44f7-9417-500418a18c73-bundle\") pod \"631bfaa9-4706-44f7-9417-500418a18c73\" (UID: \"631bfaa9-4706-44f7-9417-500418a18c73\") " Nov 26 00:39:45 crc kubenswrapper[4766]: I1126 00:39:45.379082 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzq76\" (UniqueName: \"kubernetes.io/projected/631bfaa9-4706-44f7-9417-500418a18c73-kube-api-access-nzq76\") pod \"631bfaa9-4706-44f7-9417-500418a18c73\" (UID: \"631bfaa9-4706-44f7-9417-500418a18c73\") " Nov 26 00:39:45 crc kubenswrapper[4766]: I1126 00:39:45.380149 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/631bfaa9-4706-44f7-9417-500418a18c73-bundle" (OuterVolumeSpecName: "bundle") pod "631bfaa9-4706-44f7-9417-500418a18c73" (UID: "631bfaa9-4706-44f7-9417-500418a18c73"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:39:45 crc kubenswrapper[4766]: I1126 00:39:45.394917 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/631bfaa9-4706-44f7-9417-500418a18c73-kube-api-access-nzq76" (OuterVolumeSpecName: "kube-api-access-nzq76") pod "631bfaa9-4706-44f7-9417-500418a18c73" (UID: "631bfaa9-4706-44f7-9417-500418a18c73"). InnerVolumeSpecName "kube-api-access-nzq76". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:39:45 crc kubenswrapper[4766]: I1126 00:39:45.408821 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/631bfaa9-4706-44f7-9417-500418a18c73-util" (OuterVolumeSpecName: "util") pod "631bfaa9-4706-44f7-9417-500418a18c73" (UID: "631bfaa9-4706-44f7-9417-500418a18c73"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:39:45 crc kubenswrapper[4766]: I1126 00:39:45.480070 4766 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/631bfaa9-4706-44f7-9417-500418a18c73-util\") on node \"crc\" DevicePath \"\"" Nov 26 00:39:45 crc kubenswrapper[4766]: I1126 00:39:45.480103 4766 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/631bfaa9-4706-44f7-9417-500418a18c73-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:39:45 crc kubenswrapper[4766]: I1126 00:39:45.480112 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzq76\" (UniqueName: \"kubernetes.io/projected/631bfaa9-4706-44f7-9417-500418a18c73-kube-api-access-nzq76\") on node \"crc\" DevicePath \"\"" Nov 26 00:39:46 crc kubenswrapper[4766]: I1126 00:39:46.005314 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk" event={"ID":"631bfaa9-4706-44f7-9417-500418a18c73","Type":"ContainerDied","Data":"535ff060216bed91d283920447aa8c3b473ab28edadf41ee7905bf30aa4fb520"} Nov 26 00:39:46 crc kubenswrapper[4766]: I1126 00:39:46.005358 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="535ff060216bed91d283920447aa8c3b473ab28edadf41ee7905bf30aa4fb520" Nov 26 00:39:46 crc kubenswrapper[4766]: I1126 00:39:46.005431 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.092613 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt"] Nov 26 00:39:54 crc kubenswrapper[4766]: E1126 00:39:54.093723 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="631bfaa9-4706-44f7-9417-500418a18c73" containerName="pull" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.093738 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="631bfaa9-4706-44f7-9417-500418a18c73" containerName="pull" Nov 26 00:39:54 crc kubenswrapper[4766]: E1126 00:39:54.093755 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="631bfaa9-4706-44f7-9417-500418a18c73" containerName="util" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.093762 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="631bfaa9-4706-44f7-9417-500418a18c73" containerName="util" Nov 26 00:39:54 crc kubenswrapper[4766]: E1126 00:39:54.093781 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="631bfaa9-4706-44f7-9417-500418a18c73" containerName="extract" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.093790 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="631bfaa9-4706-44f7-9417-500418a18c73" containerName="extract" Nov 26 00:39:54 crc kubenswrapper[4766]: E1126 00:39:54.093806 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4d6938a-1f6b-4b90-8c85-4616c4def2e5" containerName="console" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.093812 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4d6938a-1f6b-4b90-8c85-4616c4def2e5" containerName="console" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.093944 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4d6938a-1f6b-4b90-8c85-4616c4def2e5" containerName="console" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.093967 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="631bfaa9-4706-44f7-9417-500418a18c73" containerName="extract" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.094542 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.096304 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fd3a6275-9f47-4799-bc15-96a51b93f31b-webhook-cert\") pod \"metallb-operator-controller-manager-764c8cf56f-vrwpt\" (UID: \"fd3a6275-9f47-4799-bc15-96a51b93f31b\") " pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.096392 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dvd6\" (UniqueName: \"kubernetes.io/projected/fd3a6275-9f47-4799-bc15-96a51b93f31b-kube-api-access-6dvd6\") pod \"metallb-operator-controller-manager-764c8cf56f-vrwpt\" (UID: \"fd3a6275-9f47-4799-bc15-96a51b93f31b\") " pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.096439 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fd3a6275-9f47-4799-bc15-96a51b93f31b-apiservice-cert\") pod \"metallb-operator-controller-manager-764c8cf56f-vrwpt\" (UID: \"fd3a6275-9f47-4799-bc15-96a51b93f31b\") " pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.098991 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.099013 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-rw628" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.101112 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.101188 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.101820 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.140830 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt"] Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.197604 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fd3a6275-9f47-4799-bc15-96a51b93f31b-apiservice-cert\") pod \"metallb-operator-controller-manager-764c8cf56f-vrwpt\" (UID: \"fd3a6275-9f47-4799-bc15-96a51b93f31b\") " pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.197727 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fd3a6275-9f47-4799-bc15-96a51b93f31b-webhook-cert\") pod \"metallb-operator-controller-manager-764c8cf56f-vrwpt\" (UID: \"fd3a6275-9f47-4799-bc15-96a51b93f31b\") " pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.197809 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dvd6\" (UniqueName: \"kubernetes.io/projected/fd3a6275-9f47-4799-bc15-96a51b93f31b-kube-api-access-6dvd6\") pod \"metallb-operator-controller-manager-764c8cf56f-vrwpt\" (UID: \"fd3a6275-9f47-4799-bc15-96a51b93f31b\") " pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.208081 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fd3a6275-9f47-4799-bc15-96a51b93f31b-apiservice-cert\") pod \"metallb-operator-controller-manager-764c8cf56f-vrwpt\" (UID: \"fd3a6275-9f47-4799-bc15-96a51b93f31b\") " pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.212291 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fd3a6275-9f47-4799-bc15-96a51b93f31b-webhook-cert\") pod \"metallb-operator-controller-manager-764c8cf56f-vrwpt\" (UID: \"fd3a6275-9f47-4799-bc15-96a51b93f31b\") " pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.225345 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dvd6\" (UniqueName: \"kubernetes.io/projected/fd3a6275-9f47-4799-bc15-96a51b93f31b-kube-api-access-6dvd6\") pod \"metallb-operator-controller-manager-764c8cf56f-vrwpt\" (UID: \"fd3a6275-9f47-4799-bc15-96a51b93f31b\") " pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.413129 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.418402 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6f5d8f677c-qjhpt"] Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.419280 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6f5d8f677c-qjhpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.422908 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.423070 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-fwb2c" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.423721 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.430536 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6f5d8f677c-qjhpt"] Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.602489 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ce7c66d0-5857-46ce-a725-6531c37891cb-apiservice-cert\") pod \"metallb-operator-webhook-server-6f5d8f677c-qjhpt\" (UID: \"ce7c66d0-5857-46ce-a725-6531c37891cb\") " pod="metallb-system/metallb-operator-webhook-server-6f5d8f677c-qjhpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.602677 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6f4d\" (UniqueName: \"kubernetes.io/projected/ce7c66d0-5857-46ce-a725-6531c37891cb-kube-api-access-v6f4d\") pod \"metallb-operator-webhook-server-6f5d8f677c-qjhpt\" (UID: \"ce7c66d0-5857-46ce-a725-6531c37891cb\") " pod="metallb-system/metallb-operator-webhook-server-6f5d8f677c-qjhpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.602735 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ce7c66d0-5857-46ce-a725-6531c37891cb-webhook-cert\") pod \"metallb-operator-webhook-server-6f5d8f677c-qjhpt\" (UID: \"ce7c66d0-5857-46ce-a725-6531c37891cb\") " pod="metallb-system/metallb-operator-webhook-server-6f5d8f677c-qjhpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.704354 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ce7c66d0-5857-46ce-a725-6531c37891cb-webhook-cert\") pod \"metallb-operator-webhook-server-6f5d8f677c-qjhpt\" (UID: \"ce7c66d0-5857-46ce-a725-6531c37891cb\") " pod="metallb-system/metallb-operator-webhook-server-6f5d8f677c-qjhpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.704436 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ce7c66d0-5857-46ce-a725-6531c37891cb-apiservice-cert\") pod \"metallb-operator-webhook-server-6f5d8f677c-qjhpt\" (UID: \"ce7c66d0-5857-46ce-a725-6531c37891cb\") " pod="metallb-system/metallb-operator-webhook-server-6f5d8f677c-qjhpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.704545 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6f4d\" (UniqueName: \"kubernetes.io/projected/ce7c66d0-5857-46ce-a725-6531c37891cb-kube-api-access-v6f4d\") pod \"metallb-operator-webhook-server-6f5d8f677c-qjhpt\" (UID: \"ce7c66d0-5857-46ce-a725-6531c37891cb\") " pod="metallb-system/metallb-operator-webhook-server-6f5d8f677c-qjhpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.709388 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ce7c66d0-5857-46ce-a725-6531c37891cb-webhook-cert\") pod \"metallb-operator-webhook-server-6f5d8f677c-qjhpt\" (UID: \"ce7c66d0-5857-46ce-a725-6531c37891cb\") " pod="metallb-system/metallb-operator-webhook-server-6f5d8f677c-qjhpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.720303 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ce7c66d0-5857-46ce-a725-6531c37891cb-apiservice-cert\") pod \"metallb-operator-webhook-server-6f5d8f677c-qjhpt\" (UID: \"ce7c66d0-5857-46ce-a725-6531c37891cb\") " pod="metallb-system/metallb-operator-webhook-server-6f5d8f677c-qjhpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.723771 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6f4d\" (UniqueName: \"kubernetes.io/projected/ce7c66d0-5857-46ce-a725-6531c37891cb-kube-api-access-v6f4d\") pod \"metallb-operator-webhook-server-6f5d8f677c-qjhpt\" (UID: \"ce7c66d0-5857-46ce-a725-6531c37891cb\") " pod="metallb-system/metallb-operator-webhook-server-6f5d8f677c-qjhpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.735634 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6f5d8f677c-qjhpt" Nov 26 00:39:54 crc kubenswrapper[4766]: I1126 00:39:54.890138 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt"] Nov 26 00:39:54 crc kubenswrapper[4766]: W1126 00:39:54.943138 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd3a6275_9f47_4799_bc15_96a51b93f31b.slice/crio-7e68994ce30bd23d75d4ad47206db5f81ae7383f1a9a403588eb96ab0fe2d416 WatchSource:0}: Error finding container 7e68994ce30bd23d75d4ad47206db5f81ae7383f1a9a403588eb96ab0fe2d416: Status 404 returned error can't find the container with id 7e68994ce30bd23d75d4ad47206db5f81ae7383f1a9a403588eb96ab0fe2d416 Nov 26 00:39:55 crc kubenswrapper[4766]: I1126 00:39:55.071747 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" event={"ID":"fd3a6275-9f47-4799-bc15-96a51b93f31b","Type":"ContainerStarted","Data":"7e68994ce30bd23d75d4ad47206db5f81ae7383f1a9a403588eb96ab0fe2d416"} Nov 26 00:39:55 crc kubenswrapper[4766]: I1126 00:39:55.189511 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6f5d8f677c-qjhpt"] Nov 26 00:39:55 crc kubenswrapper[4766]: W1126 00:39:55.192802 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce7c66d0_5857_46ce_a725_6531c37891cb.slice/crio-a217cf01f6d048a48af0ba24df1b3f60e01276926038165b7fce4bbc407b92a2 WatchSource:0}: Error finding container a217cf01f6d048a48af0ba24df1b3f60e01276926038165b7fce4bbc407b92a2: Status 404 returned error can't find the container with id a217cf01f6d048a48af0ba24df1b3f60e01276926038165b7fce4bbc407b92a2 Nov 26 00:39:56 crc kubenswrapper[4766]: I1126 00:39:56.079267 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6f5d8f677c-qjhpt" event={"ID":"ce7c66d0-5857-46ce-a725-6531c37891cb","Type":"ContainerStarted","Data":"a217cf01f6d048a48af0ba24df1b3f60e01276926038165b7fce4bbc407b92a2"} Nov 26 00:40:00 crc kubenswrapper[4766]: I1126 00:40:00.118001 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" event={"ID":"fd3a6275-9f47-4799-bc15-96a51b93f31b","Type":"ContainerStarted","Data":"bd98c26b9f9e66fab6812e421b2adedc526f01a40be3a9040a7e9c259b0ff68f"} Nov 26 00:40:00 crc kubenswrapper[4766]: I1126 00:40:00.118500 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" Nov 26 00:40:00 crc kubenswrapper[4766]: I1126 00:40:00.119906 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6f5d8f677c-qjhpt" event={"ID":"ce7c66d0-5857-46ce-a725-6531c37891cb","Type":"ContainerStarted","Data":"cfd22f2ec700ff37ae5e01c57d42908ca2c7e7a75fe6ee6b508a1ee8da1752bc"} Nov 26 00:40:00 crc kubenswrapper[4766]: I1126 00:40:00.120015 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6f5d8f677c-qjhpt" Nov 26 00:40:00 crc kubenswrapper[4766]: I1126 00:40:00.138781 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" podStartSLOduration=1.329569014 podStartE2EDuration="6.138766213s" podCreationTimestamp="2025-11-26 00:39:54 +0000 UTC" firstStartedPulling="2025-11-26 00:39:54.946249 +0000 UTC m=+975.795019440" lastFinishedPulling="2025-11-26 00:39:59.755446209 +0000 UTC m=+980.604216639" observedRunningTime="2025-11-26 00:40:00.136025793 +0000 UTC m=+980.984796223" watchObservedRunningTime="2025-11-26 00:40:00.138766213 +0000 UTC m=+980.987536633" Nov 26 00:40:00 crc kubenswrapper[4766]: I1126 00:40:00.163893 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6f5d8f677c-qjhpt" podStartSLOduration=1.590085484 podStartE2EDuration="6.163876755s" podCreationTimestamp="2025-11-26 00:39:54 +0000 UTC" firstStartedPulling="2025-11-26 00:39:55.195977899 +0000 UTC m=+976.044748329" lastFinishedPulling="2025-11-26 00:39:59.76976917 +0000 UTC m=+980.618539600" observedRunningTime="2025-11-26 00:40:00.161366732 +0000 UTC m=+981.010137162" watchObservedRunningTime="2025-11-26 00:40:00.163876755 +0000 UTC m=+981.012647185" Nov 26 00:40:14 crc kubenswrapper[4766]: I1126 00:40:14.742961 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6f5d8f677c-qjhpt" Nov 26 00:40:34 crc kubenswrapper[4766]: I1126 00:40:34.417021 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.148160 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-7rtx6"] Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.151487 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.154479 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.155022 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.155048 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-7j6jd" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.158865 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-k6vct"] Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.159790 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-k6vct" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.162212 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.189514 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-k6vct"] Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.277619 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-d5tdc"] Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.283364 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-d5tdc" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.283448 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-9dzn2"] Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.284666 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-9dzn2" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.287901 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.287943 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.287949 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.287947 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.288301 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-jgphj" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.294841 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-9dzn2"] Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.321050 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/931efe26-0d06-4bcb-bbaa-f65225f02419-metrics\") pod \"frr-k8s-7rtx6\" (UID: \"931efe26-0d06-4bcb-bbaa-f65225f02419\") " pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.321108 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/931efe26-0d06-4bcb-bbaa-f65225f02419-metrics-certs\") pod \"frr-k8s-7rtx6\" (UID: \"931efe26-0d06-4bcb-bbaa-f65225f02419\") " pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.321143 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7dxs\" (UniqueName: \"kubernetes.io/projected/76a9ad7d-db95-4177-839f-db3e7cc1ac81-kube-api-access-q7dxs\") pod \"speaker-d5tdc\" (UID: \"76a9ad7d-db95-4177-839f-db3e7cc1ac81\") " pod="metallb-system/speaker-d5tdc" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.321191 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/76a9ad7d-db95-4177-839f-db3e7cc1ac81-memberlist\") pod \"speaker-d5tdc\" (UID: \"76a9ad7d-db95-4177-839f-db3e7cc1ac81\") " pod="metallb-system/speaker-d5tdc" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.321220 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0c9c337a-681d-4a68-8e51-5d690a326fe7-metrics-certs\") pod \"controller-6c7b4b5f48-9dzn2\" (UID: \"0c9c337a-681d-4a68-8e51-5d690a326fe7\") " pod="metallb-system/controller-6c7b4b5f48-9dzn2" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.321244 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/931efe26-0d06-4bcb-bbaa-f65225f02419-frr-sockets\") pod \"frr-k8s-7rtx6\" (UID: \"931efe26-0d06-4bcb-bbaa-f65225f02419\") " pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.321269 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/76a9ad7d-db95-4177-839f-db3e7cc1ac81-metallb-excludel2\") pod \"speaker-d5tdc\" (UID: \"76a9ad7d-db95-4177-839f-db3e7cc1ac81\") " pod="metallb-system/speaker-d5tdc" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.321290 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/931efe26-0d06-4bcb-bbaa-f65225f02419-frr-startup\") pod \"frr-k8s-7rtx6\" (UID: \"931efe26-0d06-4bcb-bbaa-f65225f02419\") " pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.321309 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0c9c337a-681d-4a68-8e51-5d690a326fe7-cert\") pod \"controller-6c7b4b5f48-9dzn2\" (UID: \"0c9c337a-681d-4a68-8e51-5d690a326fe7\") " pod="metallb-system/controller-6c7b4b5f48-9dzn2" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.321332 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnvb5\" (UniqueName: \"kubernetes.io/projected/0c9c337a-681d-4a68-8e51-5d690a326fe7-kube-api-access-gnvb5\") pod \"controller-6c7b4b5f48-9dzn2\" (UID: \"0c9c337a-681d-4a68-8e51-5d690a326fe7\") " pod="metallb-system/controller-6c7b4b5f48-9dzn2" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.321353 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76pnq\" (UniqueName: \"kubernetes.io/projected/88faf6a0-1824-43f6-b711-09cb78c54a5d-kube-api-access-76pnq\") pod \"frr-k8s-webhook-server-6998585d5-k6vct\" (UID: \"88faf6a0-1824-43f6-b711-09cb78c54a5d\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-k6vct" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.321380 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/931efe26-0d06-4bcb-bbaa-f65225f02419-reloader\") pod \"frr-k8s-7rtx6\" (UID: \"931efe26-0d06-4bcb-bbaa-f65225f02419\") " pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.321435 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/931efe26-0d06-4bcb-bbaa-f65225f02419-frr-conf\") pod \"frr-k8s-7rtx6\" (UID: \"931efe26-0d06-4bcb-bbaa-f65225f02419\") " pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.321456 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqsw9\" (UniqueName: \"kubernetes.io/projected/931efe26-0d06-4bcb-bbaa-f65225f02419-kube-api-access-rqsw9\") pod \"frr-k8s-7rtx6\" (UID: \"931efe26-0d06-4bcb-bbaa-f65225f02419\") " pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.321482 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/76a9ad7d-db95-4177-839f-db3e7cc1ac81-metrics-certs\") pod \"speaker-d5tdc\" (UID: \"76a9ad7d-db95-4177-839f-db3e7cc1ac81\") " pod="metallb-system/speaker-d5tdc" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.321514 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/88faf6a0-1824-43f6-b711-09cb78c54a5d-cert\") pod \"frr-k8s-webhook-server-6998585d5-k6vct\" (UID: \"88faf6a0-1824-43f6-b711-09cb78c54a5d\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-k6vct" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.422814 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/931efe26-0d06-4bcb-bbaa-f65225f02419-metrics\") pod \"frr-k8s-7rtx6\" (UID: \"931efe26-0d06-4bcb-bbaa-f65225f02419\") " pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.422898 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/931efe26-0d06-4bcb-bbaa-f65225f02419-metrics-certs\") pod \"frr-k8s-7rtx6\" (UID: \"931efe26-0d06-4bcb-bbaa-f65225f02419\") " pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.422928 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7dxs\" (UniqueName: \"kubernetes.io/projected/76a9ad7d-db95-4177-839f-db3e7cc1ac81-kube-api-access-q7dxs\") pod \"speaker-d5tdc\" (UID: \"76a9ad7d-db95-4177-839f-db3e7cc1ac81\") " pod="metallb-system/speaker-d5tdc" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.422998 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/76a9ad7d-db95-4177-839f-db3e7cc1ac81-memberlist\") pod \"speaker-d5tdc\" (UID: \"76a9ad7d-db95-4177-839f-db3e7cc1ac81\") " pod="metallb-system/speaker-d5tdc" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.423041 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0c9c337a-681d-4a68-8e51-5d690a326fe7-metrics-certs\") pod \"controller-6c7b4b5f48-9dzn2\" (UID: \"0c9c337a-681d-4a68-8e51-5d690a326fe7\") " pod="metallb-system/controller-6c7b4b5f48-9dzn2" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.423074 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/931efe26-0d06-4bcb-bbaa-f65225f02419-frr-sockets\") pod \"frr-k8s-7rtx6\" (UID: \"931efe26-0d06-4bcb-bbaa-f65225f02419\") " pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.423126 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/76a9ad7d-db95-4177-839f-db3e7cc1ac81-metallb-excludel2\") pod \"speaker-d5tdc\" (UID: \"76a9ad7d-db95-4177-839f-db3e7cc1ac81\") " pod="metallb-system/speaker-d5tdc" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.423150 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/931efe26-0d06-4bcb-bbaa-f65225f02419-frr-startup\") pod \"frr-k8s-7rtx6\" (UID: \"931efe26-0d06-4bcb-bbaa-f65225f02419\") " pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.423171 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0c9c337a-681d-4a68-8e51-5d690a326fe7-cert\") pod \"controller-6c7b4b5f48-9dzn2\" (UID: \"0c9c337a-681d-4a68-8e51-5d690a326fe7\") " pod="metallb-system/controller-6c7b4b5f48-9dzn2" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.423217 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnvb5\" (UniqueName: \"kubernetes.io/projected/0c9c337a-681d-4a68-8e51-5d690a326fe7-kube-api-access-gnvb5\") pod \"controller-6c7b4b5f48-9dzn2\" (UID: \"0c9c337a-681d-4a68-8e51-5d690a326fe7\") " pod="metallb-system/controller-6c7b4b5f48-9dzn2" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.423219 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/931efe26-0d06-4bcb-bbaa-f65225f02419-metrics\") pod \"frr-k8s-7rtx6\" (UID: \"931efe26-0d06-4bcb-bbaa-f65225f02419\") " pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.423240 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76pnq\" (UniqueName: \"kubernetes.io/projected/88faf6a0-1824-43f6-b711-09cb78c54a5d-kube-api-access-76pnq\") pod \"frr-k8s-webhook-server-6998585d5-k6vct\" (UID: \"88faf6a0-1824-43f6-b711-09cb78c54a5d\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-k6vct" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.423309 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/931efe26-0d06-4bcb-bbaa-f65225f02419-reloader\") pod \"frr-k8s-7rtx6\" (UID: \"931efe26-0d06-4bcb-bbaa-f65225f02419\") " pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.423414 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/931efe26-0d06-4bcb-bbaa-f65225f02419-frr-conf\") pod \"frr-k8s-7rtx6\" (UID: \"931efe26-0d06-4bcb-bbaa-f65225f02419\") " pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.423431 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqsw9\" (UniqueName: \"kubernetes.io/projected/931efe26-0d06-4bcb-bbaa-f65225f02419-kube-api-access-rqsw9\") pod \"frr-k8s-7rtx6\" (UID: \"931efe26-0d06-4bcb-bbaa-f65225f02419\") " pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.423465 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/76a9ad7d-db95-4177-839f-db3e7cc1ac81-metrics-certs\") pod \"speaker-d5tdc\" (UID: \"76a9ad7d-db95-4177-839f-db3e7cc1ac81\") " pod="metallb-system/speaker-d5tdc" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.423509 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/88faf6a0-1824-43f6-b711-09cb78c54a5d-cert\") pod \"frr-k8s-webhook-server-6998585d5-k6vct\" (UID: \"88faf6a0-1824-43f6-b711-09cb78c54a5d\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-k6vct" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.423578 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/931efe26-0d06-4bcb-bbaa-f65225f02419-frr-sockets\") pod \"frr-k8s-7rtx6\" (UID: \"931efe26-0d06-4bcb-bbaa-f65225f02419\") " pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: E1126 00:40:35.423641 4766 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Nov 26 00:40:35 crc kubenswrapper[4766]: E1126 00:40:35.423698 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88faf6a0-1824-43f6-b711-09cb78c54a5d-cert podName:88faf6a0-1824-43f6-b711-09cb78c54a5d nodeName:}" failed. No retries permitted until 2025-11-26 00:40:35.92368216 +0000 UTC m=+1016.772452590 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/88faf6a0-1824-43f6-b711-09cb78c54a5d-cert") pod "frr-k8s-webhook-server-6998585d5-k6vct" (UID: "88faf6a0-1824-43f6-b711-09cb78c54a5d") : secret "frr-k8s-webhook-server-cert" not found Nov 26 00:40:35 crc kubenswrapper[4766]: E1126 00:40:35.423940 4766 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.423971 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/931efe26-0d06-4bcb-bbaa-f65225f02419-reloader\") pod \"frr-k8s-7rtx6\" (UID: \"931efe26-0d06-4bcb-bbaa-f65225f02419\") " pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: E1126 00:40:35.423984 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/76a9ad7d-db95-4177-839f-db3e7cc1ac81-memberlist podName:76a9ad7d-db95-4177-839f-db3e7cc1ac81 nodeName:}" failed. No retries permitted until 2025-11-26 00:40:35.923967218 +0000 UTC m=+1016.772737648 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/76a9ad7d-db95-4177-839f-db3e7cc1ac81-memberlist") pod "speaker-d5tdc" (UID: "76a9ad7d-db95-4177-839f-db3e7cc1ac81") : secret "metallb-memberlist" not found Nov 26 00:40:35 crc kubenswrapper[4766]: E1126 00:40:35.424020 4766 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Nov 26 00:40:35 crc kubenswrapper[4766]: E1126 00:40:35.424041 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c9c337a-681d-4a68-8e51-5d690a326fe7-metrics-certs podName:0c9c337a-681d-4a68-8e51-5d690a326fe7 nodeName:}" failed. No retries permitted until 2025-11-26 00:40:35.924035909 +0000 UTC m=+1016.772806339 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0c9c337a-681d-4a68-8e51-5d690a326fe7-metrics-certs") pod "controller-6c7b4b5f48-9dzn2" (UID: "0c9c337a-681d-4a68-8e51-5d690a326fe7") : secret "controller-certs-secret" not found Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.424147 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/931efe26-0d06-4bcb-bbaa-f65225f02419-frr-conf\") pod \"frr-k8s-7rtx6\" (UID: \"931efe26-0d06-4bcb-bbaa-f65225f02419\") " pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.424907 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/76a9ad7d-db95-4177-839f-db3e7cc1ac81-metallb-excludel2\") pod \"speaker-d5tdc\" (UID: \"76a9ad7d-db95-4177-839f-db3e7cc1ac81\") " pod="metallb-system/speaker-d5tdc" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.424967 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/931efe26-0d06-4bcb-bbaa-f65225f02419-frr-startup\") pod \"frr-k8s-7rtx6\" (UID: \"931efe26-0d06-4bcb-bbaa-f65225f02419\") " pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.428743 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/76a9ad7d-db95-4177-839f-db3e7cc1ac81-metrics-certs\") pod \"speaker-d5tdc\" (UID: \"76a9ad7d-db95-4177-839f-db3e7cc1ac81\") " pod="metallb-system/speaker-d5tdc" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.435531 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0c9c337a-681d-4a68-8e51-5d690a326fe7-cert\") pod \"controller-6c7b4b5f48-9dzn2\" (UID: \"0c9c337a-681d-4a68-8e51-5d690a326fe7\") " pod="metallb-system/controller-6c7b4b5f48-9dzn2" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.444752 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/931efe26-0d06-4bcb-bbaa-f65225f02419-metrics-certs\") pod \"frr-k8s-7rtx6\" (UID: \"931efe26-0d06-4bcb-bbaa-f65225f02419\") " pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.444855 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7dxs\" (UniqueName: \"kubernetes.io/projected/76a9ad7d-db95-4177-839f-db3e7cc1ac81-kube-api-access-q7dxs\") pod \"speaker-d5tdc\" (UID: \"76a9ad7d-db95-4177-839f-db3e7cc1ac81\") " pod="metallb-system/speaker-d5tdc" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.446552 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76pnq\" (UniqueName: \"kubernetes.io/projected/88faf6a0-1824-43f6-b711-09cb78c54a5d-kube-api-access-76pnq\") pod \"frr-k8s-webhook-server-6998585d5-k6vct\" (UID: \"88faf6a0-1824-43f6-b711-09cb78c54a5d\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-k6vct" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.447818 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqsw9\" (UniqueName: \"kubernetes.io/projected/931efe26-0d06-4bcb-bbaa-f65225f02419-kube-api-access-rqsw9\") pod \"frr-k8s-7rtx6\" (UID: \"931efe26-0d06-4bcb-bbaa-f65225f02419\") " pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.449146 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnvb5\" (UniqueName: \"kubernetes.io/projected/0c9c337a-681d-4a68-8e51-5d690a326fe7-kube-api-access-gnvb5\") pod \"controller-6c7b4b5f48-9dzn2\" (UID: \"0c9c337a-681d-4a68-8e51-5d690a326fe7\") " pod="metallb-system/controller-6c7b4b5f48-9dzn2" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.474562 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.928498 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/88faf6a0-1824-43f6-b711-09cb78c54a5d-cert\") pod \"frr-k8s-webhook-server-6998585d5-k6vct\" (UID: \"88faf6a0-1824-43f6-b711-09cb78c54a5d\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-k6vct" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.928873 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/76a9ad7d-db95-4177-839f-db3e7cc1ac81-memberlist\") pod \"speaker-d5tdc\" (UID: \"76a9ad7d-db95-4177-839f-db3e7cc1ac81\") " pod="metallb-system/speaker-d5tdc" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.928895 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0c9c337a-681d-4a68-8e51-5d690a326fe7-metrics-certs\") pod \"controller-6c7b4b5f48-9dzn2\" (UID: \"0c9c337a-681d-4a68-8e51-5d690a326fe7\") " pod="metallb-system/controller-6c7b4b5f48-9dzn2" Nov 26 00:40:35 crc kubenswrapper[4766]: E1126 00:40:35.929032 4766 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 26 00:40:35 crc kubenswrapper[4766]: E1126 00:40:35.929106 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/76a9ad7d-db95-4177-839f-db3e7cc1ac81-memberlist podName:76a9ad7d-db95-4177-839f-db3e7cc1ac81 nodeName:}" failed. No retries permitted until 2025-11-26 00:40:36.929088518 +0000 UTC m=+1017.777858948 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/76a9ad7d-db95-4177-839f-db3e7cc1ac81-memberlist") pod "speaker-d5tdc" (UID: "76a9ad7d-db95-4177-839f-db3e7cc1ac81") : secret "metallb-memberlist" not found Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.934411 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/88faf6a0-1824-43f6-b711-09cb78c54a5d-cert\") pod \"frr-k8s-webhook-server-6998585d5-k6vct\" (UID: \"88faf6a0-1824-43f6-b711-09cb78c54a5d\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-k6vct" Nov 26 00:40:35 crc kubenswrapper[4766]: I1126 00:40:35.935514 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0c9c337a-681d-4a68-8e51-5d690a326fe7-metrics-certs\") pod \"controller-6c7b4b5f48-9dzn2\" (UID: \"0c9c337a-681d-4a68-8e51-5d690a326fe7\") " pod="metallb-system/controller-6c7b4b5f48-9dzn2" Nov 26 00:40:36 crc kubenswrapper[4766]: I1126 00:40:36.080236 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-k6vct" Nov 26 00:40:36 crc kubenswrapper[4766]: I1126 00:40:36.226139 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-9dzn2" Nov 26 00:40:36 crc kubenswrapper[4766]: I1126 00:40:36.348139 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7rtx6" event={"ID":"931efe26-0d06-4bcb-bbaa-f65225f02419","Type":"ContainerStarted","Data":"88e0b8a1bc8b6cd4d67e24680d20bd5bdb8b9e474011ba3a24c986c9772a411a"} Nov 26 00:40:36 crc kubenswrapper[4766]: I1126 00:40:36.472599 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-k6vct"] Nov 26 00:40:36 crc kubenswrapper[4766]: W1126 00:40:36.484895 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88faf6a0_1824_43f6_b711_09cb78c54a5d.slice/crio-46fe434257aa2603bbd0aa7fc1ed0af01316744269fc4d09aa5a83c836e3787f WatchSource:0}: Error finding container 46fe434257aa2603bbd0aa7fc1ed0af01316744269fc4d09aa5a83c836e3787f: Status 404 returned error can't find the container with id 46fe434257aa2603bbd0aa7fc1ed0af01316744269fc4d09aa5a83c836e3787f Nov 26 00:40:36 crc kubenswrapper[4766]: I1126 00:40:36.609161 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-9dzn2"] Nov 26 00:40:36 crc kubenswrapper[4766]: W1126 00:40:36.615072 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c9c337a_681d_4a68_8e51_5d690a326fe7.slice/crio-db1087a044c573ea0ed1f470ee7227045312cd77970e2fe9412e32b36d52ff62 WatchSource:0}: Error finding container db1087a044c573ea0ed1f470ee7227045312cd77970e2fe9412e32b36d52ff62: Status 404 returned error can't find the container with id db1087a044c573ea0ed1f470ee7227045312cd77970e2fe9412e32b36d52ff62 Nov 26 00:40:36 crc kubenswrapper[4766]: I1126 00:40:36.943415 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/76a9ad7d-db95-4177-839f-db3e7cc1ac81-memberlist\") pod \"speaker-d5tdc\" (UID: \"76a9ad7d-db95-4177-839f-db3e7cc1ac81\") " pod="metallb-system/speaker-d5tdc" Nov 26 00:40:36 crc kubenswrapper[4766]: I1126 00:40:36.949194 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/76a9ad7d-db95-4177-839f-db3e7cc1ac81-memberlist\") pod \"speaker-d5tdc\" (UID: \"76a9ad7d-db95-4177-839f-db3e7cc1ac81\") " pod="metallb-system/speaker-d5tdc" Nov 26 00:40:37 crc kubenswrapper[4766]: I1126 00:40:37.110241 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-d5tdc" Nov 26 00:40:37 crc kubenswrapper[4766]: W1126 00:40:37.126966 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76a9ad7d_db95_4177_839f_db3e7cc1ac81.slice/crio-8fa2927e9c146523c67b9b1a8f84dc979b2a8300f8aff0c9ab34516c12af2b33 WatchSource:0}: Error finding container 8fa2927e9c146523c67b9b1a8f84dc979b2a8300f8aff0c9ab34516c12af2b33: Status 404 returned error can't find the container with id 8fa2927e9c146523c67b9b1a8f84dc979b2a8300f8aff0c9ab34516c12af2b33 Nov 26 00:40:37 crc kubenswrapper[4766]: I1126 00:40:37.358100 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-d5tdc" event={"ID":"76a9ad7d-db95-4177-839f-db3e7cc1ac81","Type":"ContainerStarted","Data":"8fa2927e9c146523c67b9b1a8f84dc979b2a8300f8aff0c9ab34516c12af2b33"} Nov 26 00:40:37 crc kubenswrapper[4766]: I1126 00:40:37.361832 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-9dzn2" event={"ID":"0c9c337a-681d-4a68-8e51-5d690a326fe7","Type":"ContainerStarted","Data":"59d49a993f78ce8628ed1a19916100a523ab075276e115d2b025c6147523590c"} Nov 26 00:40:37 crc kubenswrapper[4766]: I1126 00:40:37.361892 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-9dzn2" event={"ID":"0c9c337a-681d-4a68-8e51-5d690a326fe7","Type":"ContainerStarted","Data":"ad79f6fb3cb1e04c3fa948f2a8355eaafbbf004a6819cd03711033f2dc4e09af"} Nov 26 00:40:37 crc kubenswrapper[4766]: I1126 00:40:37.361908 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-9dzn2" event={"ID":"0c9c337a-681d-4a68-8e51-5d690a326fe7","Type":"ContainerStarted","Data":"db1087a044c573ea0ed1f470ee7227045312cd77970e2fe9412e32b36d52ff62"} Nov 26 00:40:37 crc kubenswrapper[4766]: I1126 00:40:37.361999 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-9dzn2" Nov 26 00:40:37 crc kubenswrapper[4766]: I1126 00:40:37.367335 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-k6vct" event={"ID":"88faf6a0-1824-43f6-b711-09cb78c54a5d","Type":"ContainerStarted","Data":"46fe434257aa2603bbd0aa7fc1ed0af01316744269fc4d09aa5a83c836e3787f"} Nov 26 00:40:37 crc kubenswrapper[4766]: I1126 00:40:37.385477 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-9dzn2" podStartSLOduration=2.385405272 podStartE2EDuration="2.385405272s" podCreationTimestamp="2025-11-26 00:40:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:40:37.378278422 +0000 UTC m=+1018.227048862" watchObservedRunningTime="2025-11-26 00:40:37.385405272 +0000 UTC m=+1018.234175702" Nov 26 00:40:38 crc kubenswrapper[4766]: I1126 00:40:38.390500 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-d5tdc" event={"ID":"76a9ad7d-db95-4177-839f-db3e7cc1ac81","Type":"ContainerStarted","Data":"bc0be794af6b884b2ee7042dd071f31f7273f7057bf12a4664e3bc928e53b536"} Nov 26 00:40:38 crc kubenswrapper[4766]: I1126 00:40:38.390576 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-d5tdc" event={"ID":"76a9ad7d-db95-4177-839f-db3e7cc1ac81","Type":"ContainerStarted","Data":"dd9074118310a2e85cddd782670c81022926581e024eccf5b42600a92de3429d"} Nov 26 00:40:38 crc kubenswrapper[4766]: I1126 00:40:38.421586 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-d5tdc" podStartSLOduration=3.421567526 podStartE2EDuration="3.421567526s" podCreationTimestamp="2025-11-26 00:40:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:40:38.417837562 +0000 UTC m=+1019.266607992" watchObservedRunningTime="2025-11-26 00:40:38.421567526 +0000 UTC m=+1019.270337956" Nov 26 00:40:39 crc kubenswrapper[4766]: I1126 00:40:39.398187 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-d5tdc" Nov 26 00:40:43 crc kubenswrapper[4766]: I1126 00:40:43.433327 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7rtx6" event={"ID":"931efe26-0d06-4bcb-bbaa-f65225f02419","Type":"ContainerStarted","Data":"01e1499de9bafe4082df91b7a4b533ecf560a0235c26009a80f7f6af9b981c23"} Nov 26 00:40:43 crc kubenswrapper[4766]: I1126 00:40:43.436688 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-k6vct" event={"ID":"88faf6a0-1824-43f6-b711-09cb78c54a5d","Type":"ContainerStarted","Data":"96f73c2bdfecab70b6669b036edbb3372d228e7a9d6e5f286c38cd14aab410fa"} Nov 26 00:40:43 crc kubenswrapper[4766]: I1126 00:40:43.437050 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-k6vct" Nov 26 00:40:43 crc kubenswrapper[4766]: I1126 00:40:43.479132 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-k6vct" podStartSLOduration=1.6928548920000002 podStartE2EDuration="8.479108929s" podCreationTimestamp="2025-11-26 00:40:35 +0000 UTC" firstStartedPulling="2025-11-26 00:40:36.486479225 +0000 UTC m=+1017.335249655" lastFinishedPulling="2025-11-26 00:40:43.272733262 +0000 UTC m=+1024.121503692" observedRunningTime="2025-11-26 00:40:43.475079038 +0000 UTC m=+1024.323849468" watchObservedRunningTime="2025-11-26 00:40:43.479108929 +0000 UTC m=+1024.327879359" Nov 26 00:40:44 crc kubenswrapper[4766]: I1126 00:40:44.444081 4766 generic.go:334] "Generic (PLEG): container finished" podID="931efe26-0d06-4bcb-bbaa-f65225f02419" containerID="01e1499de9bafe4082df91b7a4b533ecf560a0235c26009a80f7f6af9b981c23" exitCode=0 Nov 26 00:40:44 crc kubenswrapper[4766]: I1126 00:40:44.444110 4766 generic.go:334] "Generic (PLEG): container finished" podID="931efe26-0d06-4bcb-bbaa-f65225f02419" containerID="17314eb029e35ca99df6e222eb39975ca9ccd0a78a1ff37b9dd6920f087fafbd" exitCode=0 Nov 26 00:40:44 crc kubenswrapper[4766]: I1126 00:40:44.445157 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7rtx6" event={"ID":"931efe26-0d06-4bcb-bbaa-f65225f02419","Type":"ContainerDied","Data":"01e1499de9bafe4082df91b7a4b533ecf560a0235c26009a80f7f6af9b981c23"} Nov 26 00:40:44 crc kubenswrapper[4766]: I1126 00:40:44.445188 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7rtx6" event={"ID":"931efe26-0d06-4bcb-bbaa-f65225f02419","Type":"ContainerDied","Data":"17314eb029e35ca99df6e222eb39975ca9ccd0a78a1ff37b9dd6920f087fafbd"} Nov 26 00:40:45 crc kubenswrapper[4766]: I1126 00:40:45.453257 4766 generic.go:334] "Generic (PLEG): container finished" podID="931efe26-0d06-4bcb-bbaa-f65225f02419" containerID="91d0ad3d8ada8f947a4d62c4e4f5b8ad90f402fc5c730d3d622d94f046b64b9a" exitCode=0 Nov 26 00:40:45 crc kubenswrapper[4766]: I1126 00:40:45.453322 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7rtx6" event={"ID":"931efe26-0d06-4bcb-bbaa-f65225f02419","Type":"ContainerDied","Data":"91d0ad3d8ada8f947a4d62c4e4f5b8ad90f402fc5c730d3d622d94f046b64b9a"} Nov 26 00:40:46 crc kubenswrapper[4766]: I1126 00:40:46.229622 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-9dzn2" Nov 26 00:40:46 crc kubenswrapper[4766]: I1126 00:40:46.463576 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7rtx6" event={"ID":"931efe26-0d06-4bcb-bbaa-f65225f02419","Type":"ContainerStarted","Data":"c8542efed5c7e4ecc97c95e9f0935cf60a9ac024e822a5e9bf0b853e20728607"} Nov 26 00:40:46 crc kubenswrapper[4766]: I1126 00:40:46.463632 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7rtx6" event={"ID":"931efe26-0d06-4bcb-bbaa-f65225f02419","Type":"ContainerStarted","Data":"736e9bb88cfef4614e9fd7a44b504dc2b7e0d60bbe3da7dd28f360c106838b69"} Nov 26 00:40:46 crc kubenswrapper[4766]: I1126 00:40:46.463642 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7rtx6" event={"ID":"931efe26-0d06-4bcb-bbaa-f65225f02419","Type":"ContainerStarted","Data":"70c4638ccd0a8e7d28762ffed2b3371b956c0a319af7211f31cc7af474282365"} Nov 26 00:40:46 crc kubenswrapper[4766]: I1126 00:40:46.463664 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7rtx6" event={"ID":"931efe26-0d06-4bcb-bbaa-f65225f02419","Type":"ContainerStarted","Data":"f8dc68b33b24c1b7dc7c0b9b0101fdb6a5f77fc53ab184163de67ac193ec747c"} Nov 26 00:40:46 crc kubenswrapper[4766]: I1126 00:40:46.463685 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7rtx6" event={"ID":"931efe26-0d06-4bcb-bbaa-f65225f02419","Type":"ContainerStarted","Data":"fce768669e9c24bab4578248e7e67b01706e8b3f3cef2ddc6d673ae7c27aa10d"} Nov 26 00:40:47 crc kubenswrapper[4766]: I1126 00:40:47.113219 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-d5tdc" Nov 26 00:40:47 crc kubenswrapper[4766]: I1126 00:40:47.474777 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7rtx6" event={"ID":"931efe26-0d06-4bcb-bbaa-f65225f02419","Type":"ContainerStarted","Data":"9783637a3c96f525e106615a19ba6bddb85394dc9bd6f8d880d749b3979ecbb7"} Nov 26 00:40:47 crc kubenswrapper[4766]: I1126 00:40:47.474976 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:47 crc kubenswrapper[4766]: I1126 00:40:47.496189 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-7rtx6" podStartSLOduration=5.179052866 podStartE2EDuration="12.496171232s" podCreationTimestamp="2025-11-26 00:40:35 +0000 UTC" firstStartedPulling="2025-11-26 00:40:35.939644504 +0000 UTC m=+1016.788414944" lastFinishedPulling="2025-11-26 00:40:43.25676288 +0000 UTC m=+1024.105533310" observedRunningTime="2025-11-26 00:40:47.492644753 +0000 UTC m=+1028.341415203" watchObservedRunningTime="2025-11-26 00:40:47.496171232 +0000 UTC m=+1028.344941662" Nov 26 00:40:49 crc kubenswrapper[4766]: I1126 00:40:49.929989 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-kt5d4"] Nov 26 00:40:49 crc kubenswrapper[4766]: I1126 00:40:49.931085 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kt5d4" Nov 26 00:40:49 crc kubenswrapper[4766]: I1126 00:40:49.935086 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-457zd" Nov 26 00:40:49 crc kubenswrapper[4766]: I1126 00:40:49.935557 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 26 00:40:49 crc kubenswrapper[4766]: I1126 00:40:49.935741 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 26 00:40:49 crc kubenswrapper[4766]: I1126 00:40:49.941552 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-kt5d4"] Nov 26 00:40:49 crc kubenswrapper[4766]: I1126 00:40:49.947450 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq5dd\" (UniqueName: \"kubernetes.io/projected/04cc18a0-b648-45be-84bd-2fd8a746d651-kube-api-access-cq5dd\") pod \"openstack-operator-index-kt5d4\" (UID: \"04cc18a0-b648-45be-84bd-2fd8a746d651\") " pod="openstack-operators/openstack-operator-index-kt5d4" Nov 26 00:40:50 crc kubenswrapper[4766]: I1126 00:40:50.048707 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq5dd\" (UniqueName: \"kubernetes.io/projected/04cc18a0-b648-45be-84bd-2fd8a746d651-kube-api-access-cq5dd\") pod \"openstack-operator-index-kt5d4\" (UID: \"04cc18a0-b648-45be-84bd-2fd8a746d651\") " pod="openstack-operators/openstack-operator-index-kt5d4" Nov 26 00:40:50 crc kubenswrapper[4766]: I1126 00:40:50.065765 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq5dd\" (UniqueName: \"kubernetes.io/projected/04cc18a0-b648-45be-84bd-2fd8a746d651-kube-api-access-cq5dd\") pod \"openstack-operator-index-kt5d4\" (UID: \"04cc18a0-b648-45be-84bd-2fd8a746d651\") " pod="openstack-operators/openstack-operator-index-kt5d4" Nov 26 00:40:50 crc kubenswrapper[4766]: I1126 00:40:50.263536 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kt5d4" Nov 26 00:40:50 crc kubenswrapper[4766]: I1126 00:40:50.475159 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:50 crc kubenswrapper[4766]: I1126 00:40:50.511662 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:50 crc kubenswrapper[4766]: I1126 00:40:50.663330 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-kt5d4"] Nov 26 00:40:50 crc kubenswrapper[4766]: W1126 00:40:50.686644 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04cc18a0_b648_45be_84bd_2fd8a746d651.slice/crio-6323fcd8db59632b194aad8f303bcc769b3397a4a73269fe96376297d00d7587 WatchSource:0}: Error finding container 6323fcd8db59632b194aad8f303bcc769b3397a4a73269fe96376297d00d7587: Status 404 returned error can't find the container with id 6323fcd8db59632b194aad8f303bcc769b3397a4a73269fe96376297d00d7587 Nov 26 00:40:51 crc kubenswrapper[4766]: I1126 00:40:51.508780 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kt5d4" event={"ID":"04cc18a0-b648-45be-84bd-2fd8a746d651","Type":"ContainerStarted","Data":"6323fcd8db59632b194aad8f303bcc769b3397a4a73269fe96376297d00d7587"} Nov 26 00:40:53 crc kubenswrapper[4766]: I1126 00:40:53.311943 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-kt5d4"] Nov 26 00:40:53 crc kubenswrapper[4766]: I1126 00:40:53.527905 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kt5d4" event={"ID":"04cc18a0-b648-45be-84bd-2fd8a746d651","Type":"ContainerStarted","Data":"e69b9121663333b1d2497475c7313f3d900ea7d05b5bd21f82ea6205224607ab"} Nov 26 00:40:53 crc kubenswrapper[4766]: I1126 00:40:53.541433 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-kt5d4" podStartSLOduration=2.211634178 podStartE2EDuration="4.541407428s" podCreationTimestamp="2025-11-26 00:40:49 +0000 UTC" firstStartedPulling="2025-11-26 00:40:50.690104324 +0000 UTC m=+1031.538874764" lastFinishedPulling="2025-11-26 00:40:53.019877594 +0000 UTC m=+1033.868648014" observedRunningTime="2025-11-26 00:40:53.538363822 +0000 UTC m=+1034.387134252" watchObservedRunningTime="2025-11-26 00:40:53.541407428 +0000 UTC m=+1034.390177858" Nov 26 00:40:53 crc kubenswrapper[4766]: I1126 00:40:53.925316 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-sp492"] Nov 26 00:40:53 crc kubenswrapper[4766]: I1126 00:40:53.928375 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-sp492" Nov 26 00:40:53 crc kubenswrapper[4766]: I1126 00:40:53.930602 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-sp492"] Nov 26 00:40:54 crc kubenswrapper[4766]: I1126 00:40:54.013554 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9q64l\" (UniqueName: \"kubernetes.io/projected/5244fd2a-7bec-4649-af25-c77d4ec96ed0-kube-api-access-9q64l\") pod \"openstack-operator-index-sp492\" (UID: \"5244fd2a-7bec-4649-af25-c77d4ec96ed0\") " pod="openstack-operators/openstack-operator-index-sp492" Nov 26 00:40:54 crc kubenswrapper[4766]: I1126 00:40:54.114396 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9q64l\" (UniqueName: \"kubernetes.io/projected/5244fd2a-7bec-4649-af25-c77d4ec96ed0-kube-api-access-9q64l\") pod \"openstack-operator-index-sp492\" (UID: \"5244fd2a-7bec-4649-af25-c77d4ec96ed0\") " pod="openstack-operators/openstack-operator-index-sp492" Nov 26 00:40:54 crc kubenswrapper[4766]: I1126 00:40:54.132888 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9q64l\" (UniqueName: \"kubernetes.io/projected/5244fd2a-7bec-4649-af25-c77d4ec96ed0-kube-api-access-9q64l\") pod \"openstack-operator-index-sp492\" (UID: \"5244fd2a-7bec-4649-af25-c77d4ec96ed0\") " pod="openstack-operators/openstack-operator-index-sp492" Nov 26 00:40:54 crc kubenswrapper[4766]: I1126 00:40:54.276638 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-sp492" Nov 26 00:40:54 crc kubenswrapper[4766]: I1126 00:40:54.533729 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-kt5d4" podUID="04cc18a0-b648-45be-84bd-2fd8a746d651" containerName="registry-server" containerID="cri-o://e69b9121663333b1d2497475c7313f3d900ea7d05b5bd21f82ea6205224607ab" gracePeriod=2 Nov 26 00:40:54 crc kubenswrapper[4766]: I1126 00:40:54.670627 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-sp492"] Nov 26 00:40:54 crc kubenswrapper[4766]: I1126 00:40:54.883748 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kt5d4" Nov 26 00:40:54 crc kubenswrapper[4766]: I1126 00:40:54.928296 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cq5dd\" (UniqueName: \"kubernetes.io/projected/04cc18a0-b648-45be-84bd-2fd8a746d651-kube-api-access-cq5dd\") pod \"04cc18a0-b648-45be-84bd-2fd8a746d651\" (UID: \"04cc18a0-b648-45be-84bd-2fd8a746d651\") " Nov 26 00:40:54 crc kubenswrapper[4766]: I1126 00:40:54.936008 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04cc18a0-b648-45be-84bd-2fd8a746d651-kube-api-access-cq5dd" (OuterVolumeSpecName: "kube-api-access-cq5dd") pod "04cc18a0-b648-45be-84bd-2fd8a746d651" (UID: "04cc18a0-b648-45be-84bd-2fd8a746d651"). InnerVolumeSpecName "kube-api-access-cq5dd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:40:55 crc kubenswrapper[4766]: I1126 00:40:55.030326 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cq5dd\" (UniqueName: \"kubernetes.io/projected/04cc18a0-b648-45be-84bd-2fd8a746d651-kube-api-access-cq5dd\") on node \"crc\" DevicePath \"\"" Nov 26 00:40:55 crc kubenswrapper[4766]: I1126 00:40:55.477035 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-7rtx6" Nov 26 00:40:55 crc kubenswrapper[4766]: I1126 00:40:55.541764 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-sp492" event={"ID":"5244fd2a-7bec-4649-af25-c77d4ec96ed0","Type":"ContainerStarted","Data":"c7197c34a427d130647a1158a3796edab2fc53cd6fe3cb9072396faf3b5f3b70"} Nov 26 00:40:55 crc kubenswrapper[4766]: I1126 00:40:55.542059 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-sp492" event={"ID":"5244fd2a-7bec-4649-af25-c77d4ec96ed0","Type":"ContainerStarted","Data":"e43b3f6e584381881ce25052d193b48e24c035b0712cec8b9f565650a5380613"} Nov 26 00:40:55 crc kubenswrapper[4766]: I1126 00:40:55.543279 4766 generic.go:334] "Generic (PLEG): container finished" podID="04cc18a0-b648-45be-84bd-2fd8a746d651" containerID="e69b9121663333b1d2497475c7313f3d900ea7d05b5bd21f82ea6205224607ab" exitCode=0 Nov 26 00:40:55 crc kubenswrapper[4766]: I1126 00:40:55.543321 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kt5d4" event={"ID":"04cc18a0-b648-45be-84bd-2fd8a746d651","Type":"ContainerDied","Data":"e69b9121663333b1d2497475c7313f3d900ea7d05b5bd21f82ea6205224607ab"} Nov 26 00:40:55 crc kubenswrapper[4766]: I1126 00:40:55.543346 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kt5d4" Nov 26 00:40:55 crc kubenswrapper[4766]: I1126 00:40:55.543351 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kt5d4" event={"ID":"04cc18a0-b648-45be-84bd-2fd8a746d651","Type":"ContainerDied","Data":"6323fcd8db59632b194aad8f303bcc769b3397a4a73269fe96376297d00d7587"} Nov 26 00:40:55 crc kubenswrapper[4766]: I1126 00:40:55.543380 4766 scope.go:117] "RemoveContainer" containerID="e69b9121663333b1d2497475c7313f3d900ea7d05b5bd21f82ea6205224607ab" Nov 26 00:40:55 crc kubenswrapper[4766]: I1126 00:40:55.565913 4766 scope.go:117] "RemoveContainer" containerID="e69b9121663333b1d2497475c7313f3d900ea7d05b5bd21f82ea6205224607ab" Nov 26 00:40:55 crc kubenswrapper[4766]: E1126 00:40:55.567051 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e69b9121663333b1d2497475c7313f3d900ea7d05b5bd21f82ea6205224607ab\": container with ID starting with e69b9121663333b1d2497475c7313f3d900ea7d05b5bd21f82ea6205224607ab not found: ID does not exist" containerID="e69b9121663333b1d2497475c7313f3d900ea7d05b5bd21f82ea6205224607ab" Nov 26 00:40:55 crc kubenswrapper[4766]: I1126 00:40:55.567059 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-sp492" podStartSLOduration=2.5277577129999997 podStartE2EDuration="2.567046012s" podCreationTimestamp="2025-11-26 00:40:53 +0000 UTC" firstStartedPulling="2025-11-26 00:40:54.680607607 +0000 UTC m=+1035.529378037" lastFinishedPulling="2025-11-26 00:40:54.719895896 +0000 UTC m=+1035.568666336" observedRunningTime="2025-11-26 00:40:55.563346979 +0000 UTC m=+1036.412117419" watchObservedRunningTime="2025-11-26 00:40:55.567046012 +0000 UTC m=+1036.415816442" Nov 26 00:40:55 crc kubenswrapper[4766]: I1126 00:40:55.567110 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e69b9121663333b1d2497475c7313f3d900ea7d05b5bd21f82ea6205224607ab"} err="failed to get container status \"e69b9121663333b1d2497475c7313f3d900ea7d05b5bd21f82ea6205224607ab\": rpc error: code = NotFound desc = could not find container \"e69b9121663333b1d2497475c7313f3d900ea7d05b5bd21f82ea6205224607ab\": container with ID starting with e69b9121663333b1d2497475c7313f3d900ea7d05b5bd21f82ea6205224607ab not found: ID does not exist" Nov 26 00:40:55 crc kubenswrapper[4766]: I1126 00:40:55.576079 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-kt5d4"] Nov 26 00:40:55 crc kubenswrapper[4766]: I1126 00:40:55.579847 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-kt5d4"] Nov 26 00:40:55 crc kubenswrapper[4766]: I1126 00:40:55.836255 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04cc18a0-b648-45be-84bd-2fd8a746d651" path="/var/lib/kubelet/pods/04cc18a0-b648-45be-84bd-2fd8a746d651/volumes" Nov 26 00:40:56 crc kubenswrapper[4766]: I1126 00:40:56.086436 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-k6vct" Nov 26 00:41:04 crc kubenswrapper[4766]: I1126 00:41:04.276792 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-sp492" Nov 26 00:41:04 crc kubenswrapper[4766]: I1126 00:41:04.277370 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-sp492" Nov 26 00:41:04 crc kubenswrapper[4766]: I1126 00:41:04.324842 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-sp492" Nov 26 00:41:04 crc kubenswrapper[4766]: I1126 00:41:04.647182 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-sp492" Nov 26 00:41:05 crc kubenswrapper[4766]: I1126 00:41:05.554002 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b"] Nov 26 00:41:05 crc kubenswrapper[4766]: E1126 00:41:05.554313 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04cc18a0-b648-45be-84bd-2fd8a746d651" containerName="registry-server" Nov 26 00:41:05 crc kubenswrapper[4766]: I1126 00:41:05.554330 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="04cc18a0-b648-45be-84bd-2fd8a746d651" containerName="registry-server" Nov 26 00:41:05 crc kubenswrapper[4766]: I1126 00:41:05.554530 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="04cc18a0-b648-45be-84bd-2fd8a746d651" containerName="registry-server" Nov 26 00:41:05 crc kubenswrapper[4766]: I1126 00:41:05.555781 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b" Nov 26 00:41:05 crc kubenswrapper[4766]: I1126 00:41:05.558252 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-qfv88" Nov 26 00:41:05 crc kubenswrapper[4766]: I1126 00:41:05.566045 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b"] Nov 26 00:41:05 crc kubenswrapper[4766]: I1126 00:41:05.679235 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/28a6a73a-8156-4df1-a41e-d2fe03ce2b25-util\") pod \"afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b\" (UID: \"28a6a73a-8156-4df1-a41e-d2fe03ce2b25\") " pod="openstack-operators/afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b" Nov 26 00:41:05 crc kubenswrapper[4766]: I1126 00:41:05.680048 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/28a6a73a-8156-4df1-a41e-d2fe03ce2b25-bundle\") pod \"afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b\" (UID: \"28a6a73a-8156-4df1-a41e-d2fe03ce2b25\") " pod="openstack-operators/afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b" Nov 26 00:41:05 crc kubenswrapper[4766]: I1126 00:41:05.680320 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gd9x\" (UniqueName: \"kubernetes.io/projected/28a6a73a-8156-4df1-a41e-d2fe03ce2b25-kube-api-access-9gd9x\") pod \"afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b\" (UID: \"28a6a73a-8156-4df1-a41e-d2fe03ce2b25\") " pod="openstack-operators/afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b" Nov 26 00:41:05 crc kubenswrapper[4766]: I1126 00:41:05.781439 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/28a6a73a-8156-4df1-a41e-d2fe03ce2b25-bundle\") pod \"afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b\" (UID: \"28a6a73a-8156-4df1-a41e-d2fe03ce2b25\") " pod="openstack-operators/afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b" Nov 26 00:41:05 crc kubenswrapper[4766]: I1126 00:41:05.781816 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gd9x\" (UniqueName: \"kubernetes.io/projected/28a6a73a-8156-4df1-a41e-d2fe03ce2b25-kube-api-access-9gd9x\") pod \"afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b\" (UID: \"28a6a73a-8156-4df1-a41e-d2fe03ce2b25\") " pod="openstack-operators/afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b" Nov 26 00:41:05 crc kubenswrapper[4766]: I1126 00:41:05.781942 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/28a6a73a-8156-4df1-a41e-d2fe03ce2b25-util\") pod \"afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b\" (UID: \"28a6a73a-8156-4df1-a41e-d2fe03ce2b25\") " pod="openstack-operators/afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b" Nov 26 00:41:05 crc kubenswrapper[4766]: I1126 00:41:05.782007 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/28a6a73a-8156-4df1-a41e-d2fe03ce2b25-bundle\") pod \"afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b\" (UID: \"28a6a73a-8156-4df1-a41e-d2fe03ce2b25\") " pod="openstack-operators/afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b" Nov 26 00:41:05 crc kubenswrapper[4766]: I1126 00:41:05.782333 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/28a6a73a-8156-4df1-a41e-d2fe03ce2b25-util\") pod \"afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b\" (UID: \"28a6a73a-8156-4df1-a41e-d2fe03ce2b25\") " pod="openstack-operators/afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b" Nov 26 00:41:05 crc kubenswrapper[4766]: I1126 00:41:05.806860 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gd9x\" (UniqueName: \"kubernetes.io/projected/28a6a73a-8156-4df1-a41e-d2fe03ce2b25-kube-api-access-9gd9x\") pod \"afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b\" (UID: \"28a6a73a-8156-4df1-a41e-d2fe03ce2b25\") " pod="openstack-operators/afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b" Nov 26 00:41:05 crc kubenswrapper[4766]: I1126 00:41:05.878970 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b" Nov 26 00:41:06 crc kubenswrapper[4766]: I1126 00:41:06.340711 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b"] Nov 26 00:41:06 crc kubenswrapper[4766]: W1126 00:41:06.346487 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28a6a73a_8156_4df1_a41e_d2fe03ce2b25.slice/crio-10d1fad443a0724ace167f0d5d35d5d6657af4050076b5a1bea284b1caa63a2a WatchSource:0}: Error finding container 10d1fad443a0724ace167f0d5d35d5d6657af4050076b5a1bea284b1caa63a2a: Status 404 returned error can't find the container with id 10d1fad443a0724ace167f0d5d35d5d6657af4050076b5a1bea284b1caa63a2a Nov 26 00:41:06 crc kubenswrapper[4766]: I1126 00:41:06.630747 4766 generic.go:334] "Generic (PLEG): container finished" podID="28a6a73a-8156-4df1-a41e-d2fe03ce2b25" containerID="e19aaa49b3c058989b984883cacaeba7771a794c97a7c8b71b7a90c62124a9cc" exitCode=0 Nov 26 00:41:06 crc kubenswrapper[4766]: I1126 00:41:06.630790 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b" event={"ID":"28a6a73a-8156-4df1-a41e-d2fe03ce2b25","Type":"ContainerDied","Data":"e19aaa49b3c058989b984883cacaeba7771a794c97a7c8b71b7a90c62124a9cc"} Nov 26 00:41:06 crc kubenswrapper[4766]: I1126 00:41:06.630815 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b" event={"ID":"28a6a73a-8156-4df1-a41e-d2fe03ce2b25","Type":"ContainerStarted","Data":"10d1fad443a0724ace167f0d5d35d5d6657af4050076b5a1bea284b1caa63a2a"} Nov 26 00:41:07 crc kubenswrapper[4766]: I1126 00:41:07.639075 4766 generic.go:334] "Generic (PLEG): container finished" podID="28a6a73a-8156-4df1-a41e-d2fe03ce2b25" containerID="f386e30a37e124cde3f0271e5958e9b5f8b564d1eef798da4d58af9a2c24ba7b" exitCode=0 Nov 26 00:41:07 crc kubenswrapper[4766]: I1126 00:41:07.639152 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b" event={"ID":"28a6a73a-8156-4df1-a41e-d2fe03ce2b25","Type":"ContainerDied","Data":"f386e30a37e124cde3f0271e5958e9b5f8b564d1eef798da4d58af9a2c24ba7b"} Nov 26 00:41:08 crc kubenswrapper[4766]: I1126 00:41:08.653575 4766 generic.go:334] "Generic (PLEG): container finished" podID="28a6a73a-8156-4df1-a41e-d2fe03ce2b25" containerID="fb91f10684f95761776719bfc725219ba3df9e78727a02577e242a35596356ad" exitCode=0 Nov 26 00:41:08 crc kubenswrapper[4766]: I1126 00:41:08.653623 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b" event={"ID":"28a6a73a-8156-4df1-a41e-d2fe03ce2b25","Type":"ContainerDied","Data":"fb91f10684f95761776719bfc725219ba3df9e78727a02577e242a35596356ad"} Nov 26 00:41:09 crc kubenswrapper[4766]: I1126 00:41:09.922417 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b" Nov 26 00:41:10 crc kubenswrapper[4766]: I1126 00:41:10.071625 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gd9x\" (UniqueName: \"kubernetes.io/projected/28a6a73a-8156-4df1-a41e-d2fe03ce2b25-kube-api-access-9gd9x\") pod \"28a6a73a-8156-4df1-a41e-d2fe03ce2b25\" (UID: \"28a6a73a-8156-4df1-a41e-d2fe03ce2b25\") " Nov 26 00:41:10 crc kubenswrapper[4766]: I1126 00:41:10.071785 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/28a6a73a-8156-4df1-a41e-d2fe03ce2b25-bundle\") pod \"28a6a73a-8156-4df1-a41e-d2fe03ce2b25\" (UID: \"28a6a73a-8156-4df1-a41e-d2fe03ce2b25\") " Nov 26 00:41:10 crc kubenswrapper[4766]: I1126 00:41:10.071940 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/28a6a73a-8156-4df1-a41e-d2fe03ce2b25-util\") pod \"28a6a73a-8156-4df1-a41e-d2fe03ce2b25\" (UID: \"28a6a73a-8156-4df1-a41e-d2fe03ce2b25\") " Nov 26 00:41:10 crc kubenswrapper[4766]: I1126 00:41:10.072854 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28a6a73a-8156-4df1-a41e-d2fe03ce2b25-bundle" (OuterVolumeSpecName: "bundle") pod "28a6a73a-8156-4df1-a41e-d2fe03ce2b25" (UID: "28a6a73a-8156-4df1-a41e-d2fe03ce2b25"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:41:10 crc kubenswrapper[4766]: I1126 00:41:10.076715 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28a6a73a-8156-4df1-a41e-d2fe03ce2b25-kube-api-access-9gd9x" (OuterVolumeSpecName: "kube-api-access-9gd9x") pod "28a6a73a-8156-4df1-a41e-d2fe03ce2b25" (UID: "28a6a73a-8156-4df1-a41e-d2fe03ce2b25"). InnerVolumeSpecName "kube-api-access-9gd9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:41:10 crc kubenswrapper[4766]: I1126 00:41:10.089165 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28a6a73a-8156-4df1-a41e-d2fe03ce2b25-util" (OuterVolumeSpecName: "util") pod "28a6a73a-8156-4df1-a41e-d2fe03ce2b25" (UID: "28a6a73a-8156-4df1-a41e-d2fe03ce2b25"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:41:10 crc kubenswrapper[4766]: I1126 00:41:10.174146 4766 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/28a6a73a-8156-4df1-a41e-d2fe03ce2b25-util\") on node \"crc\" DevicePath \"\"" Nov 26 00:41:10 crc kubenswrapper[4766]: I1126 00:41:10.174198 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gd9x\" (UniqueName: \"kubernetes.io/projected/28a6a73a-8156-4df1-a41e-d2fe03ce2b25-kube-api-access-9gd9x\") on node \"crc\" DevicePath \"\"" Nov 26 00:41:10 crc kubenswrapper[4766]: I1126 00:41:10.174210 4766 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/28a6a73a-8156-4df1-a41e-d2fe03ce2b25-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:41:10 crc kubenswrapper[4766]: I1126 00:41:10.669521 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b" event={"ID":"28a6a73a-8156-4df1-a41e-d2fe03ce2b25","Type":"ContainerDied","Data":"10d1fad443a0724ace167f0d5d35d5d6657af4050076b5a1bea284b1caa63a2a"} Nov 26 00:41:10 crc kubenswrapper[4766]: I1126 00:41:10.669571 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10d1fad443a0724ace167f0d5d35d5d6657af4050076b5a1bea284b1caa63a2a" Nov 26 00:41:10 crc kubenswrapper[4766]: I1126 00:41:10.669606 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b" Nov 26 00:41:28 crc kubenswrapper[4766]: I1126 00:41:28.688146 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7d5fdccdc4-5sh97"] Nov 26 00:41:28 crc kubenswrapper[4766]: E1126 00:41:28.689019 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28a6a73a-8156-4df1-a41e-d2fe03ce2b25" containerName="extract" Nov 26 00:41:28 crc kubenswrapper[4766]: I1126 00:41:28.689036 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="28a6a73a-8156-4df1-a41e-d2fe03ce2b25" containerName="extract" Nov 26 00:41:28 crc kubenswrapper[4766]: E1126 00:41:28.689060 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28a6a73a-8156-4df1-a41e-d2fe03ce2b25" containerName="util" Nov 26 00:41:28 crc kubenswrapper[4766]: I1126 00:41:28.689067 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="28a6a73a-8156-4df1-a41e-d2fe03ce2b25" containerName="util" Nov 26 00:41:28 crc kubenswrapper[4766]: E1126 00:41:28.689078 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28a6a73a-8156-4df1-a41e-d2fe03ce2b25" containerName="pull" Nov 26 00:41:28 crc kubenswrapper[4766]: I1126 00:41:28.689085 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="28a6a73a-8156-4df1-a41e-d2fe03ce2b25" containerName="pull" Nov 26 00:41:28 crc kubenswrapper[4766]: I1126 00:41:28.689240 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="28a6a73a-8156-4df1-a41e-d2fe03ce2b25" containerName="extract" Nov 26 00:41:28 crc kubenswrapper[4766]: I1126 00:41:28.689864 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7d5fdccdc4-5sh97" Nov 26 00:41:28 crc kubenswrapper[4766]: I1126 00:41:28.692026 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-gdhck" Nov 26 00:41:28 crc kubenswrapper[4766]: I1126 00:41:28.697612 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsdv9\" (UniqueName: \"kubernetes.io/projected/62c72069-cf63-4244-a55e-0028aaaed380-kube-api-access-dsdv9\") pod \"openstack-operator-controller-operator-7d5fdccdc4-5sh97\" (UID: \"62c72069-cf63-4244-a55e-0028aaaed380\") " pod="openstack-operators/openstack-operator-controller-operator-7d5fdccdc4-5sh97" Nov 26 00:41:28 crc kubenswrapper[4766]: I1126 00:41:28.721547 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7d5fdccdc4-5sh97"] Nov 26 00:41:28 crc kubenswrapper[4766]: I1126 00:41:28.798620 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsdv9\" (UniqueName: \"kubernetes.io/projected/62c72069-cf63-4244-a55e-0028aaaed380-kube-api-access-dsdv9\") pod \"openstack-operator-controller-operator-7d5fdccdc4-5sh97\" (UID: \"62c72069-cf63-4244-a55e-0028aaaed380\") " pod="openstack-operators/openstack-operator-controller-operator-7d5fdccdc4-5sh97" Nov 26 00:41:28 crc kubenswrapper[4766]: I1126 00:41:28.819211 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsdv9\" (UniqueName: \"kubernetes.io/projected/62c72069-cf63-4244-a55e-0028aaaed380-kube-api-access-dsdv9\") pod \"openstack-operator-controller-operator-7d5fdccdc4-5sh97\" (UID: \"62c72069-cf63-4244-a55e-0028aaaed380\") " pod="openstack-operators/openstack-operator-controller-operator-7d5fdccdc4-5sh97" Nov 26 00:41:29 crc kubenswrapper[4766]: I1126 00:41:29.074067 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7d5fdccdc4-5sh97" Nov 26 00:41:29 crc kubenswrapper[4766]: I1126 00:41:29.485937 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7d5fdccdc4-5sh97"] Nov 26 00:41:29 crc kubenswrapper[4766]: W1126 00:41:29.494517 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62c72069_cf63_4244_a55e_0028aaaed380.slice/crio-516c0891cdf2f7b46beb7a1641e5b15621bbb8431a73fd8d2fe2b37d7cb94975 WatchSource:0}: Error finding container 516c0891cdf2f7b46beb7a1641e5b15621bbb8431a73fd8d2fe2b37d7cb94975: Status 404 returned error can't find the container with id 516c0891cdf2f7b46beb7a1641e5b15621bbb8431a73fd8d2fe2b37d7cb94975 Nov 26 00:41:29 crc kubenswrapper[4766]: I1126 00:41:29.802625 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7d5fdccdc4-5sh97" event={"ID":"62c72069-cf63-4244-a55e-0028aaaed380","Type":"ContainerStarted","Data":"516c0891cdf2f7b46beb7a1641e5b15621bbb8431a73fd8d2fe2b37d7cb94975"} Nov 26 00:41:36 crc kubenswrapper[4766]: I1126 00:41:36.851555 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7d5fdccdc4-5sh97" event={"ID":"62c72069-cf63-4244-a55e-0028aaaed380","Type":"ContainerStarted","Data":"8f4c6185fb7a867b4fb5901e3bfa554b6d34fb4271abaf7dfaee99cdc49113be"} Nov 26 00:41:36 crc kubenswrapper[4766]: I1126 00:41:36.852102 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7d5fdccdc4-5sh97" Nov 26 00:41:36 crc kubenswrapper[4766]: I1126 00:41:36.893519 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7d5fdccdc4-5sh97" podStartSLOduration=2.427655673 podStartE2EDuration="8.893494846s" podCreationTimestamp="2025-11-26 00:41:28 +0000 UTC" firstStartedPulling="2025-11-26 00:41:29.496658495 +0000 UTC m=+1070.345428925" lastFinishedPulling="2025-11-26 00:41:35.962497668 +0000 UTC m=+1076.811268098" observedRunningTime="2025-11-26 00:41:36.885233959 +0000 UTC m=+1077.734004479" watchObservedRunningTime="2025-11-26 00:41:36.893494846 +0000 UTC m=+1077.742265306" Nov 26 00:41:41 crc kubenswrapper[4766]: I1126 00:41:41.479457 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:41:41 crc kubenswrapper[4766]: I1126 00:41:41.480048 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:41:49 crc kubenswrapper[4766]: I1126 00:41:49.077823 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7d5fdccdc4-5sh97" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.134434 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.136093 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.138501 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-gbglb" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.138884 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.139962 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.143494 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-4w6wk" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.146400 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.154586 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.164802 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-dqtcj"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.166529 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.171991 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-dhtlj" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.202601 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.204127 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.208973 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-dqtcj"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.210026 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-94zcs" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.219419 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.220518 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.234173 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.237487 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-gd78k" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.276147 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.287816 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-686hb\" (UniqueName: \"kubernetes.io/projected/31b1b73b-1d10-49ac-8549-b202c965b63e-kube-api-access-686hb\") pod \"cinder-operator-controller-manager-6b7f75547b-jgz6c\" (UID: \"31b1b73b-1d10-49ac-8549-b202c965b63e\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.288265 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvmlj\" (UniqueName: \"kubernetes.io/projected/cc87f631-ba4e-4943-b51d-bf1012cca754-kube-api-access-wvmlj\") pod \"barbican-operator-controller-manager-7b64f4fb85-gbdhz\" (UID: \"cc87f631-ba4e-4943-b51d-bf1012cca754\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.288502 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnshx\" (UniqueName: \"kubernetes.io/projected/0e6cfef3-86af-4cf0-ab06-b3eef687836b-kube-api-access-vnshx\") pod \"designate-operator-controller-manager-955677c94-dqtcj\" (UID: \"0e6cfef3-86af-4cf0-ab06-b3eef687836b\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.306734 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.341079 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-glg62" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.380767 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.401584 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.403605 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-686hb\" (UniqueName: \"kubernetes.io/projected/31b1b73b-1d10-49ac-8549-b202c965b63e-kube-api-access-686hb\") pod \"cinder-operator-controller-manager-6b7f75547b-jgz6c\" (UID: \"31b1b73b-1d10-49ac-8549-b202c965b63e\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.403681 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvmlj\" (UniqueName: \"kubernetes.io/projected/cc87f631-ba4e-4943-b51d-bf1012cca754-kube-api-access-wvmlj\") pod \"barbican-operator-controller-manager-7b64f4fb85-gbdhz\" (UID: \"cc87f631-ba4e-4943-b51d-bf1012cca754\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.403708 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnshx\" (UniqueName: \"kubernetes.io/projected/0e6cfef3-86af-4cf0-ab06-b3eef687836b-kube-api-access-vnshx\") pod \"designate-operator-controller-manager-955677c94-dqtcj\" (UID: \"0e6cfef3-86af-4cf0-ab06-b3eef687836b\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.403758 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqtsw\" (UniqueName: \"kubernetes.io/projected/5284ad70-a36c-4666-8202-84633d83f50b-kube-api-access-mqtsw\") pod \"horizon-operator-controller-manager-5d494799bf-h4mrn\" (UID: \"5284ad70-a36c-4666-8202-84633d83f50b\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.403780 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn798\" (UniqueName: \"kubernetes.io/projected/1ecbd1d2-cb7f-432f-bc75-903d0643ea29-kube-api-access-qn798\") pod \"heat-operator-controller-manager-5b77f656f-tkj9b\" (UID: \"1ecbd1d2-cb7f-432f-bc75-903d0643ea29\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.403796 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtbzh\" (UniqueName: \"kubernetes.io/projected/45353fc4-9955-460b-9276-9792dfdaf8f4-kube-api-access-qtbzh\") pod \"glance-operator-controller-manager-589cbd6b5b-k68q9\" (UID: \"45353fc4-9955-460b-9276-9792dfdaf8f4\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.406834 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.410333 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-pnfcn" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.410886 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.414779 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.428086 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.450783 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-686hb\" (UniqueName: \"kubernetes.io/projected/31b1b73b-1d10-49ac-8549-b202c965b63e-kube-api-access-686hb\") pod \"cinder-operator-controller-manager-6b7f75547b-jgz6c\" (UID: \"31b1b73b-1d10-49ac-8549-b202c965b63e\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.456589 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnshx\" (UniqueName: \"kubernetes.io/projected/0e6cfef3-86af-4cf0-ab06-b3eef687836b-kube-api-access-vnshx\") pod \"designate-operator-controller-manager-955677c94-dqtcj\" (UID: \"0e6cfef3-86af-4cf0-ab06-b3eef687836b\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.460138 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvmlj\" (UniqueName: \"kubernetes.io/projected/cc87f631-ba4e-4943-b51d-bf1012cca754-kube-api-access-wvmlj\") pod \"barbican-operator-controller-manager-7b64f4fb85-gbdhz\" (UID: \"cc87f631-ba4e-4943-b51d-bf1012cca754\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.465929 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.467671 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.469224 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.471143 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-qv6sf" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.479124 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.479177 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.497781 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.499063 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.499671 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.504399 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-n8pfm" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.505019 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8fgz\" (UniqueName: \"kubernetes.io/projected/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-kube-api-access-f8fgz\") pod \"infra-operator-controller-manager-57548d458d-x6h4v\" (UID: \"c3cec209-fdef-44a6-90d0-2d06cfd7e5cc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.505069 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-cert\") pod \"infra-operator-controller-manager-57548d458d-x6h4v\" (UID: \"c3cec209-fdef-44a6-90d0-2d06cfd7e5cc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.505138 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqtsw\" (UniqueName: \"kubernetes.io/projected/5284ad70-a36c-4666-8202-84633d83f50b-kube-api-access-mqtsw\") pod \"horizon-operator-controller-manager-5d494799bf-h4mrn\" (UID: \"5284ad70-a36c-4666-8202-84633d83f50b\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.505163 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn798\" (UniqueName: \"kubernetes.io/projected/1ecbd1d2-cb7f-432f-bc75-903d0643ea29-kube-api-access-qn798\") pod \"heat-operator-controller-manager-5b77f656f-tkj9b\" (UID: \"1ecbd1d2-cb7f-432f-bc75-903d0643ea29\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.505195 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtbzh\" (UniqueName: \"kubernetes.io/projected/45353fc4-9955-460b-9276-9792dfdaf8f4-kube-api-access-qtbzh\") pod \"glance-operator-controller-manager-589cbd6b5b-k68q9\" (UID: \"45353fc4-9955-460b-9276-9792dfdaf8f4\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.529803 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.531260 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.535063 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-6wqr6" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.548884 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.553923 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn798\" (UniqueName: \"kubernetes.io/projected/1ecbd1d2-cb7f-432f-bc75-903d0643ea29-kube-api-access-qn798\") pod \"heat-operator-controller-manager-5b77f656f-tkj9b\" (UID: \"1ecbd1d2-cb7f-432f-bc75-903d0643ea29\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.556982 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.582455 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqtsw\" (UniqueName: \"kubernetes.io/projected/5284ad70-a36c-4666-8202-84633d83f50b-kube-api-access-mqtsw\") pod \"horizon-operator-controller-manager-5d494799bf-h4mrn\" (UID: \"5284ad70-a36c-4666-8202-84633d83f50b\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.583207 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.583330 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtbzh\" (UniqueName: \"kubernetes.io/projected/45353fc4-9955-460b-9276-9792dfdaf8f4-kube-api-access-qtbzh\") pod \"glance-operator-controller-manager-589cbd6b5b-k68q9\" (UID: \"45353fc4-9955-460b-9276-9792dfdaf8f4\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.606638 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vsp8\" (UniqueName: \"kubernetes.io/projected/3d0d59b2-0486-40a2-ae60-b738a4f76dc4-kube-api-access-6vsp8\") pod \"keystone-operator-controller-manager-7b4567c7cf-fcr76\" (UID: \"3d0d59b2-0486-40a2-ae60-b738a4f76dc4\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.606784 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkvlr\" (UniqueName: \"kubernetes.io/projected/08bd3753-3b70-44f8-92ff-98bedd88f662-kube-api-access-tkvlr\") pod \"ironic-operator-controller-manager-67cb4dc6d4-qtnxq\" (UID: \"08bd3753-3b70-44f8-92ff-98bedd88f662\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.606956 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5rsm\" (UniqueName: \"kubernetes.io/projected/a16c5563-750f-4e91-93bd-4044c656cb88-kube-api-access-p5rsm\") pod \"manila-operator-controller-manager-5d499bf58b-rs4kn\" (UID: \"a16c5563-750f-4e91-93bd-4044c656cb88\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.607031 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8fgz\" (UniqueName: \"kubernetes.io/projected/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-kube-api-access-f8fgz\") pod \"infra-operator-controller-manager-57548d458d-x6h4v\" (UID: \"c3cec209-fdef-44a6-90d0-2d06cfd7e5cc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.607066 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-cert\") pod \"infra-operator-controller-manager-57548d458d-x6h4v\" (UID: \"c3cec209-fdef-44a6-90d0-2d06cfd7e5cc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:42:11 crc kubenswrapper[4766]: E1126 00:42:11.607438 4766 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 26 00:42:11 crc kubenswrapper[4766]: E1126 00:42:11.607598 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-cert podName:c3cec209-fdef-44a6-90d0-2d06cfd7e5cc nodeName:}" failed. No retries permitted until 2025-11-26 00:42:12.107564408 +0000 UTC m=+1112.956334838 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-cert") pod "infra-operator-controller-manager-57548d458d-x6h4v" (UID: "c3cec209-fdef-44a6-90d0-2d06cfd7e5cc") : secret "infra-operator-webhook-server-cert" not found Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.610665 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.611959 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.612524 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.616269 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-tbnmt" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.624091 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.627782 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.630061 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-5mqxk" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.638745 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8fgz\" (UniqueName: \"kubernetes.io/projected/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-kube-api-access-f8fgz\") pod \"infra-operator-controller-manager-57548d458d-x6h4v\" (UID: \"c3cec209-fdef-44a6-90d0-2d06cfd7e5cc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.657116 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.667090 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.668819 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.675157 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-vzktk" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.688067 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.694614 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.702702 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.710784 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vsp8\" (UniqueName: \"kubernetes.io/projected/3d0d59b2-0486-40a2-ae60-b738a4f76dc4-kube-api-access-6vsp8\") pod \"keystone-operator-controller-manager-7b4567c7cf-fcr76\" (UID: \"3d0d59b2-0486-40a2-ae60-b738a4f76dc4\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.710884 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkvlr\" (UniqueName: \"kubernetes.io/projected/08bd3753-3b70-44f8-92ff-98bedd88f662-kube-api-access-tkvlr\") pod \"ironic-operator-controller-manager-67cb4dc6d4-qtnxq\" (UID: \"08bd3753-3b70-44f8-92ff-98bedd88f662\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.710960 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lccc6\" (UniqueName: \"kubernetes.io/projected/2bb23a3b-1f4a-409f-9d9c-0afe73a33352-kube-api-access-lccc6\") pod \"neutron-operator-controller-manager-6fdcddb789-wdzgw\" (UID: \"2bb23a3b-1f4a-409f-9d9c-0afe73a33352\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.711041 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxmbh\" (UniqueName: \"kubernetes.io/projected/749a25e6-23eb-4fdd-8e15-ac2874cdee7f-kube-api-access-cxmbh\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-z865g\" (UID: \"749a25e6-23eb-4fdd-8e15-ac2874cdee7f\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.711075 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5rsm\" (UniqueName: \"kubernetes.io/projected/a16c5563-750f-4e91-93bd-4044c656cb88-kube-api-access-p5rsm\") pod \"manila-operator-controller-manager-5d499bf58b-rs4kn\" (UID: \"a16c5563-750f-4e91-93bd-4044c656cb88\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.715332 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.717834 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.722383 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.724257 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-gjrqk" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.745060 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vsp8\" (UniqueName: \"kubernetes.io/projected/3d0d59b2-0486-40a2-ae60-b738a4f76dc4-kube-api-access-6vsp8\") pod \"keystone-operator-controller-manager-7b4567c7cf-fcr76\" (UID: \"3d0d59b2-0486-40a2-ae60-b738a4f76dc4\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.746108 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkvlr\" (UniqueName: \"kubernetes.io/projected/08bd3753-3b70-44f8-92ff-98bedd88f662-kube-api-access-tkvlr\") pod \"ironic-operator-controller-manager-67cb4dc6d4-qtnxq\" (UID: \"08bd3753-3b70-44f8-92ff-98bedd88f662\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.760023 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.774185 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.775811 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.779156 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-dq898" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.786621 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5rsm\" (UniqueName: \"kubernetes.io/projected/a16c5563-750f-4e91-93bd-4044c656cb88-kube-api-access-p5rsm\") pod \"manila-operator-controller-manager-5d499bf58b-rs4kn\" (UID: \"a16c5563-750f-4e91-93bd-4044c656cb88\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.812350 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bhxg\" (UniqueName: \"kubernetes.io/projected/8969652f-14c9-4618-ad2d-1b3fd1ad3627-kube-api-access-8bhxg\") pod \"nova-operator-controller-manager-79556f57fc-dm7xb\" (UID: \"8969652f-14c9-4618-ad2d-1b3fd1ad3627\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.812438 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxmbh\" (UniqueName: \"kubernetes.io/projected/749a25e6-23eb-4fdd-8e15-ac2874cdee7f-kube-api-access-cxmbh\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-z865g\" (UID: \"749a25e6-23eb-4fdd-8e15-ac2874cdee7f\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.812598 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lccc6\" (UniqueName: \"kubernetes.io/projected/2bb23a3b-1f4a-409f-9d9c-0afe73a33352-kube-api-access-lccc6\") pod \"neutron-operator-controller-manager-6fdcddb789-wdzgw\" (UID: \"2bb23a3b-1f4a-409f-9d9c-0afe73a33352\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.812645 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9kb2\" (UniqueName: \"kubernetes.io/projected/91d3e9c8-8bbc-43b9-947b-71c4bb474b21-kube-api-access-n9kb2\") pod \"octavia-operator-controller-manager-64cdc6ff96-bfw5f\" (UID: \"91d3e9c8-8bbc-43b9-947b-71c4bb474b21\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.825627 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.838691 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxmbh\" (UniqueName: \"kubernetes.io/projected/749a25e6-23eb-4fdd-8e15-ac2874cdee7f-kube-api-access-cxmbh\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-z865g\" (UID: \"749a25e6-23eb-4fdd-8e15-ac2874cdee7f\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.838971 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.839047 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.842397 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lccc6\" (UniqueName: \"kubernetes.io/projected/2bb23a3b-1f4a-409f-9d9c-0afe73a33352-kube-api-access-lccc6\") pod \"neutron-operator-controller-manager-6fdcddb789-wdzgw\" (UID: \"2bb23a3b-1f4a-409f-9d9c-0afe73a33352\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.843971 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-f6npz" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.845157 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.879391 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.879443 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.883050 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.884677 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.887629 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-c7k77" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.919033 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzm2q\" (UniqueName: \"kubernetes.io/projected/a2e7c098-bf9d-4d99-84bf-adfb49ee6b12-kube-api-access-gzm2q\") pod \"ovn-operator-controller-manager-56897c768d-9zhql\" (UID: \"a2e7c098-bf9d-4d99-84bf-adfb49ee6b12\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.919072 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/150d5aa9-549d-4b43-b5c3-e6c81e942146-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t\" (UID: \"150d5aa9-549d-4b43-b5c3-e6c81e942146\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.919195 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9kb2\" (UniqueName: \"kubernetes.io/projected/91d3e9c8-8bbc-43b9-947b-71c4bb474b21-kube-api-access-n9kb2\") pod \"octavia-operator-controller-manager-64cdc6ff96-bfw5f\" (UID: \"91d3e9c8-8bbc-43b9-947b-71c4bb474b21\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.928716 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bhxg\" (UniqueName: \"kubernetes.io/projected/8969652f-14c9-4618-ad2d-1b3fd1ad3627-kube-api-access-8bhxg\") pod \"nova-operator-controller-manager-79556f57fc-dm7xb\" (UID: \"8969652f-14c9-4618-ad2d-1b3fd1ad3627\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.928814 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wm6m\" (UniqueName: \"kubernetes.io/projected/150d5aa9-549d-4b43-b5c3-e6c81e942146-kube-api-access-6wm6m\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t\" (UID: \"150d5aa9-549d-4b43-b5c3-e6c81e942146\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.931670 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.948824 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9kb2\" (UniqueName: \"kubernetes.io/projected/91d3e9c8-8bbc-43b9-947b-71c4bb474b21-kube-api-access-n9kb2\") pod \"octavia-operator-controller-manager-64cdc6ff96-bfw5f\" (UID: \"91d3e9c8-8bbc-43b9-947b-71c4bb474b21\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.975818 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bhxg\" (UniqueName: \"kubernetes.io/projected/8969652f-14c9-4618-ad2d-1b3fd1ad3627-kube-api-access-8bhxg\") pod \"nova-operator-controller-manager-79556f57fc-dm7xb\" (UID: \"8969652f-14c9-4618-ad2d-1b3fd1ad3627\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.975884 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2"] Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.990482 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.996721 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" Nov 26 00:42:11 crc kubenswrapper[4766]: I1126 00:42:11.997061 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-x9jkp" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.027949 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.039802 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wm6m\" (UniqueName: \"kubernetes.io/projected/150d5aa9-549d-4b43-b5c3-e6c81e942146-kube-api-access-6wm6m\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t\" (UID: \"150d5aa9-549d-4b43-b5c3-e6c81e942146\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.039887 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkpqv\" (UniqueName: \"kubernetes.io/projected/97419f32-8768-495f-a652-8a1e8701546f-kube-api-access-hkpqv\") pod \"placement-operator-controller-manager-57988cc5b5-xlxgj\" (UID: \"97419f32-8768-495f-a652-8a1e8701546f\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.039944 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzm2q\" (UniqueName: \"kubernetes.io/projected/a2e7c098-bf9d-4d99-84bf-adfb49ee6b12-kube-api-access-gzm2q\") pod \"ovn-operator-controller-manager-56897c768d-9zhql\" (UID: \"a2e7c098-bf9d-4d99-84bf-adfb49ee6b12\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.039993 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/150d5aa9-549d-4b43-b5c3-e6c81e942146-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t\" (UID: \"150d5aa9-549d-4b43-b5c3-e6c81e942146\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" Nov 26 00:42:12 crc kubenswrapper[4766]: E1126 00:42:12.040248 4766 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 00:42:12 crc kubenswrapper[4766]: E1126 00:42:12.040303 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/150d5aa9-549d-4b43-b5c3-e6c81e942146-cert podName:150d5aa9-549d-4b43-b5c3-e6c81e942146 nodeName:}" failed. No retries permitted until 2025-11-26 00:42:12.54028655 +0000 UTC m=+1113.389056980 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/150d5aa9-549d-4b43-b5c3-e6c81e942146-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" (UID: "150d5aa9-549d-4b43-b5c3-e6c81e942146") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.074914 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft"] Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.076590 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzm2q\" (UniqueName: \"kubernetes.io/projected/a2e7c098-bf9d-4d99-84bf-adfb49ee6b12-kube-api-access-gzm2q\") pod \"ovn-operator-controller-manager-56897c768d-9zhql\" (UID: \"a2e7c098-bf9d-4d99-84bf-adfb49ee6b12\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.077873 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.082084 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.085586 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wm6m\" (UniqueName: \"kubernetes.io/projected/150d5aa9-549d-4b43-b5c3-e6c81e942146-kube-api-access-6wm6m\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t\" (UID: \"150d5aa9-549d-4b43-b5c3-e6c81e942146\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.095023 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2"] Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.109772 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.111195 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-kkdmz" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.143344 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8d97\" (UniqueName: \"kubernetes.io/projected/00a32e6c-edcc-4625-ac35-26fed19f1638-kube-api-access-w8d97\") pod \"swift-operator-controller-manager-d77b94747-2t9w2\" (UID: \"00a32e6c-edcc-4625-ac35-26fed19f1638\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.143477 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-cert\") pod \"infra-operator-controller-manager-57548d458d-x6h4v\" (UID: \"c3cec209-fdef-44a6-90d0-2d06cfd7e5cc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.143504 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkpqv\" (UniqueName: \"kubernetes.io/projected/97419f32-8768-495f-a652-8a1e8701546f-kube-api-access-hkpqv\") pod \"placement-operator-controller-manager-57988cc5b5-xlxgj\" (UID: \"97419f32-8768-495f-a652-8a1e8701546f\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" Nov 26 00:42:12 crc kubenswrapper[4766]: E1126 00:42:12.144000 4766 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 26 00:42:12 crc kubenswrapper[4766]: E1126 00:42:12.144062 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-cert podName:c3cec209-fdef-44a6-90d0-2d06cfd7e5cc nodeName:}" failed. No retries permitted until 2025-11-26 00:42:13.144040982 +0000 UTC m=+1113.992811442 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-cert") pod "infra-operator-controller-manager-57548d458d-x6h4v" (UID: "c3cec209-fdef-44a6-90d0-2d06cfd7e5cc") : secret "infra-operator-webhook-server-cert" not found Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.144431 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft"] Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.144783 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.156514 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-glkl6"] Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.159340 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-glkl6" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.162707 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.165703 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-glkl6"] Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.169375 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkpqv\" (UniqueName: \"kubernetes.io/projected/97419f32-8768-495f-a652-8a1e8701546f-kube-api-access-hkpqv\") pod \"placement-operator-controller-manager-57988cc5b5-xlxgj\" (UID: \"97419f32-8768-495f-a652-8a1e8701546f\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.169702 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-kdksf" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.178759 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8"] Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.180388 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.199462 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.202561 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8"] Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.202986 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-5b4dx" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.220430 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.234692 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m"] Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.241998 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.244326 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpcdc\" (UniqueName: \"kubernetes.io/projected/40780971-20ed-4b4c-8d19-e74f6582ef8c-kube-api-access-qpcdc\") pod \"test-operator-controller-manager-5cd6c7f4c8-glkl6\" (UID: \"40780971-20ed-4b4c-8d19-e74f6582ef8c\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-glkl6" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.244385 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8drj\" (UniqueName: \"kubernetes.io/projected/04e53d09-de84-4682-b17b-61f9ea7bf152-kube-api-access-r8drj\") pod \"telemetry-operator-controller-manager-5ccf99cfdd-9cfft\" (UID: \"04e53d09-de84-4682-b17b-61f9ea7bf152\") " pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.244451 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9scq\" (UniqueName: \"kubernetes.io/projected/751b8e60-46d4-4f5e-b62d-800f641ccd40-kube-api-access-z9scq\") pod \"watcher-operator-controller-manager-656dcb59d4-2hpg8\" (UID: \"751b8e60-46d4-4f5e-b62d-800f641ccd40\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.244478 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8d97\" (UniqueName: \"kubernetes.io/projected/00a32e6c-edcc-4625-ac35-26fed19f1638-kube-api-access-w8d97\") pod \"swift-operator-controller-manager-d77b94747-2t9w2\" (UID: \"00a32e6c-edcc-4625-ac35-26fed19f1638\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.245715 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.261598 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.261947 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.262299 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-kbwvs" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.264788 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m"] Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.288050 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8d97\" (UniqueName: \"kubernetes.io/projected/00a32e6c-edcc-4625-ac35-26fed19f1638-kube-api-access-w8d97\") pod \"swift-operator-controller-manager-d77b94747-2t9w2\" (UID: \"00a32e6c-edcc-4625-ac35-26fed19f1638\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.289418 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw"] Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.290907 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.293987 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-m2b9k" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.310557 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw"] Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.345391 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9scq\" (UniqueName: \"kubernetes.io/projected/751b8e60-46d4-4f5e-b62d-800f641ccd40-kube-api-access-z9scq\") pod \"watcher-operator-controller-manager-656dcb59d4-2hpg8\" (UID: \"751b8e60-46d4-4f5e-b62d-800f641ccd40\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.345431 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-webhook-certs\") pod \"openstack-operator-controller-manager-86674bdb4b-vq98m\" (UID: \"d00b9c60-0fb7-49c6-944a-fac598cc22a4\") " pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.345484 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpcdc\" (UniqueName: \"kubernetes.io/projected/40780971-20ed-4b4c-8d19-e74f6582ef8c-kube-api-access-qpcdc\") pod \"test-operator-controller-manager-5cd6c7f4c8-glkl6\" (UID: \"40780971-20ed-4b4c-8d19-e74f6582ef8c\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-glkl6" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.345568 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-metrics-certs\") pod \"openstack-operator-controller-manager-86674bdb4b-vq98m\" (UID: \"d00b9c60-0fb7-49c6-944a-fac598cc22a4\") " pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.345584 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g967\" (UniqueName: \"kubernetes.io/projected/d00b9c60-0fb7-49c6-944a-fac598cc22a4-kube-api-access-9g967\") pod \"openstack-operator-controller-manager-86674bdb4b-vq98m\" (UID: \"d00b9c60-0fb7-49c6-944a-fac598cc22a4\") " pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.345604 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8drj\" (UniqueName: \"kubernetes.io/projected/04e53d09-de84-4682-b17b-61f9ea7bf152-kube-api-access-r8drj\") pod \"telemetry-operator-controller-manager-5ccf99cfdd-9cfft\" (UID: \"04e53d09-de84-4682-b17b-61f9ea7bf152\") " pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.345662 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgkkx\" (UniqueName: \"kubernetes.io/projected/91ee52b5-0e2e-4732-87bf-809e378d33d3-kube-api-access-cgkkx\") pod \"rabbitmq-cluster-operator-manager-668c99d594-mx9vw\" (UID: \"91ee52b5-0e2e-4732-87bf-809e378d33d3\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.349622 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-dqtcj"] Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.353467 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.375419 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8drj\" (UniqueName: \"kubernetes.io/projected/04e53d09-de84-4682-b17b-61f9ea7bf152-kube-api-access-r8drj\") pod \"telemetry-operator-controller-manager-5ccf99cfdd-9cfft\" (UID: \"04e53d09-de84-4682-b17b-61f9ea7bf152\") " pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.386715 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9scq\" (UniqueName: \"kubernetes.io/projected/751b8e60-46d4-4f5e-b62d-800f641ccd40-kube-api-access-z9scq\") pod \"watcher-operator-controller-manager-656dcb59d4-2hpg8\" (UID: \"751b8e60-46d4-4f5e-b62d-800f641ccd40\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.389913 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpcdc\" (UniqueName: \"kubernetes.io/projected/40780971-20ed-4b4c-8d19-e74f6582ef8c-kube-api-access-qpcdc\") pod \"test-operator-controller-manager-5cd6c7f4c8-glkl6\" (UID: \"40780971-20ed-4b4c-8d19-e74f6582ef8c\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-glkl6" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.461628 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgkkx\" (UniqueName: \"kubernetes.io/projected/91ee52b5-0e2e-4732-87bf-809e378d33d3-kube-api-access-cgkkx\") pod \"rabbitmq-cluster-operator-manager-668c99d594-mx9vw\" (UID: \"91ee52b5-0e2e-4732-87bf-809e378d33d3\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.461734 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-webhook-certs\") pod \"openstack-operator-controller-manager-86674bdb4b-vq98m\" (UID: \"d00b9c60-0fb7-49c6-944a-fac598cc22a4\") " pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.461929 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-metrics-certs\") pod \"openstack-operator-controller-manager-86674bdb4b-vq98m\" (UID: \"d00b9c60-0fb7-49c6-944a-fac598cc22a4\") " pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.461958 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g967\" (UniqueName: \"kubernetes.io/projected/d00b9c60-0fb7-49c6-944a-fac598cc22a4-kube-api-access-9g967\") pod \"openstack-operator-controller-manager-86674bdb4b-vq98m\" (UID: \"d00b9c60-0fb7-49c6-944a-fac598cc22a4\") " pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:12 crc kubenswrapper[4766]: E1126 00:42:12.462128 4766 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 26 00:42:12 crc kubenswrapper[4766]: E1126 00:42:12.462209 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-webhook-certs podName:d00b9c60-0fb7-49c6-944a-fac598cc22a4 nodeName:}" failed. No retries permitted until 2025-11-26 00:42:12.962175501 +0000 UTC m=+1113.810946001 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-webhook-certs") pod "openstack-operator-controller-manager-86674bdb4b-vq98m" (UID: "d00b9c60-0fb7-49c6-944a-fac598cc22a4") : secret "webhook-server-cert" not found Nov 26 00:42:12 crc kubenswrapper[4766]: E1126 00:42:12.462418 4766 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 26 00:42:12 crc kubenswrapper[4766]: E1126 00:42:12.462471 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-metrics-certs podName:d00b9c60-0fb7-49c6-944a-fac598cc22a4 nodeName:}" failed. No retries permitted until 2025-11-26 00:42:12.962454128 +0000 UTC m=+1113.811224648 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-metrics-certs") pod "openstack-operator-controller-manager-86674bdb4b-vq98m" (UID: "d00b9c60-0fb7-49c6-944a-fac598cc22a4") : secret "metrics-server-cert" not found Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.484610 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgkkx\" (UniqueName: \"kubernetes.io/projected/91ee52b5-0e2e-4732-87bf-809e378d33d3-kube-api-access-cgkkx\") pod \"rabbitmq-cluster-operator-manager-668c99d594-mx9vw\" (UID: \"91ee52b5-0e2e-4732-87bf-809e378d33d3\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.504693 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g967\" (UniqueName: \"kubernetes.io/projected/d00b9c60-0fb7-49c6-944a-fac598cc22a4-kube-api-access-9g967\") pod \"openstack-operator-controller-manager-86674bdb4b-vq98m\" (UID: \"d00b9c60-0fb7-49c6-944a-fac598cc22a4\") " pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.563286 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.563786 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/150d5aa9-549d-4b43-b5c3-e6c81e942146-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t\" (UID: \"150d5aa9-549d-4b43-b5c3-e6c81e942146\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" Nov 26 00:42:12 crc kubenswrapper[4766]: E1126 00:42:12.563983 4766 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 00:42:12 crc kubenswrapper[4766]: E1126 00:42:12.564033 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/150d5aa9-549d-4b43-b5c3-e6c81e942146-cert podName:150d5aa9-549d-4b43-b5c3-e6c81e942146 nodeName:}" failed. No retries permitted until 2025-11-26 00:42:13.564014445 +0000 UTC m=+1114.412784885 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/150d5aa9-549d-4b43-b5c3-e6c81e942146-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" (UID: "150d5aa9-549d-4b43-b5c3-e6c81e942146") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.601267 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-glkl6" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.627081 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.629268 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c"] Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.658867 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.971948 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-metrics-certs\") pod \"openstack-operator-controller-manager-86674bdb4b-vq98m\" (UID: \"d00b9c60-0fb7-49c6-944a-fac598cc22a4\") " pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:12 crc kubenswrapper[4766]: I1126 00:42:12.972354 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-webhook-certs\") pod \"openstack-operator-controller-manager-86674bdb4b-vq98m\" (UID: \"d00b9c60-0fb7-49c6-944a-fac598cc22a4\") " pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:12 crc kubenswrapper[4766]: E1126 00:42:12.972129 4766 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 26 00:42:12 crc kubenswrapper[4766]: E1126 00:42:12.972448 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-metrics-certs podName:d00b9c60-0fb7-49c6-944a-fac598cc22a4 nodeName:}" failed. No retries permitted until 2025-11-26 00:42:13.972428987 +0000 UTC m=+1114.821199487 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-metrics-certs") pod "openstack-operator-controller-manager-86674bdb4b-vq98m" (UID: "d00b9c60-0fb7-49c6-944a-fac598cc22a4") : secret "metrics-server-cert" not found Nov 26 00:42:12 crc kubenswrapper[4766]: E1126 00:42:12.972516 4766 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 26 00:42:12 crc kubenswrapper[4766]: E1126 00:42:12.972549 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-webhook-certs podName:d00b9c60-0fb7-49c6-944a-fac598cc22a4 nodeName:}" failed. No retries permitted until 2025-11-26 00:42:13.97254049 +0000 UTC m=+1114.821310910 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-webhook-certs") pod "openstack-operator-controller-manager-86674bdb4b-vq98m" (UID: "d00b9c60-0fb7-49c6-944a-fac598cc22a4") : secret "webhook-server-cert" not found Nov 26 00:42:13 crc kubenswrapper[4766]: I1126 00:42:13.125022 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" event={"ID":"31b1b73b-1d10-49ac-8549-b202c965b63e","Type":"ContainerStarted","Data":"bdec2c340233f74547ebdc93ba48ac4104bbc1efc92732877cdf8a3076092eca"} Nov 26 00:42:13 crc kubenswrapper[4766]: I1126 00:42:13.126533 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" event={"ID":"0e6cfef3-86af-4cf0-ab06-b3eef687836b","Type":"ContainerStarted","Data":"21a3391be2b617f1915dddfb0ec09df96944aaa4c21f5fe3d98eb7913af54901"} Nov 26 00:42:13 crc kubenswrapper[4766]: I1126 00:42:13.372627 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-cert\") pod \"infra-operator-controller-manager-57548d458d-x6h4v\" (UID: \"c3cec209-fdef-44a6-90d0-2d06cfd7e5cc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:42:13 crc kubenswrapper[4766]: E1126 00:42:13.376913 4766 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 26 00:42:13 crc kubenswrapper[4766]: E1126 00:42:13.377134 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-cert podName:c3cec209-fdef-44a6-90d0-2d06cfd7e5cc nodeName:}" failed. No retries permitted until 2025-11-26 00:42:15.377109136 +0000 UTC m=+1116.225879566 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-cert") pod "infra-operator-controller-manager-57548d458d-x6h4v" (UID: "c3cec209-fdef-44a6-90d0-2d06cfd7e5cc") : secret "infra-operator-webhook-server-cert" not found Nov 26 00:42:13 crc kubenswrapper[4766]: I1126 00:42:13.581935 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/150d5aa9-549d-4b43-b5c3-e6c81e942146-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t\" (UID: \"150d5aa9-549d-4b43-b5c3-e6c81e942146\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" Nov 26 00:42:13 crc kubenswrapper[4766]: E1126 00:42:13.582116 4766 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 00:42:13 crc kubenswrapper[4766]: E1126 00:42:13.582186 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/150d5aa9-549d-4b43-b5c3-e6c81e942146-cert podName:150d5aa9-549d-4b43-b5c3-e6c81e942146 nodeName:}" failed. No retries permitted until 2025-11-26 00:42:15.582167789 +0000 UTC m=+1116.430938219 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/150d5aa9-549d-4b43-b5c3-e6c81e942146-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" (UID: "150d5aa9-549d-4b43-b5c3-e6c81e942146") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 00:42:13 crc kubenswrapper[4766]: I1126 00:42:13.614577 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9"] Nov 26 00:42:13 crc kubenswrapper[4766]: I1126 00:42:13.622602 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b"] Nov 26 00:42:13 crc kubenswrapper[4766]: I1126 00:42:13.884475 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn"] Nov 26 00:42:13 crc kubenswrapper[4766]: I1126 00:42:13.892288 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz"] Nov 26 00:42:13 crc kubenswrapper[4766]: W1126 00:42:13.893071 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc87f631_ba4e_4943_b51d_bf1012cca754.slice/crio-af16bc6de132eb354554b2255ab7202023cc4ba996afaebedec54642bfb74795 WatchSource:0}: Error finding container af16bc6de132eb354554b2255ab7202023cc4ba996afaebedec54642bfb74795: Status 404 returned error can't find the container with id af16bc6de132eb354554b2255ab7202023cc4ba996afaebedec54642bfb74795 Nov 26 00:42:13 crc kubenswrapper[4766]: W1126 00:42:13.898207 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5284ad70_a36c_4666_8202_84633d83f50b.slice/crio-a35cf14f2bfe5a56a22c9ac42741e2ab610f40584419cb9a275454d5499c0a7e WatchSource:0}: Error finding container a35cf14f2bfe5a56a22c9ac42741e2ab610f40584419cb9a275454d5499c0a7e: Status 404 returned error can't find the container with id a35cf14f2bfe5a56a22c9ac42741e2ab610f40584419cb9a275454d5499c0a7e Nov 26 00:42:13 crc kubenswrapper[4766]: I1126 00:42:13.994211 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-webhook-certs\") pod \"openstack-operator-controller-manager-86674bdb4b-vq98m\" (UID: \"d00b9c60-0fb7-49c6-944a-fac598cc22a4\") " pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:13 crc kubenswrapper[4766]: I1126 00:42:13.994357 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-metrics-certs\") pod \"openstack-operator-controller-manager-86674bdb4b-vq98m\" (UID: \"d00b9c60-0fb7-49c6-944a-fac598cc22a4\") " pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:13 crc kubenswrapper[4766]: E1126 00:42:13.994518 4766 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 26 00:42:13 crc kubenswrapper[4766]: E1126 00:42:13.994561 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-metrics-certs podName:d00b9c60-0fb7-49c6-944a-fac598cc22a4 nodeName:}" failed. No retries permitted until 2025-11-26 00:42:15.99454756 +0000 UTC m=+1116.843317990 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-metrics-certs") pod "openstack-operator-controller-manager-86674bdb4b-vq98m" (UID: "d00b9c60-0fb7-49c6-944a-fac598cc22a4") : secret "metrics-server-cert" not found Nov 26 00:42:13 crc kubenswrapper[4766]: E1126 00:42:13.994793 4766 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 26 00:42:13 crc kubenswrapper[4766]: E1126 00:42:13.995741 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-webhook-certs podName:d00b9c60-0fb7-49c6-944a-fac598cc22a4 nodeName:}" failed. No retries permitted until 2025-11-26 00:42:15.995725949 +0000 UTC m=+1116.844496379 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-webhook-certs") pod "openstack-operator-controller-manager-86674bdb4b-vq98m" (UID: "d00b9c60-0fb7-49c6-944a-fac598cc22a4") : secret "webhook-server-cert" not found Nov 26 00:42:14 crc kubenswrapper[4766]: I1126 00:42:14.135764 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" event={"ID":"45353fc4-9955-460b-9276-9792dfdaf8f4","Type":"ContainerStarted","Data":"329ff93413ed233873912167105d14d453ddfd39e5d394dfa9a49873bb7309d5"} Nov 26 00:42:14 crc kubenswrapper[4766]: I1126 00:42:14.136818 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" event={"ID":"1ecbd1d2-cb7f-432f-bc75-903d0643ea29","Type":"ContainerStarted","Data":"512a39e916645b84e4065838538338798d2b19d6b23900f2f379aa3b68bdc52c"} Nov 26 00:42:14 crc kubenswrapper[4766]: I1126 00:42:14.138905 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" event={"ID":"cc87f631-ba4e-4943-b51d-bf1012cca754","Type":"ContainerStarted","Data":"af16bc6de132eb354554b2255ab7202023cc4ba996afaebedec54642bfb74795"} Nov 26 00:42:14 crc kubenswrapper[4766]: I1126 00:42:14.140186 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" event={"ID":"5284ad70-a36c-4666-8202-84633d83f50b","Type":"ContainerStarted","Data":"a35cf14f2bfe5a56a22c9ac42741e2ab610f40584419cb9a275454d5499c0a7e"} Nov 26 00:42:14 crc kubenswrapper[4766]: I1126 00:42:14.347823 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw"] Nov 26 00:42:14 crc kubenswrapper[4766]: W1126 00:42:14.359876 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2bb23a3b_1f4a_409f_9d9c_0afe73a33352.slice/crio-10788554c963508fe63998ff42ded53c8183ecf3699810e3a79cba213400b279 WatchSource:0}: Error finding container 10788554c963508fe63998ff42ded53c8183ecf3699810e3a79cba213400b279: Status 404 returned error can't find the container with id 10788554c963508fe63998ff42ded53c8183ecf3699810e3a79cba213400b279 Nov 26 00:42:14 crc kubenswrapper[4766]: I1126 00:42:14.364108 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f"] Nov 26 00:42:14 crc kubenswrapper[4766]: I1126 00:42:14.390062 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq"] Nov 26 00:42:14 crc kubenswrapper[4766]: I1126 00:42:14.405725 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2"] Nov 26 00:42:14 crc kubenswrapper[4766]: W1126 00:42:14.411114 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40780971_20ed_4b4c_8d19_e74f6582ef8c.slice/crio-cd8c661a345109c74e395f0d9d748ff2e5260ef29e923469607ab03d3e855e34 WatchSource:0}: Error finding container cd8c661a345109c74e395f0d9d748ff2e5260ef29e923469607ab03d3e855e34: Status 404 returned error can't find the container with id cd8c661a345109c74e395f0d9d748ff2e5260ef29e923469607ab03d3e855e34 Nov 26 00:42:14 crc kubenswrapper[4766]: I1126 00:42:14.412893 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql"] Nov 26 00:42:14 crc kubenswrapper[4766]: I1126 00:42:14.429061 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft"] Nov 26 00:42:14 crc kubenswrapper[4766]: W1126 00:42:14.434625 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04e53d09_de84_4682_b17b_61f9ea7bf152.slice/crio-c57f0b070ffc93a4d6f5ad20fcd9511a644bd3caf1237f5f9bdaa05ac620b47c WatchSource:0}: Error finding container c57f0b070ffc93a4d6f5ad20fcd9511a644bd3caf1237f5f9bdaa05ac620b47c: Status 404 returned error can't find the container with id c57f0b070ffc93a4d6f5ad20fcd9511a644bd3caf1237f5f9bdaa05ac620b47c Nov 26 00:42:14 crc kubenswrapper[4766]: W1126 00:42:14.438692 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00a32e6c_edcc_4625_ac35_26fed19f1638.slice/crio-c3b737def77cc0a7cf849116a03d976e6e90b0cbaf9061ed846541738ffcdfb4 WatchSource:0}: Error finding container c3b737def77cc0a7cf849116a03d976e6e90b0cbaf9061ed846541738ffcdfb4: Status 404 returned error can't find the container with id c3b737def77cc0a7cf849116a03d976e6e90b0cbaf9061ed846541738ffcdfb4 Nov 26 00:42:14 crc kubenswrapper[4766]: I1126 00:42:14.442530 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn"] Nov 26 00:42:14 crc kubenswrapper[4766]: W1126 00:42:14.444204 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97419f32_8768_495f_a652_8a1e8701546f.slice/crio-f69d6cf8d886daa39ebb6b18cd9b535c495957feafc2c002ca49b3e5623c133c WatchSource:0}: Error finding container f69d6cf8d886daa39ebb6b18cd9b535c495957feafc2c002ca49b3e5623c133c: Status 404 returned error can't find the container with id f69d6cf8d886daa39ebb6b18cd9b535c495957feafc2c002ca49b3e5623c133c Nov 26 00:42:14 crc kubenswrapper[4766]: W1126 00:42:14.445878 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda16c5563_750f_4e91_93bd_4044c656cb88.slice/crio-1755f3ec6e7769e0444b42c56210036491f269ec4a996a070c234c5f5b0fcb83 WatchSource:0}: Error finding container 1755f3ec6e7769e0444b42c56210036491f269ec4a996a070c234c5f5b0fcb83: Status 404 returned error can't find the container with id 1755f3ec6e7769e0444b42c56210036491f269ec4a996a070c234c5f5b0fcb83 Nov 26 00:42:14 crc kubenswrapper[4766]: W1126 00:42:14.453820 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2e7c098_bf9d_4d99_84bf_adfb49ee6b12.slice/crio-270d78353a7337926026aca120a6ef2aee454e74fa1f7b2783a85a7d4aa59306 WatchSource:0}: Error finding container 270d78353a7337926026aca120a6ef2aee454e74fa1f7b2783a85a7d4aa59306: Status 404 returned error can't find the container with id 270d78353a7337926026aca120a6ef2aee454e74fa1f7b2783a85a7d4aa59306 Nov 26 00:42:14 crc kubenswrapper[4766]: W1126 00:42:14.457041 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod749a25e6_23eb_4fdd_8e15_ac2874cdee7f.slice/crio-886119a6cb002624679afa6c8c80d9acc57ab4b23d1a395f1e47b3f534f62f52 WatchSource:0}: Error finding container 886119a6cb002624679afa6c8c80d9acc57ab4b23d1a395f1e47b3f534f62f52: Status 404 returned error can't find the container with id 886119a6cb002624679afa6c8c80d9acc57ab4b23d1a395f1e47b3f534f62f52 Nov 26 00:42:14 crc kubenswrapper[4766]: W1126 00:42:14.459972 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d0d59b2_0486_40a2_ae60_b738a4f76dc4.slice/crio-cc47626890fba42a8e46d06d73d12ea3bf25620ee8ec900ecd25d722c8868bdd WatchSource:0}: Error finding container cc47626890fba42a8e46d06d73d12ea3bf25620ee8ec900ecd25d722c8868bdd: Status 404 returned error can't find the container with id cc47626890fba42a8e46d06d73d12ea3bf25620ee8ec900ecd25d722c8868bdd Nov 26 00:42:14 crc kubenswrapper[4766]: E1126 00:42:14.461066 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:888edf6f432e52eaa5fc3caeae616fe38a3302b006bbba0e38885b2beba9f0f2,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cxmbh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-66f4dd4bc7-z865g_openstack-operators(749a25e6-23eb-4fdd-8e15-ac2874cdee7f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 00:42:14 crc kubenswrapper[4766]: E1126 00:42:14.463891 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cxmbh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-66f4dd4bc7-z865g_openstack-operators(749a25e6-23eb-4fdd-8e15-ac2874cdee7f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 00:42:14 crc kubenswrapper[4766]: E1126 00:42:14.464024 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:25faa5b0e4801d4d3b01a28b877ed3188eee71f33ad66f3c2e86b7921758e711,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6vsp8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7b4567c7cf-fcr76_openstack-operators(3d0d59b2-0486-40a2-ae60-b738a4f76dc4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 00:42:14 crc kubenswrapper[4766]: E1126 00:42:14.465777 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" podUID="749a25e6-23eb-4fdd-8e15-ac2874cdee7f" Nov 26 00:42:14 crc kubenswrapper[4766]: E1126 00:42:14.468696 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6vsp8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7b4567c7cf-fcr76_openstack-operators(3d0d59b2-0486-40a2-ae60-b738a4f76dc4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 00:42:14 crc kubenswrapper[4766]: E1126 00:42:14.470054 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" podUID="3d0d59b2-0486-40a2-ae60-b738a4f76dc4" Nov 26 00:42:14 crc kubenswrapper[4766]: E1126 00:42:14.470334 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8bhxg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-79556f57fc-dm7xb_openstack-operators(8969652f-14c9-4618-ad2d-1b3fd1ad3627): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 00:42:14 crc kubenswrapper[4766]: E1126 00:42:14.473361 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8bhxg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-79556f57fc-dm7xb_openstack-operators(8969652f-14c9-4618-ad2d-1b3fd1ad3627): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 00:42:14 crc kubenswrapper[4766]: E1126 00:42:14.474580 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" podUID="8969652f-14c9-4618-ad2d-1b3fd1ad3627" Nov 26 00:42:14 crc kubenswrapper[4766]: I1126 00:42:14.485410 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-glkl6"] Nov 26 00:42:14 crc kubenswrapper[4766]: I1126 00:42:14.516829 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj"] Nov 26 00:42:14 crc kubenswrapper[4766]: I1126 00:42:14.534836 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g"] Nov 26 00:42:14 crc kubenswrapper[4766]: I1126 00:42:14.550151 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb"] Nov 26 00:42:14 crc kubenswrapper[4766]: W1126 00:42:14.553143 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod751b8e60_46d4_4f5e_b62d_800f641ccd40.slice/crio-6c4643b73479cbc6feffae558d921413d85a9f8a647130307cc1d06d2cbea351 WatchSource:0}: Error finding container 6c4643b73479cbc6feffae558d921413d85a9f8a647130307cc1d06d2cbea351: Status 404 returned error can't find the container with id 6c4643b73479cbc6feffae558d921413d85a9f8a647130307cc1d06d2cbea351 Nov 26 00:42:14 crc kubenswrapper[4766]: I1126 00:42:14.555732 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76"] Nov 26 00:42:14 crc kubenswrapper[4766]: I1126 00:42:14.562096 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8"] Nov 26 00:42:14 crc kubenswrapper[4766]: I1126 00:42:14.567141 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw"] Nov 26 00:42:14 crc kubenswrapper[4766]: W1126 00:42:14.573273 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91ee52b5_0e2e_4732_87bf_809e378d33d3.slice/crio-b2b78ed128ab8a9453a619f8acc3ca1db56dbbd38a4c76762f6955e8a56f8261 WatchSource:0}: Error finding container b2b78ed128ab8a9453a619f8acc3ca1db56dbbd38a4c76762f6955e8a56f8261: Status 404 returned error can't find the container with id b2b78ed128ab8a9453a619f8acc3ca1db56dbbd38a4c76762f6955e8a56f8261 Nov 26 00:42:14 crc kubenswrapper[4766]: E1126 00:42:14.577869 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cgkkx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-mx9vw_openstack-operators(91ee52b5-0e2e-4732-87bf-809e378d33d3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 00:42:14 crc kubenswrapper[4766]: E1126 00:42:14.579397 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw" podUID="91ee52b5-0e2e-4732-87bf-809e378d33d3" Nov 26 00:42:15 crc kubenswrapper[4766]: I1126 00:42:15.150677 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" event={"ID":"751b8e60-46d4-4f5e-b62d-800f641ccd40","Type":"ContainerStarted","Data":"6c4643b73479cbc6feffae558d921413d85a9f8a647130307cc1d06d2cbea351"} Nov 26 00:42:15 crc kubenswrapper[4766]: I1126 00:42:15.152155 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" event={"ID":"00a32e6c-edcc-4625-ac35-26fed19f1638","Type":"ContainerStarted","Data":"c3b737def77cc0a7cf849116a03d976e6e90b0cbaf9061ed846541738ffcdfb4"} Nov 26 00:42:15 crc kubenswrapper[4766]: I1126 00:42:15.153349 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw" event={"ID":"91ee52b5-0e2e-4732-87bf-809e378d33d3","Type":"ContainerStarted","Data":"b2b78ed128ab8a9453a619f8acc3ca1db56dbbd38a4c76762f6955e8a56f8261"} Nov 26 00:42:15 crc kubenswrapper[4766]: I1126 00:42:15.154467 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" event={"ID":"8969652f-14c9-4618-ad2d-1b3fd1ad3627","Type":"ContainerStarted","Data":"7fce33e195885755594701d6054f676e540cdfd196b7c58015d40a100e1806fc"} Nov 26 00:42:15 crc kubenswrapper[4766]: E1126 00:42:15.154761 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw" podUID="91ee52b5-0e2e-4732-87bf-809e378d33d3" Nov 26 00:42:15 crc kubenswrapper[4766]: E1126 00:42:15.155914 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" podUID="8969652f-14c9-4618-ad2d-1b3fd1ad3627" Nov 26 00:42:15 crc kubenswrapper[4766]: I1126 00:42:15.156042 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" event={"ID":"2bb23a3b-1f4a-409f-9d9c-0afe73a33352","Type":"ContainerStarted","Data":"10788554c963508fe63998ff42ded53c8183ecf3699810e3a79cba213400b279"} Nov 26 00:42:15 crc kubenswrapper[4766]: I1126 00:42:15.158368 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" event={"ID":"749a25e6-23eb-4fdd-8e15-ac2874cdee7f","Type":"ContainerStarted","Data":"886119a6cb002624679afa6c8c80d9acc57ab4b23d1a395f1e47b3f534f62f52"} Nov 26 00:42:15 crc kubenswrapper[4766]: I1126 00:42:15.161041 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-glkl6" event={"ID":"40780971-20ed-4b4c-8d19-e74f6582ef8c","Type":"ContainerStarted","Data":"cd8c661a345109c74e395f0d9d748ff2e5260ef29e923469607ab03d3e855e34"} Nov 26 00:42:15 crc kubenswrapper[4766]: I1126 00:42:15.162578 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" event={"ID":"91d3e9c8-8bbc-43b9-947b-71c4bb474b21","Type":"ContainerStarted","Data":"c6f82d59070828dda6e403355385ba2b1877280e8ac9ea1d57d7a4897a74d9f0"} Nov 26 00:42:15 crc kubenswrapper[4766]: I1126 00:42:15.165558 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn" event={"ID":"a16c5563-750f-4e91-93bd-4044c656cb88","Type":"ContainerStarted","Data":"1755f3ec6e7769e0444b42c56210036491f269ec4a996a070c234c5f5b0fcb83"} Nov 26 00:42:15 crc kubenswrapper[4766]: I1126 00:42:15.169803 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" event={"ID":"3d0d59b2-0486-40a2-ae60-b738a4f76dc4","Type":"ContainerStarted","Data":"cc47626890fba42a8e46d06d73d12ea3bf25620ee8ec900ecd25d722c8868bdd"} Nov 26 00:42:15 crc kubenswrapper[4766]: I1126 00:42:15.172956 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" event={"ID":"97419f32-8768-495f-a652-8a1e8701546f","Type":"ContainerStarted","Data":"f69d6cf8d886daa39ebb6b18cd9b535c495957feafc2c002ca49b3e5623c133c"} Nov 26 00:42:15 crc kubenswrapper[4766]: E1126 00:42:15.172967 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:888edf6f432e52eaa5fc3caeae616fe38a3302b006bbba0e38885b2beba9f0f2\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" podUID="749a25e6-23eb-4fdd-8e15-ac2874cdee7f" Nov 26 00:42:15 crc kubenswrapper[4766]: E1126 00:42:15.175061 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:25faa5b0e4801d4d3b01a28b877ed3188eee71f33ad66f3c2e86b7921758e711\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" podUID="3d0d59b2-0486-40a2-ae60-b738a4f76dc4" Nov 26 00:42:15 crc kubenswrapper[4766]: I1126 00:42:15.179685 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" event={"ID":"08bd3753-3b70-44f8-92ff-98bedd88f662","Type":"ContainerStarted","Data":"6a8f4162cccb0b7c6780b337e7d1a34a09855289e761ab39db7b56de8b0065c5"} Nov 26 00:42:15 crc kubenswrapper[4766]: I1126 00:42:15.182386 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" event={"ID":"a2e7c098-bf9d-4d99-84bf-adfb49ee6b12","Type":"ContainerStarted","Data":"270d78353a7337926026aca120a6ef2aee454e74fa1f7b2783a85a7d4aa59306"} Nov 26 00:42:15 crc kubenswrapper[4766]: I1126 00:42:15.217283 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" event={"ID":"04e53d09-de84-4682-b17b-61f9ea7bf152","Type":"ContainerStarted","Data":"c57f0b070ffc93a4d6f5ad20fcd9511a644bd3caf1237f5f9bdaa05ac620b47c"} Nov 26 00:42:15 crc kubenswrapper[4766]: I1126 00:42:15.426880 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-cert\") pod \"infra-operator-controller-manager-57548d458d-x6h4v\" (UID: \"c3cec209-fdef-44a6-90d0-2d06cfd7e5cc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:42:15 crc kubenswrapper[4766]: E1126 00:42:15.427057 4766 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 26 00:42:15 crc kubenswrapper[4766]: E1126 00:42:15.427101 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-cert podName:c3cec209-fdef-44a6-90d0-2d06cfd7e5cc nodeName:}" failed. No retries permitted until 2025-11-26 00:42:19.427087666 +0000 UTC m=+1120.275858096 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-cert") pod "infra-operator-controller-manager-57548d458d-x6h4v" (UID: "c3cec209-fdef-44a6-90d0-2d06cfd7e5cc") : secret "infra-operator-webhook-server-cert" not found Nov 26 00:42:15 crc kubenswrapper[4766]: I1126 00:42:15.629810 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/150d5aa9-549d-4b43-b5c3-e6c81e942146-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t\" (UID: \"150d5aa9-549d-4b43-b5c3-e6c81e942146\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" Nov 26 00:42:15 crc kubenswrapper[4766]: E1126 00:42:15.629930 4766 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 00:42:15 crc kubenswrapper[4766]: E1126 00:42:15.629995 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/150d5aa9-549d-4b43-b5c3-e6c81e942146-cert podName:150d5aa9-549d-4b43-b5c3-e6c81e942146 nodeName:}" failed. No retries permitted until 2025-11-26 00:42:19.629975374 +0000 UTC m=+1120.478745804 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/150d5aa9-549d-4b43-b5c3-e6c81e942146-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" (UID: "150d5aa9-549d-4b43-b5c3-e6c81e942146") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 00:42:16 crc kubenswrapper[4766]: I1126 00:42:16.035815 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-metrics-certs\") pod \"openstack-operator-controller-manager-86674bdb4b-vq98m\" (UID: \"d00b9c60-0fb7-49c6-944a-fac598cc22a4\") " pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:16 crc kubenswrapper[4766]: I1126 00:42:16.036189 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-webhook-certs\") pod \"openstack-operator-controller-manager-86674bdb4b-vq98m\" (UID: \"d00b9c60-0fb7-49c6-944a-fac598cc22a4\") " pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:16 crc kubenswrapper[4766]: E1126 00:42:16.035996 4766 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 26 00:42:16 crc kubenswrapper[4766]: E1126 00:42:16.036296 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-metrics-certs podName:d00b9c60-0fb7-49c6-944a-fac598cc22a4 nodeName:}" failed. No retries permitted until 2025-11-26 00:42:20.036274083 +0000 UTC m=+1120.885044563 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-metrics-certs") pod "openstack-operator-controller-manager-86674bdb4b-vq98m" (UID: "d00b9c60-0fb7-49c6-944a-fac598cc22a4") : secret "metrics-server-cert" not found Nov 26 00:42:16 crc kubenswrapper[4766]: E1126 00:42:16.036369 4766 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 26 00:42:16 crc kubenswrapper[4766]: E1126 00:42:16.036417 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-webhook-certs podName:d00b9c60-0fb7-49c6-944a-fac598cc22a4 nodeName:}" failed. No retries permitted until 2025-11-26 00:42:20.036403747 +0000 UTC m=+1120.885174177 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-webhook-certs") pod "openstack-operator-controller-manager-86674bdb4b-vq98m" (UID: "d00b9c60-0fb7-49c6-944a-fac598cc22a4") : secret "webhook-server-cert" not found Nov 26 00:42:16 crc kubenswrapper[4766]: E1126 00:42:16.246383 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw" podUID="91ee52b5-0e2e-4732-87bf-809e378d33d3" Nov 26 00:42:16 crc kubenswrapper[4766]: E1126 00:42:16.247475 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:25faa5b0e4801d4d3b01a28b877ed3188eee71f33ad66f3c2e86b7921758e711\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" podUID="3d0d59b2-0486-40a2-ae60-b738a4f76dc4" Nov 26 00:42:16 crc kubenswrapper[4766]: E1126 00:42:16.247694 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" podUID="8969652f-14c9-4618-ad2d-1b3fd1ad3627" Nov 26 00:42:16 crc kubenswrapper[4766]: E1126 00:42:16.248288 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:888edf6f432e52eaa5fc3caeae616fe38a3302b006bbba0e38885b2beba9f0f2\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" podUID="749a25e6-23eb-4fdd-8e15-ac2874cdee7f" Nov 26 00:42:19 crc kubenswrapper[4766]: I1126 00:42:19.516481 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-cert\") pod \"infra-operator-controller-manager-57548d458d-x6h4v\" (UID: \"c3cec209-fdef-44a6-90d0-2d06cfd7e5cc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:42:19 crc kubenswrapper[4766]: E1126 00:42:19.516689 4766 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 26 00:42:19 crc kubenswrapper[4766]: E1126 00:42:19.517195 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-cert podName:c3cec209-fdef-44a6-90d0-2d06cfd7e5cc nodeName:}" failed. No retries permitted until 2025-11-26 00:42:27.517173868 +0000 UTC m=+1128.365944308 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-cert") pod "infra-operator-controller-manager-57548d458d-x6h4v" (UID: "c3cec209-fdef-44a6-90d0-2d06cfd7e5cc") : secret "infra-operator-webhook-server-cert" not found Nov 26 00:42:19 crc kubenswrapper[4766]: I1126 00:42:19.719941 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/150d5aa9-549d-4b43-b5c3-e6c81e942146-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t\" (UID: \"150d5aa9-549d-4b43-b5c3-e6c81e942146\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" Nov 26 00:42:19 crc kubenswrapper[4766]: E1126 00:42:19.720250 4766 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 00:42:19 crc kubenswrapper[4766]: E1126 00:42:19.720334 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/150d5aa9-549d-4b43-b5c3-e6c81e942146-cert podName:150d5aa9-549d-4b43-b5c3-e6c81e942146 nodeName:}" failed. No retries permitted until 2025-11-26 00:42:27.720319723 +0000 UTC m=+1128.569090153 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/150d5aa9-549d-4b43-b5c3-e6c81e942146-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" (UID: "150d5aa9-549d-4b43-b5c3-e6c81e942146") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 00:42:20 crc kubenswrapper[4766]: I1126 00:42:20.127319 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-metrics-certs\") pod \"openstack-operator-controller-manager-86674bdb4b-vq98m\" (UID: \"d00b9c60-0fb7-49c6-944a-fac598cc22a4\") " pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:20 crc kubenswrapper[4766]: I1126 00:42:20.127394 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-webhook-certs\") pod \"openstack-operator-controller-manager-86674bdb4b-vq98m\" (UID: \"d00b9c60-0fb7-49c6-944a-fac598cc22a4\") " pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:20 crc kubenswrapper[4766]: E1126 00:42:20.127507 4766 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 26 00:42:20 crc kubenswrapper[4766]: E1126 00:42:20.127526 4766 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 26 00:42:20 crc kubenswrapper[4766]: E1126 00:42:20.127591 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-webhook-certs podName:d00b9c60-0fb7-49c6-944a-fac598cc22a4 nodeName:}" failed. No retries permitted until 2025-11-26 00:42:28.127571816 +0000 UTC m=+1128.976342246 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-webhook-certs") pod "openstack-operator-controller-manager-86674bdb4b-vq98m" (UID: "d00b9c60-0fb7-49c6-944a-fac598cc22a4") : secret "webhook-server-cert" not found Nov 26 00:42:20 crc kubenswrapper[4766]: E1126 00:42:20.127609 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-metrics-certs podName:d00b9c60-0fb7-49c6-944a-fac598cc22a4 nodeName:}" failed. No retries permitted until 2025-11-26 00:42:28.127602487 +0000 UTC m=+1128.976372917 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-metrics-certs") pod "openstack-operator-controller-manager-86674bdb4b-vq98m" (UID: "d00b9c60-0fb7-49c6-944a-fac598cc22a4") : secret "metrics-server-cert" not found Nov 26 00:42:25 crc kubenswrapper[4766]: E1126 00:42:25.908121 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9413ed1bc2ae1a6bd28c59b1c7f7e91e1638de7b2a7d4729ed3fa2135182465d" Nov 26 00:42:25 crc kubenswrapper[4766]: E1126 00:42:25.908896 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9413ed1bc2ae1a6bd28c59b1c7f7e91e1638de7b2a7d4729ed3fa2135182465d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mqtsw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-5d494799bf-h4mrn_openstack-operators(5284ad70-a36c-4666-8202-84633d83f50b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 00:42:26 crc kubenswrapper[4766]: E1126 00:42:26.414966 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423" Nov 26 00:42:26 crc kubenswrapper[4766]: E1126 00:42:26.415120 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hkpqv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-57988cc5b5-xlxgj_openstack-operators(97419f32-8768-495f-a652-8a1e8701546f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 00:42:27 crc kubenswrapper[4766]: I1126 00:42:27.563930 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-cert\") pod \"infra-operator-controller-manager-57548d458d-x6h4v\" (UID: \"c3cec209-fdef-44a6-90d0-2d06cfd7e5cc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:42:27 crc kubenswrapper[4766]: E1126 00:42:27.564185 4766 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 26 00:42:27 crc kubenswrapper[4766]: E1126 00:42:27.564277 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-cert podName:c3cec209-fdef-44a6-90d0-2d06cfd7e5cc nodeName:}" failed. No retries permitted until 2025-11-26 00:42:43.564257656 +0000 UTC m=+1144.413028096 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-cert") pod "infra-operator-controller-manager-57548d458d-x6h4v" (UID: "c3cec209-fdef-44a6-90d0-2d06cfd7e5cc") : secret "infra-operator-webhook-server-cert" not found Nov 26 00:42:27 crc kubenswrapper[4766]: I1126 00:42:27.766279 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/150d5aa9-549d-4b43-b5c3-e6c81e942146-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t\" (UID: \"150d5aa9-549d-4b43-b5c3-e6c81e942146\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" Nov 26 00:42:27 crc kubenswrapper[4766]: E1126 00:42:27.766513 4766 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 00:42:27 crc kubenswrapper[4766]: E1126 00:42:27.767226 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/150d5aa9-549d-4b43-b5c3-e6c81e942146-cert podName:150d5aa9-549d-4b43-b5c3-e6c81e942146 nodeName:}" failed. No retries permitted until 2025-11-26 00:42:43.767205136 +0000 UTC m=+1144.615975566 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/150d5aa9-549d-4b43-b5c3-e6c81e942146-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" (UID: "150d5aa9-549d-4b43-b5c3-e6c81e942146") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 00:42:28 crc kubenswrapper[4766]: I1126 00:42:28.176132 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-webhook-certs\") pod \"openstack-operator-controller-manager-86674bdb4b-vq98m\" (UID: \"d00b9c60-0fb7-49c6-944a-fac598cc22a4\") " pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:28 crc kubenswrapper[4766]: I1126 00:42:28.176326 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-metrics-certs\") pod \"openstack-operator-controller-manager-86674bdb4b-vq98m\" (UID: \"d00b9c60-0fb7-49c6-944a-fac598cc22a4\") " pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:28 crc kubenswrapper[4766]: I1126 00:42:28.181985 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-webhook-certs\") pod \"openstack-operator-controller-manager-86674bdb4b-vq98m\" (UID: \"d00b9c60-0fb7-49c6-944a-fac598cc22a4\") " pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:28 crc kubenswrapper[4766]: I1126 00:42:28.182981 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d00b9c60-0fb7-49c6-944a-fac598cc22a4-metrics-certs\") pod \"openstack-operator-controller-manager-86674bdb4b-vq98m\" (UID: \"d00b9c60-0fb7-49c6-944a-fac598cc22a4\") " pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:28 crc kubenswrapper[4766]: I1126 00:42:28.240922 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:33 crc kubenswrapper[4766]: E1126 00:42:33.194281 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:3dbf9fd9dce75f1fb250ee4c4097ad77d2f34110b61d85e37abd9c472e022e6c" Nov 26 00:42:33 crc kubenswrapper[4766]: E1126 00:42:33.195890 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:3dbf9fd9dce75f1fb250ee4c4097ad77d2f34110b61d85e37abd9c472e022e6c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wvmlj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7b64f4fb85-gbdhz_openstack-operators(cc87f631-ba4e-4943-b51d-bf1012cca754): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 00:42:33 crc kubenswrapper[4766]: E1126 00:42:33.707624 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:45ae665ce2ea81aef212ee402cb02693ee49001a7c88c40c9598ff2859b838a2" Nov 26 00:42:33 crc kubenswrapper[4766]: E1126 00:42:33.708194 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:45ae665ce2ea81aef212ee402cb02693ee49001a7c88c40c9598ff2859b838a2,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qtbzh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-589cbd6b5b-k68q9_openstack-operators(45353fc4-9955-460b-9276-9792dfdaf8f4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 00:42:34 crc kubenswrapper[4766]: E1126 00:42:34.404056 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:bbb543d2d67c73e5df5d6357c3251363eb34a99575c5bf10416edd45dbdae2f6" Nov 26 00:42:34 crc kubenswrapper[4766]: E1126 00:42:34.404276 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:bbb543d2d67c73e5df5d6357c3251363eb34a99575c5bf10416edd45dbdae2f6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gzm2q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-56897c768d-9zhql_openstack-operators(a2e7c098-bf9d-4d99-84bf-adfb49ee6b12): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 00:42:34 crc kubenswrapper[4766]: E1126 00:42:34.899609 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:e00a9ed0ab26c5b745bd804ab1fe6b22428d026f17ea05a05f045e060342f46c" Nov 26 00:42:34 crc kubenswrapper[4766]: E1126 00:42:34.899865 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:e00a9ed0ab26c5b745bd804ab1fe6b22428d026f17ea05a05f045e060342f46c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lccc6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-6fdcddb789-wdzgw_openstack-operators(2bb23a3b-1f4a-409f-9d9c-0afe73a33352): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 00:42:34 crc kubenswrapper[4766]: E1126 00:42:34.975419 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.248:5001/openstack-k8s-operators/telemetry-operator:bf35154a77d3f7d42763b9d6bf295684481cdc52" Nov 26 00:42:34 crc kubenswrapper[4766]: E1126 00:42:34.976097 4766 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.248:5001/openstack-k8s-operators/telemetry-operator:bf35154a77d3f7d42763b9d6bf295684481cdc52" Nov 26 00:42:34 crc kubenswrapper[4766]: E1126 00:42:34.976334 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.248:5001/openstack-k8s-operators/telemetry-operator:bf35154a77d3f7d42763b9d6bf295684481cdc52,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r8drj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5ccf99cfdd-9cfft_openstack-operators(04e53d09-de84-4682-b17b-61f9ea7bf152): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 00:42:36 crc kubenswrapper[4766]: I1126 00:42:36.096162 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m"] Nov 26 00:42:36 crc kubenswrapper[4766]: W1126 00:42:36.247761 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd00b9c60_0fb7_49c6_944a_fac598cc22a4.slice/crio-9457e8126f8e9e82d5429d5aa23867de6de7e64d364078145881793f81c98ad6 WatchSource:0}: Error finding container 9457e8126f8e9e82d5429d5aa23867de6de7e64d364078145881793f81c98ad6: Status 404 returned error can't find the container with id 9457e8126f8e9e82d5429d5aa23867de6de7e64d364078145881793f81c98ad6 Nov 26 00:42:36 crc kubenswrapper[4766]: I1126 00:42:36.397933 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" event={"ID":"d00b9c60-0fb7-49c6-944a-fac598cc22a4","Type":"ContainerStarted","Data":"9457e8126f8e9e82d5429d5aa23867de6de7e64d364078145881793f81c98ad6"} Nov 26 00:42:40 crc kubenswrapper[4766]: I1126 00:42:40.489899 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" event={"ID":"d00b9c60-0fb7-49c6-944a-fac598cc22a4","Type":"ContainerStarted","Data":"35bfe8abff8194805e0514db2803bd9c014d0cb93366a09b57b6684dc470002d"} Nov 26 00:42:40 crc kubenswrapper[4766]: I1126 00:42:40.490657 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:40 crc kubenswrapper[4766]: I1126 00:42:40.518237 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" event={"ID":"00a32e6c-edcc-4625-ac35-26fed19f1638","Type":"ContainerStarted","Data":"90575fc710573493e63d6004e10aaf0090eed703ac12ecf62af5a2bbbe5bff23"} Nov 26 00:42:40 crc kubenswrapper[4766]: I1126 00:42:40.523668 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" event={"ID":"91d3e9c8-8bbc-43b9-947b-71c4bb474b21","Type":"ContainerStarted","Data":"5c707b0b94e22890a9fb6d29850435a4aca9ff411003ed90468f5dcf5adef6c1"} Nov 26 00:42:40 crc kubenswrapper[4766]: I1126 00:42:40.557297 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" event={"ID":"31b1b73b-1d10-49ac-8549-b202c965b63e","Type":"ContainerStarted","Data":"f1706c458c961181fe0f0079d94cf794bd65e747edbd15a93fd1d4c83ed950ce"} Nov 26 00:42:40 crc kubenswrapper[4766]: I1126 00:42:40.591677 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn" event={"ID":"a16c5563-750f-4e91-93bd-4044c656cb88","Type":"ContainerStarted","Data":"7a8ccd36fe913294b5a86afc0b6c563f7643d1547af8414850c17e3801762226"} Nov 26 00:42:40 crc kubenswrapper[4766]: I1126 00:42:40.597035 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" event={"ID":"0e6cfef3-86af-4cf0-ab06-b3eef687836b","Type":"ContainerStarted","Data":"b27130c628e9b6ed27e2271bbe305fc8f86a893172736a4bf2cd4a9053f2dc60"} Nov 26 00:42:41 crc kubenswrapper[4766]: I1126 00:42:41.480039 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:42:41 crc kubenswrapper[4766]: I1126 00:42:41.480335 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:42:41 crc kubenswrapper[4766]: I1126 00:42:41.480379 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:42:41 crc kubenswrapper[4766]: I1126 00:42:41.481008 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cef0ca8e4338b5dbd39bc68daed18f6f5044b46711899f9b65da90146983e184"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 00:42:41 crc kubenswrapper[4766]: I1126 00:42:41.481054 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://cef0ca8e4338b5dbd39bc68daed18f6f5044b46711899f9b65da90146983e184" gracePeriod=600 Nov 26 00:42:41 crc kubenswrapper[4766]: I1126 00:42:41.620530 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" event={"ID":"1ecbd1d2-cb7f-432f-bc75-903d0643ea29","Type":"ContainerStarted","Data":"8a2e74efad7d541f04c8dc635f49437996e65e1ca2c77cfd523ba5f59dee0347"} Nov 26 00:42:41 crc kubenswrapper[4766]: I1126 00:42:41.621634 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-glkl6" event={"ID":"40780971-20ed-4b4c-8d19-e74f6582ef8c","Type":"ContainerStarted","Data":"d771a6533701416a700cc1e18bc1cd25b44543865ae2251cd56deca2cba74e69"} Nov 26 00:42:41 crc kubenswrapper[4766]: I1126 00:42:41.623303 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" event={"ID":"08bd3753-3b70-44f8-92ff-98bedd88f662","Type":"ContainerStarted","Data":"c8697279442bcd764b62b25e4990741b03dd670e88bec4559ad22b620126eab6"} Nov 26 00:42:41 crc kubenswrapper[4766]: I1126 00:42:41.626015 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" event={"ID":"751b8e60-46d4-4f5e-b62d-800f641ccd40","Type":"ContainerStarted","Data":"529ad9f55b98547ef2e983ec1a27611aee84050866339d4ad48ffac041ae848e"} Nov 26 00:42:42 crc kubenswrapper[4766]: I1126 00:42:42.636929 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" event={"ID":"3d0d59b2-0486-40a2-ae60-b738a4f76dc4","Type":"ContainerStarted","Data":"282e0f3e513c0848d9385350f540459f60d5cca52c05ea2585f15b371786e987"} Nov 26 00:42:42 crc kubenswrapper[4766]: I1126 00:42:42.639070 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" event={"ID":"8969652f-14c9-4618-ad2d-1b3fd1ad3627","Type":"ContainerStarted","Data":"7ca895f781b91bd56703eeb8cbdcc0e9ef3d10ec65456779ca8985abe2fadd88"} Nov 26 00:42:42 crc kubenswrapper[4766]: I1126 00:42:42.641367 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="cef0ca8e4338b5dbd39bc68daed18f6f5044b46711899f9b65da90146983e184" exitCode=0 Nov 26 00:42:42 crc kubenswrapper[4766]: I1126 00:42:42.641409 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"cef0ca8e4338b5dbd39bc68daed18f6f5044b46711899f9b65da90146983e184"} Nov 26 00:42:42 crc kubenswrapper[4766]: I1126 00:42:42.641459 4766 scope.go:117] "RemoveContainer" containerID="d5bd12ab3efcc84d754246df84a11f62a9d2d5fd80aa034846fa0733f2acc7ee" Nov 26 00:42:42 crc kubenswrapper[4766]: I1126 00:42:42.645860 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" event={"ID":"749a25e6-23eb-4fdd-8e15-ac2874cdee7f","Type":"ContainerStarted","Data":"e8d48edf8701473cece2bdce0d7107d87dd054530b4e46cc47521fd87e92ac5a"} Nov 26 00:42:43 crc kubenswrapper[4766]: I1126 00:42:43.639326 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-cert\") pod \"infra-operator-controller-manager-57548d458d-x6h4v\" (UID: \"c3cec209-fdef-44a6-90d0-2d06cfd7e5cc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:42:43 crc kubenswrapper[4766]: I1126 00:42:43.645712 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3cec209-fdef-44a6-90d0-2d06cfd7e5cc-cert\") pod \"infra-operator-controller-manager-57548d458d-x6h4v\" (UID: \"c3cec209-fdef-44a6-90d0-2d06cfd7e5cc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:42:43 crc kubenswrapper[4766]: E1126 00:42:43.647773 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" podUID="97419f32-8768-495f-a652-8a1e8701546f" Nov 26 00:42:43 crc kubenswrapper[4766]: I1126 00:42:43.655550 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" event={"ID":"97419f32-8768-495f-a652-8a1e8701546f","Type":"ContainerStarted","Data":"adda5f308be0603de2ba6d4ac8a7737e537070861e9864428df8b1e39ea390c0"} Nov 26 00:42:43 crc kubenswrapper[4766]: I1126 00:42:43.662506 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw" event={"ID":"91ee52b5-0e2e-4732-87bf-809e378d33d3","Type":"ContainerStarted","Data":"0a0ef28abd9114b5dfc1b827a78e10c6f12f44b41063b3480c57b485a5daea67"} Nov 26 00:42:43 crc kubenswrapper[4766]: I1126 00:42:43.666044 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"8a2702c35239a61bcf6c33a8721b25d955ee16f1fd71838bddfceaf1a1a62d99"} Nov 26 00:42:43 crc kubenswrapper[4766]: I1126 00:42:43.688977 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" podStartSLOduration=32.688958297 podStartE2EDuration="32.688958297s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:42:40.559972157 +0000 UTC m=+1141.408742597" watchObservedRunningTime="2025-11-26 00:42:43.688958297 +0000 UTC m=+1144.537728727" Nov 26 00:42:43 crc kubenswrapper[4766]: E1126 00:42:43.696329 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" podUID="45353fc4-9955-460b-9276-9792dfdaf8f4" Nov 26 00:42:43 crc kubenswrapper[4766]: I1126 00:42:43.708932 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw" podStartSLOduration=5.793913381 podStartE2EDuration="31.708916057s" podCreationTimestamp="2025-11-26 00:42:12 +0000 UTC" firstStartedPulling="2025-11-26 00:42:14.577736975 +0000 UTC m=+1115.426507405" lastFinishedPulling="2025-11-26 00:42:40.492739651 +0000 UTC m=+1141.341510081" observedRunningTime="2025-11-26 00:42:43.707730448 +0000 UTC m=+1144.556500878" watchObservedRunningTime="2025-11-26 00:42:43.708916057 +0000 UTC m=+1144.557686487" Nov 26 00:42:43 crc kubenswrapper[4766]: I1126 00:42:43.834212 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-pnfcn" Nov 26 00:42:43 crc kubenswrapper[4766]: I1126 00:42:43.839487 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:42:43 crc kubenswrapper[4766]: I1126 00:42:43.842095 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/150d5aa9-549d-4b43-b5c3-e6c81e942146-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t\" (UID: \"150d5aa9-549d-4b43-b5c3-e6c81e942146\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" Nov 26 00:42:43 crc kubenswrapper[4766]: I1126 00:42:43.848060 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/150d5aa9-549d-4b43-b5c3-e6c81e942146-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t\" (UID: \"150d5aa9-549d-4b43-b5c3-e6c81e942146\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" Nov 26 00:42:43 crc kubenswrapper[4766]: E1126 00:42:43.851727 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" podUID="04e53d09-de84-4682-b17b-61f9ea7bf152" Nov 26 00:42:43 crc kubenswrapper[4766]: E1126 00:42:43.931620 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" podUID="2bb23a3b-1f4a-409f-9d9c-0afe73a33352" Nov 26 00:42:43 crc kubenswrapper[4766]: I1126 00:42:43.981281 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-f6npz" Nov 26 00:42:43 crc kubenswrapper[4766]: I1126 00:42:43.990946 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" Nov 26 00:42:44 crc kubenswrapper[4766]: I1126 00:42:44.235664 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v"] Nov 26 00:42:44 crc kubenswrapper[4766]: W1126 00:42:44.272989 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3cec209_fdef_44a6_90d0_2d06cfd7e5cc.slice/crio-8d3dc006a727986090bb635fd29351ba81c1e19c25b6aecdd4cf5f7792d20b0c WatchSource:0}: Error finding container 8d3dc006a727986090bb635fd29351ba81c1e19c25b6aecdd4cf5f7792d20b0c: Status 404 returned error can't find the container with id 8d3dc006a727986090bb635fd29351ba81c1e19c25b6aecdd4cf5f7792d20b0c Nov 26 00:42:44 crc kubenswrapper[4766]: I1126 00:42:44.643344 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t"] Nov 26 00:42:44 crc kubenswrapper[4766]: W1126 00:42:44.667122 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod150d5aa9_549d_4b43_b5c3_e6c81e942146.slice/crio-4076e7eb0e6e85ec559b374a266c7886c1db91131056eae2ee5c17f876a63f92 WatchSource:0}: Error finding container 4076e7eb0e6e85ec559b374a266c7886c1db91131056eae2ee5c17f876a63f92: Status 404 returned error can't find the container with id 4076e7eb0e6e85ec559b374a266c7886c1db91131056eae2ee5c17f876a63f92 Nov 26 00:42:44 crc kubenswrapper[4766]: I1126 00:42:44.692328 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" event={"ID":"2bb23a3b-1f4a-409f-9d9c-0afe73a33352","Type":"ContainerStarted","Data":"00800c5b8ec8525537c354673bdc9b1573bbe462b5a0cb46014d04ab1e4e5c18"} Nov 26 00:42:44 crc kubenswrapper[4766]: I1126 00:42:44.694615 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" event={"ID":"04e53d09-de84-4682-b17b-61f9ea7bf152","Type":"ContainerStarted","Data":"8b02abbc20ae438abe2ae55880bdc2200a53f3d775f905584b3bdfb9113d3d41"} Nov 26 00:42:44 crc kubenswrapper[4766]: I1126 00:42:44.701773 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" event={"ID":"8969652f-14c9-4618-ad2d-1b3fd1ad3627","Type":"ContainerStarted","Data":"240076d57356378bc0de5975e9cc797ae7a311e1945cae17709ea67fa6dff8b1"} Nov 26 00:42:44 crc kubenswrapper[4766]: I1126 00:42:44.702444 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" Nov 26 00:42:44 crc kubenswrapper[4766]: I1126 00:42:44.703664 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" event={"ID":"c3cec209-fdef-44a6-90d0-2d06cfd7e5cc","Type":"ContainerStarted","Data":"8d3dc006a727986090bb635fd29351ba81c1e19c25b6aecdd4cf5f7792d20b0c"} Nov 26 00:42:44 crc kubenswrapper[4766]: E1126 00:42:44.708763 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:e00a9ed0ab26c5b745bd804ab1fe6b22428d026f17ea05a05f045e060342f46c\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" podUID="2bb23a3b-1f4a-409f-9d9c-0afe73a33352" Nov 26 00:42:44 crc kubenswrapper[4766]: E1126 00:42:44.709071 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.248:5001/openstack-k8s-operators/telemetry-operator:bf35154a77d3f7d42763b9d6bf295684481cdc52\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" podUID="04e53d09-de84-4682-b17b-61f9ea7bf152" Nov 26 00:42:44 crc kubenswrapper[4766]: I1126 00:42:44.709429 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" event={"ID":"45353fc4-9955-460b-9276-9792dfdaf8f4","Type":"ContainerStarted","Data":"f694b3181746155e6a668faf90c0904cc4d2ac015ac76fc9c02be99e8317a3f1"} Nov 26 00:42:44 crc kubenswrapper[4766]: I1126 00:42:44.718584 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-glkl6" event={"ID":"40780971-20ed-4b4c-8d19-e74f6582ef8c","Type":"ContainerStarted","Data":"02ad7a3e86eb8022c5353239e9ee6d099b6e2e2c1fbd4e53f7279ff915870f39"} Nov 26 00:42:44 crc kubenswrapper[4766]: I1126 00:42:44.718619 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-glkl6" Nov 26 00:42:44 crc kubenswrapper[4766]: I1126 00:42:44.789981 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" podStartSLOduration=4.282416427 podStartE2EDuration="33.789959098s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="2025-11-26 00:42:14.470227159 +0000 UTC m=+1115.318997589" lastFinishedPulling="2025-11-26 00:42:43.97776983 +0000 UTC m=+1144.826540260" observedRunningTime="2025-11-26 00:42:44.776507321 +0000 UTC m=+1145.625277761" watchObservedRunningTime="2025-11-26 00:42:44.789959098 +0000 UTC m=+1145.638729528" Nov 26 00:42:44 crc kubenswrapper[4766]: I1126 00:42:44.802381 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-glkl6" podStartSLOduration=4.262837606 podStartE2EDuration="33.802356669s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="2025-11-26 00:42:14.427734643 +0000 UTC m=+1115.276505073" lastFinishedPulling="2025-11-26 00:42:43.967253706 +0000 UTC m=+1144.816024136" observedRunningTime="2025-11-26 00:42:44.7976312 +0000 UTC m=+1145.646401630" watchObservedRunningTime="2025-11-26 00:42:44.802356669 +0000 UTC m=+1145.651127099" Nov 26 00:42:44 crc kubenswrapper[4766]: E1126 00:42:44.818412 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" podUID="cc87f631-ba4e-4943-b51d-bf1012cca754" Nov 26 00:42:44 crc kubenswrapper[4766]: E1126 00:42:44.819206 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" podUID="5284ad70-a36c-4666-8202-84633d83f50b" Nov 26 00:42:45 crc kubenswrapper[4766]: E1126 00:42:45.184993 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" podUID="a2e7c098-bf9d-4d99-84bf-adfb49ee6b12" Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.757566 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" event={"ID":"91d3e9c8-8bbc-43b9-947b-71c4bb474b21","Type":"ContainerStarted","Data":"4a985708e1cd5672b893439abd06877da524e1e67743025a497df7d8b4eb927e"} Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.760307 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.766588 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.774932 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" event={"ID":"97419f32-8768-495f-a652-8a1e8701546f","Type":"ContainerStarted","Data":"f4b0af14fec9ed185546e3a6bc7c56dad525b00ceaba8d4801f280f5b12cb82e"} Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.775782 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.776815 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" event={"ID":"cc87f631-ba4e-4943-b51d-bf1012cca754","Type":"ContainerStarted","Data":"21e3578331e2d70d717875269c730fd533b677321286c3625bd80e25c12f11cd"} Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.778269 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" event={"ID":"5284ad70-a36c-4666-8202-84633d83f50b","Type":"ContainerStarted","Data":"b3668272a6280fb20a2fbec3d0110b14f26914d3f5b40948640aae53eb1c40a5"} Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.785941 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" event={"ID":"08bd3753-3b70-44f8-92ff-98bedd88f662","Type":"ContainerStarted","Data":"7817abbbde1a4102eff98f74f05461354a36046c42326180f68c8f346cf803b2"} Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.786923 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.788992 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" event={"ID":"150d5aa9-549d-4b43-b5c3-e6c81e942146","Type":"ContainerStarted","Data":"4076e7eb0e6e85ec559b374a266c7886c1db91131056eae2ee5c17f876a63f92"} Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.789073 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" podStartSLOduration=4.616015394 podStartE2EDuration="34.789061944s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="2025-11-26 00:42:14.357211265 +0000 UTC m=+1115.205981695" lastFinishedPulling="2025-11-26 00:42:44.530257815 +0000 UTC m=+1145.379028245" observedRunningTime="2025-11-26 00:42:45.788197552 +0000 UTC m=+1146.636967982" watchObservedRunningTime="2025-11-26 00:42:45.789061944 +0000 UTC m=+1146.637832374" Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.791569 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" event={"ID":"45353fc4-9955-460b-9276-9792dfdaf8f4","Type":"ContainerStarted","Data":"07f0c7157997d9fbe76f790c8c21c65a71f1d39b9ab836201135a61d80cb901e"} Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.792364 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.792565 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.796606 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" event={"ID":"a2e7c098-bf9d-4d99-84bf-adfb49ee6b12","Type":"ContainerStarted","Data":"c4525da49b407e47118482e101f953fdaf4e1bd6332dc4cc62511f53d1777748"} Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.800920 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" event={"ID":"31b1b73b-1d10-49ac-8549-b202c965b63e","Type":"ContainerStarted","Data":"ca913b778214e7ce2c89de0c36803278dadeb4cbaa170ad6c3915c72550fa958"} Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.811704 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" podStartSLOduration=5.152428816 podStartE2EDuration="34.811665781s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="2025-11-26 00:42:14.427750414 +0000 UTC m=+1115.276520844" lastFinishedPulling="2025-11-26 00:42:44.086987389 +0000 UTC m=+1144.935757809" observedRunningTime="2025-11-26 00:42:45.805191288 +0000 UTC m=+1146.653961728" watchObservedRunningTime="2025-11-26 00:42:45.811665781 +0000 UTC m=+1146.660436211" Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.817021 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn" event={"ID":"a16c5563-750f-4e91-93bd-4044c656cb88","Type":"ContainerStarted","Data":"9ec9f3977e5acabc9cfac1c7a21846d24362af6821f420f4f88179fb213843f8"} Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.818165 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn" Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.853521 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn" Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.857731 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" event={"ID":"3d0d59b2-0486-40a2-ae60-b738a4f76dc4","Type":"ContainerStarted","Data":"01beee0d179d46a54dd3f499027a38f655fbba5e214ecff89082c7afb7b57296"} Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.857766 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.862224 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" podStartSLOduration=5.072661126 podStartE2EDuration="34.862204498s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="2025-11-26 00:42:14.451746746 +0000 UTC m=+1115.300517176" lastFinishedPulling="2025-11-26 00:42:44.241290118 +0000 UTC m=+1145.090060548" observedRunningTime="2025-11-26 00:42:45.857640634 +0000 UTC m=+1146.706411074" watchObservedRunningTime="2025-11-26 00:42:45.862204498 +0000 UTC m=+1146.710974928" Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.874985 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" event={"ID":"1ecbd1d2-cb7f-432f-bc75-903d0643ea29","Type":"ContainerStarted","Data":"5067b3d9c280622d0c82460f5f6e981c806f9f982f406a54a9342efd6c4bab80"} Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.878517 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" event={"ID":"0e6cfef3-86af-4cf0-ab06-b3eef687836b","Type":"ContainerStarted","Data":"dafca054fc4b421b3f373fc52c14e6c0efc66e796dd01a756bcc72dabbd40c6b"} Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.883836 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.887684 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.888692 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" event={"ID":"749a25e6-23eb-4fdd-8e15-ac2874cdee7f","Type":"ContainerStarted","Data":"5b976136ea0527dc414c1ffe7073c473783a5f61165172b8d99c35e38895efa9"} Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.889522 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.892709 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" event={"ID":"00a32e6c-edcc-4625-ac35-26fed19f1638","Type":"ContainerStarted","Data":"00745e6236e47471b85087e4010a77a24e51d6bc4451ff93b092423fcd0020fd"} Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.893706 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.902772 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.920479 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" event={"ID":"751b8e60-46d4-4f5e-b62d-800f641ccd40","Type":"ContainerStarted","Data":"0c74165af8903b09956cc50899df2d2dc8bad574ae2f6cbf9e3fc1a97d767e05"} Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.920530 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.926387 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-glkl6" Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.958336 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" podStartSLOduration=2.75329457 podStartE2EDuration="34.958314128s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="2025-11-26 00:42:12.407674654 +0000 UTC m=+1113.256445084" lastFinishedPulling="2025-11-26 00:42:44.612694212 +0000 UTC m=+1145.461464642" observedRunningTime="2025-11-26 00:42:45.954940174 +0000 UTC m=+1146.803710604" watchObservedRunningTime="2025-11-26 00:42:45.958314128 +0000 UTC m=+1146.807084558" Nov 26 00:42:45 crc kubenswrapper[4766]: I1126 00:42:45.979778 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" podStartSLOduration=3.329642114 podStartE2EDuration="34.979757166s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="2025-11-26 00:42:12.710183571 +0000 UTC m=+1113.558954001" lastFinishedPulling="2025-11-26 00:42:44.360298623 +0000 UTC m=+1145.209069053" observedRunningTime="2025-11-26 00:42:45.971830157 +0000 UTC m=+1146.820600587" watchObservedRunningTime="2025-11-26 00:42:45.979757166 +0000 UTC m=+1146.828527596" Nov 26 00:42:46 crc kubenswrapper[4766]: I1126 00:42:46.068115 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" podStartSLOduration=5.421785739 podStartE2EDuration="35.06808102s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="2025-11-26 00:42:14.440694588 +0000 UTC m=+1115.289465018" lastFinishedPulling="2025-11-26 00:42:44.086989869 +0000 UTC m=+1144.935760299" observedRunningTime="2025-11-26 00:42:46.056391157 +0000 UTC m=+1146.905161587" watchObservedRunningTime="2025-11-26 00:42:46.06808102 +0000 UTC m=+1146.916851470" Nov 26 00:42:46 crc kubenswrapper[4766]: I1126 00:42:46.098708 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn" podStartSLOduration=4.792283204 podStartE2EDuration="35.098683798s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="2025-11-26 00:42:14.448076854 +0000 UTC m=+1115.296847284" lastFinishedPulling="2025-11-26 00:42:44.754477448 +0000 UTC m=+1145.603247878" observedRunningTime="2025-11-26 00:42:46.081806144 +0000 UTC m=+1146.930576574" watchObservedRunningTime="2025-11-26 00:42:46.098683798 +0000 UTC m=+1146.947454228" Nov 26 00:42:46 crc kubenswrapper[4766]: I1126 00:42:46.114245 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" podStartSLOduration=5.042090647 podStartE2EDuration="35.114219217s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="2025-11-26 00:42:14.460912495 +0000 UTC m=+1115.309682925" lastFinishedPulling="2025-11-26 00:42:44.533041065 +0000 UTC m=+1145.381811495" observedRunningTime="2025-11-26 00:42:46.110200757 +0000 UTC m=+1146.958971207" watchObservedRunningTime="2025-11-26 00:42:46.114219217 +0000 UTC m=+1146.962989647" Nov 26 00:42:46 crc kubenswrapper[4766]: I1126 00:42:46.148063 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" podStartSLOduration=5.217428906 podStartE2EDuration="35.148041206s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="2025-11-26 00:42:14.555385085 +0000 UTC m=+1115.404155515" lastFinishedPulling="2025-11-26 00:42:44.485997385 +0000 UTC m=+1145.334767815" observedRunningTime="2025-11-26 00:42:46.138291191 +0000 UTC m=+1146.987061621" watchObservedRunningTime="2025-11-26 00:42:46.148041206 +0000 UTC m=+1146.996811626" Nov 26 00:42:46 crc kubenswrapper[4766]: I1126 00:42:46.174143 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" podStartSLOduration=4.81805848 podStartE2EDuration="35.17412541s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="2025-11-26 00:42:14.46387118 +0000 UTC m=+1115.312641610" lastFinishedPulling="2025-11-26 00:42:44.81993811 +0000 UTC m=+1145.668708540" observedRunningTime="2025-11-26 00:42:46.160053367 +0000 UTC m=+1147.008823797" watchObservedRunningTime="2025-11-26 00:42:46.17412541 +0000 UTC m=+1147.022895840" Nov 26 00:42:46 crc kubenswrapper[4766]: I1126 00:42:46.240472 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" podStartSLOduration=4.782080008 podStartE2EDuration="35.240448883s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="2025-11-26 00:42:13.675964421 +0000 UTC m=+1114.524734851" lastFinishedPulling="2025-11-26 00:42:44.134333296 +0000 UTC m=+1144.983103726" observedRunningTime="2025-11-26 00:42:46.234843282 +0000 UTC m=+1147.083613712" watchObservedRunningTime="2025-11-26 00:42:46.240448883 +0000 UTC m=+1147.089219313" Nov 26 00:42:46 crc kubenswrapper[4766]: I1126 00:42:46.303278 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" podStartSLOduration=3.693430047 podStartE2EDuration="35.303255358s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="2025-11-26 00:42:13.715911123 +0000 UTC m=+1114.564681553" lastFinishedPulling="2025-11-26 00:42:45.325736424 +0000 UTC m=+1146.174506864" observedRunningTime="2025-11-26 00:42:46.296098829 +0000 UTC m=+1147.144869259" watchObservedRunningTime="2025-11-26 00:42:46.303255358 +0000 UTC m=+1147.152025798" Nov 26 00:42:46 crc kubenswrapper[4766]: I1126 00:42:46.928723 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" event={"ID":"04e53d09-de84-4682-b17b-61f9ea7bf152","Type":"ContainerStarted","Data":"e2d340c49871daeed42f4ca0db9efe9f2362548fc9ff8ab87a1d99962b1a93c3"} Nov 26 00:42:46 crc kubenswrapper[4766]: I1126 00:42:46.929266 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" Nov 26 00:42:46 crc kubenswrapper[4766]: I1126 00:42:46.929332 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" Nov 26 00:42:46 crc kubenswrapper[4766]: I1126 00:42:46.931906 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" Nov 26 00:42:46 crc kubenswrapper[4766]: I1126 00:42:46.932595 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" Nov 26 00:42:46 crc kubenswrapper[4766]: I1126 00:42:46.932968 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" Nov 26 00:42:46 crc kubenswrapper[4766]: I1126 00:42:46.951901 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" podStartSLOduration=4.372182878 podStartE2EDuration="35.951885895s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="2025-11-26 00:42:14.440172525 +0000 UTC m=+1115.288942955" lastFinishedPulling="2025-11-26 00:42:46.019875542 +0000 UTC m=+1146.868645972" observedRunningTime="2025-11-26 00:42:46.94571853 +0000 UTC m=+1147.794488960" watchObservedRunningTime="2025-11-26 00:42:46.951885895 +0000 UTC m=+1147.800656325" Nov 26 00:42:48 crc kubenswrapper[4766]: I1126 00:42:48.248800 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:42:50 crc kubenswrapper[4766]: I1126 00:42:50.959971 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" event={"ID":"c3cec209-fdef-44a6-90d0-2d06cfd7e5cc","Type":"ContainerStarted","Data":"6a140667962d79fbd5490523c75046936a2cbbc4d994d23df2283e5282a10a5f"} Nov 26 00:42:50 crc kubenswrapper[4766]: I1126 00:42:50.960207 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" event={"ID":"c3cec209-fdef-44a6-90d0-2d06cfd7e5cc","Type":"ContainerStarted","Data":"6361803ac24912c90abb031d2a560bba65c8b89ddf1b832d483be9a3151edb53"} Nov 26 00:42:50 crc kubenswrapper[4766]: I1126 00:42:50.960463 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:42:50 crc kubenswrapper[4766]: I1126 00:42:50.962343 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" event={"ID":"cc87f631-ba4e-4943-b51d-bf1012cca754","Type":"ContainerStarted","Data":"f7cdc453d3359bcc3627ad6f9177b79e5434324bdb9b07dd2d35a4e00b037322"} Nov 26 00:42:50 crc kubenswrapper[4766]: I1126 00:42:50.962644 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" Nov 26 00:42:50 crc kubenswrapper[4766]: I1126 00:42:50.963855 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" event={"ID":"5284ad70-a36c-4666-8202-84633d83f50b","Type":"ContainerStarted","Data":"adbc6b083454dd91829f6392ff5941d676cb8c4d44bbae613815465810d69a3e"} Nov 26 00:42:50 crc kubenswrapper[4766]: I1126 00:42:50.963967 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" Nov 26 00:42:50 crc kubenswrapper[4766]: I1126 00:42:50.965905 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" event={"ID":"2bb23a3b-1f4a-409f-9d9c-0afe73a33352","Type":"ContainerStarted","Data":"2f60c1d7aa6d4d7875d73101d3c8630a065629b9bbc495ba4eb16c81c1076282"} Nov 26 00:42:50 crc kubenswrapper[4766]: I1126 00:42:50.966444 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" Nov 26 00:42:50 crc kubenswrapper[4766]: I1126 00:42:50.968590 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" event={"ID":"a2e7c098-bf9d-4d99-84bf-adfb49ee6b12","Type":"ContainerStarted","Data":"973fb49329012ed9c8a7426429d28b69e518c0f2507cbfeacc05ff401e64aa19"} Nov 26 00:42:50 crc kubenswrapper[4766]: I1126 00:42:50.969066 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" Nov 26 00:42:50 crc kubenswrapper[4766]: I1126 00:42:50.970686 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" event={"ID":"150d5aa9-549d-4b43-b5c3-e6c81e942146","Type":"ContainerStarted","Data":"3504b845aa2362ac4d2d801677a8679707e409fe90321d8c55f9b1e6c59c00b1"} Nov 26 00:42:50 crc kubenswrapper[4766]: I1126 00:42:50.970709 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" event={"ID":"150d5aa9-549d-4b43-b5c3-e6c81e942146","Type":"ContainerStarted","Data":"8fa8e16236e2c4d7e14a8760e1cba9898e839125d1530d58220f2a9afcd34ba3"} Nov 26 00:42:50 crc kubenswrapper[4766]: I1126 00:42:50.971165 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" Nov 26 00:42:50 crc kubenswrapper[4766]: I1126 00:42:50.986339 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" podStartSLOduration=34.048690787 podStartE2EDuration="39.986322782s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="2025-11-26 00:42:44.314219437 +0000 UTC m=+1145.162989857" lastFinishedPulling="2025-11-26 00:42:50.251851422 +0000 UTC m=+1151.100621852" observedRunningTime="2025-11-26 00:42:50.980748002 +0000 UTC m=+1151.829518432" watchObservedRunningTime="2025-11-26 00:42:50.986322782 +0000 UTC m=+1151.835093212" Nov 26 00:42:51 crc kubenswrapper[4766]: I1126 00:42:51.011755 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" podStartSLOduration=4.628467845 podStartE2EDuration="40.011736359s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="2025-11-26 00:42:13.90004011 +0000 UTC m=+1114.748810540" lastFinishedPulling="2025-11-26 00:42:49.283308624 +0000 UTC m=+1150.132079054" observedRunningTime="2025-11-26 00:42:51.006523869 +0000 UTC m=+1151.855294299" watchObservedRunningTime="2025-11-26 00:42:51.011736359 +0000 UTC m=+1151.860506789" Nov 26 00:42:51 crc kubenswrapper[4766]: I1126 00:42:51.029355 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" podStartSLOduration=4.219688715 podStartE2EDuration="40.029332021s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="2025-11-26 00:42:14.457209903 +0000 UTC m=+1115.305980333" lastFinishedPulling="2025-11-26 00:42:50.266853209 +0000 UTC m=+1151.115623639" observedRunningTime="2025-11-26 00:42:51.024489329 +0000 UTC m=+1151.873259769" watchObservedRunningTime="2025-11-26 00:42:51.029332021 +0000 UTC m=+1151.878102451" Nov 26 00:42:51 crc kubenswrapper[4766]: I1126 00:42:51.060884 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" podStartSLOduration=34.517749709 podStartE2EDuration="40.060867711s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="2025-11-26 00:42:44.70830029 +0000 UTC m=+1145.557070720" lastFinishedPulling="2025-11-26 00:42:50.251418292 +0000 UTC m=+1151.100188722" observedRunningTime="2025-11-26 00:42:51.057109007 +0000 UTC m=+1151.905879457" watchObservedRunningTime="2025-11-26 00:42:51.060867711 +0000 UTC m=+1151.909638131" Nov 26 00:42:51 crc kubenswrapper[4766]: I1126 00:42:51.084366 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" podStartSLOduration=4.439163838 podStartE2EDuration="40.08434941s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="2025-11-26 00:42:13.897578998 +0000 UTC m=+1114.746349428" lastFinishedPulling="2025-11-26 00:42:49.54276457 +0000 UTC m=+1150.391535000" observedRunningTime="2025-11-26 00:42:51.07796773 +0000 UTC m=+1151.926738160" watchObservedRunningTime="2025-11-26 00:42:51.08434941 +0000 UTC m=+1151.933119840" Nov 26 00:42:51 crc kubenswrapper[4766]: I1126 00:42:51.094927 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" podStartSLOduration=5.183873503 podStartE2EDuration="40.094912365s" podCreationTimestamp="2025-11-26 00:42:11 +0000 UTC" firstStartedPulling="2025-11-26 00:42:14.373588315 +0000 UTC m=+1115.222358745" lastFinishedPulling="2025-11-26 00:42:49.284627187 +0000 UTC m=+1150.133397607" observedRunningTime="2025-11-26 00:42:51.093327716 +0000 UTC m=+1151.942098146" watchObservedRunningTime="2025-11-26 00:42:51.094912365 +0000 UTC m=+1151.943682795" Nov 26 00:42:51 crc kubenswrapper[4766]: I1126 00:42:51.616089 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" Nov 26 00:42:52 crc kubenswrapper[4766]: I1126 00:42:52.031744 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" Nov 26 00:42:52 crc kubenswrapper[4766]: I1126 00:42:52.113447 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" Nov 26 00:42:52 crc kubenswrapper[4766]: I1126 00:42:52.169872 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" Nov 26 00:42:52 crc kubenswrapper[4766]: I1126 00:42:52.252242 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" Nov 26 00:42:52 crc kubenswrapper[4766]: I1126 00:42:52.563510 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" Nov 26 00:42:52 crc kubenswrapper[4766]: I1126 00:42:52.565784 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" Nov 26 00:43:01 crc kubenswrapper[4766]: I1126 00:43:01.692472 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" Nov 26 00:43:01 crc kubenswrapper[4766]: I1126 00:43:01.763589 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" Nov 26 00:43:02 crc kubenswrapper[4766]: I1126 00:43:02.149090 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" Nov 26 00:43:02 crc kubenswrapper[4766]: I1126 00:43:02.231828 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" Nov 26 00:43:03 crc kubenswrapper[4766]: I1126 00:43:03.845819 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:43:03 crc kubenswrapper[4766]: I1126 00:43:03.998051 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" Nov 26 00:43:19 crc kubenswrapper[4766]: I1126 00:43:19.907128 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nnfjf"] Nov 26 00:43:19 crc kubenswrapper[4766]: I1126 00:43:19.909986 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-nnfjf" Nov 26 00:43:19 crc kubenswrapper[4766]: I1126 00:43:19.919171 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 26 00:43:19 crc kubenswrapper[4766]: I1126 00:43:19.919575 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-pc94r" Nov 26 00:43:19 crc kubenswrapper[4766]: I1126 00:43:19.919778 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 26 00:43:19 crc kubenswrapper[4766]: I1126 00:43:19.919911 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 26 00:43:19 crc kubenswrapper[4766]: I1126 00:43:19.929995 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nnfjf"] Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.022971 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bsfs9"] Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.025046 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-bsfs9" Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.027500 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.031362 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bsfs9"] Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.065446 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5633e35b-d324-4bf0-b039-d1a3969481bd-config\") pod \"dnsmasq-dns-675f4bcbfc-nnfjf\" (UID: \"5633e35b-d324-4bf0-b039-d1a3969481bd\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nnfjf" Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.065751 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84hqx\" (UniqueName: \"kubernetes.io/projected/5633e35b-d324-4bf0-b039-d1a3969481bd-kube-api-access-84hqx\") pod \"dnsmasq-dns-675f4bcbfc-nnfjf\" (UID: \"5633e35b-d324-4bf0-b039-d1a3969481bd\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nnfjf" Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.167080 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4187edff-cbce-40bc-b499-44eac6eb103e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-bsfs9\" (UID: \"4187edff-cbce-40bc-b499-44eac6eb103e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bsfs9" Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.167236 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fsf7\" (UniqueName: \"kubernetes.io/projected/4187edff-cbce-40bc-b499-44eac6eb103e-kube-api-access-7fsf7\") pod \"dnsmasq-dns-78dd6ddcc-bsfs9\" (UID: \"4187edff-cbce-40bc-b499-44eac6eb103e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bsfs9" Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.167293 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84hqx\" (UniqueName: \"kubernetes.io/projected/5633e35b-d324-4bf0-b039-d1a3969481bd-kube-api-access-84hqx\") pod \"dnsmasq-dns-675f4bcbfc-nnfjf\" (UID: \"5633e35b-d324-4bf0-b039-d1a3969481bd\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nnfjf" Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.167334 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4187edff-cbce-40bc-b499-44eac6eb103e-config\") pod \"dnsmasq-dns-78dd6ddcc-bsfs9\" (UID: \"4187edff-cbce-40bc-b499-44eac6eb103e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bsfs9" Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.167369 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5633e35b-d324-4bf0-b039-d1a3969481bd-config\") pod \"dnsmasq-dns-675f4bcbfc-nnfjf\" (UID: \"5633e35b-d324-4bf0-b039-d1a3969481bd\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nnfjf" Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.168555 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5633e35b-d324-4bf0-b039-d1a3969481bd-config\") pod \"dnsmasq-dns-675f4bcbfc-nnfjf\" (UID: \"5633e35b-d324-4bf0-b039-d1a3969481bd\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nnfjf" Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.190911 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84hqx\" (UniqueName: \"kubernetes.io/projected/5633e35b-d324-4bf0-b039-d1a3969481bd-kube-api-access-84hqx\") pod \"dnsmasq-dns-675f4bcbfc-nnfjf\" (UID: \"5633e35b-d324-4bf0-b039-d1a3969481bd\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nnfjf" Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.230611 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-nnfjf" Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.268542 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fsf7\" (UniqueName: \"kubernetes.io/projected/4187edff-cbce-40bc-b499-44eac6eb103e-kube-api-access-7fsf7\") pod \"dnsmasq-dns-78dd6ddcc-bsfs9\" (UID: \"4187edff-cbce-40bc-b499-44eac6eb103e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bsfs9" Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.268920 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4187edff-cbce-40bc-b499-44eac6eb103e-config\") pod \"dnsmasq-dns-78dd6ddcc-bsfs9\" (UID: \"4187edff-cbce-40bc-b499-44eac6eb103e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bsfs9" Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.269029 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4187edff-cbce-40bc-b499-44eac6eb103e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-bsfs9\" (UID: \"4187edff-cbce-40bc-b499-44eac6eb103e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bsfs9" Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.270325 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4187edff-cbce-40bc-b499-44eac6eb103e-config\") pod \"dnsmasq-dns-78dd6ddcc-bsfs9\" (UID: \"4187edff-cbce-40bc-b499-44eac6eb103e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bsfs9" Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.271093 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4187edff-cbce-40bc-b499-44eac6eb103e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-bsfs9\" (UID: \"4187edff-cbce-40bc-b499-44eac6eb103e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bsfs9" Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.292818 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fsf7\" (UniqueName: \"kubernetes.io/projected/4187edff-cbce-40bc-b499-44eac6eb103e-kube-api-access-7fsf7\") pod \"dnsmasq-dns-78dd6ddcc-bsfs9\" (UID: \"4187edff-cbce-40bc-b499-44eac6eb103e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bsfs9" Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.351493 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-bsfs9" Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.685670 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nnfjf"] Nov 26 00:43:20 crc kubenswrapper[4766]: I1126 00:43:20.798298 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bsfs9"] Nov 26 00:43:20 crc kubenswrapper[4766]: W1126 00:43:20.807762 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4187edff_cbce_40bc_b499_44eac6eb103e.slice/crio-965eb157337206e3b32c4cc1f26746d006c192144ac9ef5f2ffa8243963a3b38 WatchSource:0}: Error finding container 965eb157337206e3b32c4cc1f26746d006c192144ac9ef5f2ffa8243963a3b38: Status 404 returned error can't find the container with id 965eb157337206e3b32c4cc1f26746d006c192144ac9ef5f2ffa8243963a3b38 Nov 26 00:43:21 crc kubenswrapper[4766]: I1126 00:43:21.199073 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-bsfs9" event={"ID":"4187edff-cbce-40bc-b499-44eac6eb103e","Type":"ContainerStarted","Data":"965eb157337206e3b32c4cc1f26746d006c192144ac9ef5f2ffa8243963a3b38"} Nov 26 00:43:21 crc kubenswrapper[4766]: I1126 00:43:21.200034 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-nnfjf" event={"ID":"5633e35b-d324-4bf0-b039-d1a3969481bd","Type":"ContainerStarted","Data":"3f11a5510d7a9c7ef4b0640fa9f37fa8625ad473e561269a96df1393b97acbbe"} Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.115901 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nnfjf"] Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.135146 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jb2rz"] Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.138241 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.147779 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jb2rz"] Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.221886 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cab5b48f-5422-4e6b-a207-c02958c69f79-config\") pod \"dnsmasq-dns-666b6646f7-jb2rz\" (UID: \"cab5b48f-5422-4e6b-a207-c02958c69f79\") " pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.221940 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdq7v\" (UniqueName: \"kubernetes.io/projected/cab5b48f-5422-4e6b-a207-c02958c69f79-kube-api-access-zdq7v\") pod \"dnsmasq-dns-666b6646f7-jb2rz\" (UID: \"cab5b48f-5422-4e6b-a207-c02958c69f79\") " pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.222084 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cab5b48f-5422-4e6b-a207-c02958c69f79-dns-svc\") pod \"dnsmasq-dns-666b6646f7-jb2rz\" (UID: \"cab5b48f-5422-4e6b-a207-c02958c69f79\") " pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.323834 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cab5b48f-5422-4e6b-a207-c02958c69f79-dns-svc\") pod \"dnsmasq-dns-666b6646f7-jb2rz\" (UID: \"cab5b48f-5422-4e6b-a207-c02958c69f79\") " pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.323924 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cab5b48f-5422-4e6b-a207-c02958c69f79-config\") pod \"dnsmasq-dns-666b6646f7-jb2rz\" (UID: \"cab5b48f-5422-4e6b-a207-c02958c69f79\") " pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.323955 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdq7v\" (UniqueName: \"kubernetes.io/projected/cab5b48f-5422-4e6b-a207-c02958c69f79-kube-api-access-zdq7v\") pod \"dnsmasq-dns-666b6646f7-jb2rz\" (UID: \"cab5b48f-5422-4e6b-a207-c02958c69f79\") " pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.325071 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cab5b48f-5422-4e6b-a207-c02958c69f79-config\") pod \"dnsmasq-dns-666b6646f7-jb2rz\" (UID: \"cab5b48f-5422-4e6b-a207-c02958c69f79\") " pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.326082 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cab5b48f-5422-4e6b-a207-c02958c69f79-dns-svc\") pod \"dnsmasq-dns-666b6646f7-jb2rz\" (UID: \"cab5b48f-5422-4e6b-a207-c02958c69f79\") " pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.349412 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdq7v\" (UniqueName: \"kubernetes.io/projected/cab5b48f-5422-4e6b-a207-c02958c69f79-kube-api-access-zdq7v\") pod \"dnsmasq-dns-666b6646f7-jb2rz\" (UID: \"cab5b48f-5422-4e6b-a207-c02958c69f79\") " pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.424939 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bsfs9"] Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.448486 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-chhxg"] Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.449773 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.461109 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-chhxg"] Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.475041 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.527564 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b064e590-4c62-4d62-841e-28a7d134632c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-chhxg\" (UID: \"b064e590-4c62-4d62-841e-28a7d134632c\") " pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.530804 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89mf2\" (UniqueName: \"kubernetes.io/projected/b064e590-4c62-4d62-841e-28a7d134632c-kube-api-access-89mf2\") pod \"dnsmasq-dns-57d769cc4f-chhxg\" (UID: \"b064e590-4c62-4d62-841e-28a7d134632c\") " pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.531046 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b064e590-4c62-4d62-841e-28a7d134632c-config\") pod \"dnsmasq-dns-57d769cc4f-chhxg\" (UID: \"b064e590-4c62-4d62-841e-28a7d134632c\") " pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.632676 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89mf2\" (UniqueName: \"kubernetes.io/projected/b064e590-4c62-4d62-841e-28a7d134632c-kube-api-access-89mf2\") pod \"dnsmasq-dns-57d769cc4f-chhxg\" (UID: \"b064e590-4c62-4d62-841e-28a7d134632c\") " pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.633044 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b064e590-4c62-4d62-841e-28a7d134632c-config\") pod \"dnsmasq-dns-57d769cc4f-chhxg\" (UID: \"b064e590-4c62-4d62-841e-28a7d134632c\") " pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.633105 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b064e590-4c62-4d62-841e-28a7d134632c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-chhxg\" (UID: \"b064e590-4c62-4d62-841e-28a7d134632c\") " pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.633990 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b064e590-4c62-4d62-841e-28a7d134632c-config\") pod \"dnsmasq-dns-57d769cc4f-chhxg\" (UID: \"b064e590-4c62-4d62-841e-28a7d134632c\") " pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.634285 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b064e590-4c62-4d62-841e-28a7d134632c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-chhxg\" (UID: \"b064e590-4c62-4d62-841e-28a7d134632c\") " pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.672429 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89mf2\" (UniqueName: \"kubernetes.io/projected/b064e590-4c62-4d62-841e-28a7d134632c-kube-api-access-89mf2\") pod \"dnsmasq-dns-57d769cc4f-chhxg\" (UID: \"b064e590-4c62-4d62-841e-28a7d134632c\") " pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" Nov 26 00:43:23 crc kubenswrapper[4766]: I1126 00:43:23.767272 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" Nov 26 00:43:24 crc kubenswrapper[4766]: W1126 00:43:24.032005 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcab5b48f_5422_4e6b_a207_c02958c69f79.slice/crio-327ee74afcb829fc22cadcd55f4b211ec2ef5da0b2d99562d6128964a792c7fd WatchSource:0}: Error finding container 327ee74afcb829fc22cadcd55f4b211ec2ef5da0b2d99562d6128964a792c7fd: Status 404 returned error can't find the container with id 327ee74afcb829fc22cadcd55f4b211ec2ef5da0b2d99562d6128964a792c7fd Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.039873 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jb2rz"] Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.240446 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-chhxg"] Nov 26 00:43:24 crc kubenswrapper[4766]: W1126 00:43:24.251625 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb064e590_4c62_4d62_841e_28a7d134632c.slice/crio-81199ac01bc23f7877aa38c040004ebb21d79db326a83af3d16610b4d6f447dd WatchSource:0}: Error finding container 81199ac01bc23f7877aa38c040004ebb21d79db326a83af3d16610b4d6f447dd: Status 404 returned error can't find the container with id 81199ac01bc23f7877aa38c040004ebb21d79db326a83af3d16610b4d6f447dd Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.260125 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" event={"ID":"cab5b48f-5422-4e6b-a207-c02958c69f79","Type":"ContainerStarted","Data":"327ee74afcb829fc22cadcd55f4b211ec2ef5da0b2d99562d6128964a792c7fd"} Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.280982 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.286866 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.293320 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.300963 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.306564 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.306734 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.307544 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.307549 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7g6b2" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.307627 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.307621 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.344539 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dpv8\" (UniqueName: \"kubernetes.io/projected/8a944def-677c-4bdc-a7df-2586d0c7937f-kube-api-access-5dpv8\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.344628 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.344712 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a944def-677c-4bdc-a7df-2586d0c7937f-config-data\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.344841 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.345097 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.345128 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8a944def-677c-4bdc-a7df-2586d0c7937f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.345219 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8a944def-677c-4bdc-a7df-2586d0c7937f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.345274 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8a944def-677c-4bdc-a7df-2586d0c7937f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.346871 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.346981 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.347026 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8a944def-677c-4bdc-a7df-2586d0c7937f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.448855 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dpv8\" (UniqueName: \"kubernetes.io/projected/8a944def-677c-4bdc-a7df-2586d0c7937f-kube-api-access-5dpv8\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.448908 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.448936 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a944def-677c-4bdc-a7df-2586d0c7937f-config-data\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.448965 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.448984 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.449000 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8a944def-677c-4bdc-a7df-2586d0c7937f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.449028 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8a944def-677c-4bdc-a7df-2586d0c7937f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.449048 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8a944def-677c-4bdc-a7df-2586d0c7937f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.449070 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.449108 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.449126 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8a944def-677c-4bdc-a7df-2586d0c7937f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.450782 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.451029 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.451945 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.452143 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8a944def-677c-4bdc-a7df-2586d0c7937f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.452443 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a944def-677c-4bdc-a7df-2586d0c7937f-config-data\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.452756 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8a944def-677c-4bdc-a7df-2586d0c7937f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.456146 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8a944def-677c-4bdc-a7df-2586d0c7937f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.456541 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.457122 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8a944def-677c-4bdc-a7df-2586d0c7937f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.467373 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.467515 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dpv8\" (UniqueName: \"kubernetes.io/projected/8a944def-677c-4bdc-a7df-2586d0c7937f-kube-api-access-5dpv8\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.481992 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.608421 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.610739 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.613734 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.613761 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-llt2n" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.614012 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.615030 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.615231 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.615437 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.615637 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.622166 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.631837 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.651561 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.651623 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.651685 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.651714 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqrq7\" (UniqueName: \"kubernetes.io/projected/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-kube-api-access-dqrq7\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.651748 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.651824 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.651877 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.651941 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.651962 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.652021 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.652048 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.754596 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.755376 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.755463 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.755516 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.755601 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.755639 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.755700 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.755733 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqrq7\" (UniqueName: \"kubernetes.io/projected/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-kube-api-access-dqrq7\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.755768 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.756195 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.756147 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.755937 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.756215 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.756524 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.756580 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.756920 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.758687 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.760476 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.760907 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.764072 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.773848 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqrq7\" (UniqueName: \"kubernetes.io/projected/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-kube-api-access-dqrq7\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.775117 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.795350 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:24 crc kubenswrapper[4766]: I1126 00:43:24.938908 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:43:25 crc kubenswrapper[4766]: I1126 00:43:25.294208 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" event={"ID":"b064e590-4c62-4d62-841e-28a7d134632c","Type":"ContainerStarted","Data":"81199ac01bc23f7877aa38c040004ebb21d79db326a83af3d16610b4d6f447dd"} Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.127943 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.130098 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.132923 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-5zcn8" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.132926 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.133258 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.133735 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.142888 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.163254 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.193586 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/54c97191-0bfc-4a65-8168-10833aab2814-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.193630 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54c97191-0bfc-4a65-8168-10833aab2814-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.193732 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/54c97191-0bfc-4a65-8168-10833aab2814-config-data-default\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.193749 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxr5s\" (UniqueName: \"kubernetes.io/projected/54c97191-0bfc-4a65-8168-10833aab2814-kube-api-access-fxr5s\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.193768 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/54c97191-0bfc-4a65-8168-10833aab2814-kolla-config\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.193826 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.193886 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54c97191-0bfc-4a65-8168-10833aab2814-operator-scripts\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.193905 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/54c97191-0bfc-4a65-8168-10833aab2814-config-data-generated\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.295810 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/54c97191-0bfc-4a65-8168-10833aab2814-config-data-default\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.295865 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxr5s\" (UniqueName: \"kubernetes.io/projected/54c97191-0bfc-4a65-8168-10833aab2814-kube-api-access-fxr5s\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.295917 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/54c97191-0bfc-4a65-8168-10833aab2814-kolla-config\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.295944 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.296030 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54c97191-0bfc-4a65-8168-10833aab2814-operator-scripts\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.296057 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/54c97191-0bfc-4a65-8168-10833aab2814-config-data-generated\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.296104 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/54c97191-0bfc-4a65-8168-10833aab2814-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.296134 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54c97191-0bfc-4a65-8168-10833aab2814-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.296536 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.296901 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/54c97191-0bfc-4a65-8168-10833aab2814-kolla-config\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.297132 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/54c97191-0bfc-4a65-8168-10833aab2814-config-data-default\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.297537 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/54c97191-0bfc-4a65-8168-10833aab2814-config-data-generated\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.301264 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54c97191-0bfc-4a65-8168-10833aab2814-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.301304 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/54c97191-0bfc-4a65-8168-10833aab2814-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.322939 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxr5s\" (UniqueName: \"kubernetes.io/projected/54c97191-0bfc-4a65-8168-10833aab2814-kube-api-access-fxr5s\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.331708 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54c97191-0bfc-4a65-8168-10833aab2814-operator-scripts\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.359075 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"54c97191-0bfc-4a65-8168-10833aab2814\") " pod="openstack/openstack-galera-0" Nov 26 00:43:26 crc kubenswrapper[4766]: I1126 00:43:26.479725 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.457537 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.459812 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.461964 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-qwbfd" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.462314 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.462886 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.463039 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.469287 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.618383 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/44f36c8b-51b8-4bce-afbb-1962eb317dec-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.618440 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44f36c8b-51b8-4bce-afbb-1962eb317dec-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.618927 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvwbb\" (UniqueName: \"kubernetes.io/projected/44f36c8b-51b8-4bce-afbb-1962eb317dec-kube-api-access-nvwbb\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.619020 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.619070 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/44f36c8b-51b8-4bce-afbb-1962eb317dec-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.619130 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/44f36c8b-51b8-4bce-afbb-1962eb317dec-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.619306 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/44f36c8b-51b8-4bce-afbb-1962eb317dec-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.619386 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44f36c8b-51b8-4bce-afbb-1962eb317dec-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.720612 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.720710 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/44f36c8b-51b8-4bce-afbb-1962eb317dec-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.720750 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/44f36c8b-51b8-4bce-afbb-1962eb317dec-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.720815 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/44f36c8b-51b8-4bce-afbb-1962eb317dec-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.720849 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44f36c8b-51b8-4bce-afbb-1962eb317dec-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.720945 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/44f36c8b-51b8-4bce-afbb-1962eb317dec-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.720993 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44f36c8b-51b8-4bce-afbb-1962eb317dec-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.721084 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvwbb\" (UniqueName: \"kubernetes.io/projected/44f36c8b-51b8-4bce-afbb-1962eb317dec-kube-api-access-nvwbb\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.721137 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.721292 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/44f36c8b-51b8-4bce-afbb-1962eb317dec-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.721687 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/44f36c8b-51b8-4bce-afbb-1962eb317dec-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.722256 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44f36c8b-51b8-4bce-afbb-1962eb317dec-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.722543 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/44f36c8b-51b8-4bce-afbb-1962eb317dec-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.725825 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/44f36c8b-51b8-4bce-afbb-1962eb317dec-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.727866 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44f36c8b-51b8-4bce-afbb-1962eb317dec-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.739446 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvwbb\" (UniqueName: \"kubernetes.io/projected/44f36c8b-51b8-4bce-afbb-1962eb317dec-kube-api-access-nvwbb\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.751407 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"44f36c8b-51b8-4bce-afbb-1962eb317dec\") " pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.818408 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.824900 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.825954 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.830515 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.830846 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-6kd8s" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.830848 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.848850 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.923935 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f3679589-d23a-48c7-8b4a-712935d229f1-config-data\") pod \"memcached-0\" (UID: \"f3679589-d23a-48c7-8b4a-712935d229f1\") " pod="openstack/memcached-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.924551 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxggw\" (UniqueName: \"kubernetes.io/projected/f3679589-d23a-48c7-8b4a-712935d229f1-kube-api-access-gxggw\") pod \"memcached-0\" (UID: \"f3679589-d23a-48c7-8b4a-712935d229f1\") " pod="openstack/memcached-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.924613 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f3679589-d23a-48c7-8b4a-712935d229f1-kolla-config\") pod \"memcached-0\" (UID: \"f3679589-d23a-48c7-8b4a-712935d229f1\") " pod="openstack/memcached-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.924678 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3679589-d23a-48c7-8b4a-712935d229f1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f3679589-d23a-48c7-8b4a-712935d229f1\") " pod="openstack/memcached-0" Nov 26 00:43:27 crc kubenswrapper[4766]: I1126 00:43:27.924915 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3679589-d23a-48c7-8b4a-712935d229f1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f3679589-d23a-48c7-8b4a-712935d229f1\") " pod="openstack/memcached-0" Nov 26 00:43:28 crc kubenswrapper[4766]: I1126 00:43:28.025418 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3679589-d23a-48c7-8b4a-712935d229f1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f3679589-d23a-48c7-8b4a-712935d229f1\") " pod="openstack/memcached-0" Nov 26 00:43:28 crc kubenswrapper[4766]: I1126 00:43:28.025498 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f3679589-d23a-48c7-8b4a-712935d229f1-config-data\") pod \"memcached-0\" (UID: \"f3679589-d23a-48c7-8b4a-712935d229f1\") " pod="openstack/memcached-0" Nov 26 00:43:28 crc kubenswrapper[4766]: I1126 00:43:28.025530 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxggw\" (UniqueName: \"kubernetes.io/projected/f3679589-d23a-48c7-8b4a-712935d229f1-kube-api-access-gxggw\") pod \"memcached-0\" (UID: \"f3679589-d23a-48c7-8b4a-712935d229f1\") " pod="openstack/memcached-0" Nov 26 00:43:28 crc kubenswrapper[4766]: I1126 00:43:28.025551 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f3679589-d23a-48c7-8b4a-712935d229f1-kolla-config\") pod \"memcached-0\" (UID: \"f3679589-d23a-48c7-8b4a-712935d229f1\") " pod="openstack/memcached-0" Nov 26 00:43:28 crc kubenswrapper[4766]: I1126 00:43:28.025565 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3679589-d23a-48c7-8b4a-712935d229f1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f3679589-d23a-48c7-8b4a-712935d229f1\") " pod="openstack/memcached-0" Nov 26 00:43:28 crc kubenswrapper[4766]: I1126 00:43:28.026829 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f3679589-d23a-48c7-8b4a-712935d229f1-config-data\") pod \"memcached-0\" (UID: \"f3679589-d23a-48c7-8b4a-712935d229f1\") " pod="openstack/memcached-0" Nov 26 00:43:28 crc kubenswrapper[4766]: I1126 00:43:28.027022 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f3679589-d23a-48c7-8b4a-712935d229f1-kolla-config\") pod \"memcached-0\" (UID: \"f3679589-d23a-48c7-8b4a-712935d229f1\") " pod="openstack/memcached-0" Nov 26 00:43:28 crc kubenswrapper[4766]: I1126 00:43:28.034211 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3679589-d23a-48c7-8b4a-712935d229f1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f3679589-d23a-48c7-8b4a-712935d229f1\") " pod="openstack/memcached-0" Nov 26 00:43:28 crc kubenswrapper[4766]: I1126 00:43:28.034223 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3679589-d23a-48c7-8b4a-712935d229f1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f3679589-d23a-48c7-8b4a-712935d229f1\") " pod="openstack/memcached-0" Nov 26 00:43:28 crc kubenswrapper[4766]: I1126 00:43:28.044350 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxggw\" (UniqueName: \"kubernetes.io/projected/f3679589-d23a-48c7-8b4a-712935d229f1-kube-api-access-gxggw\") pod \"memcached-0\" (UID: \"f3679589-d23a-48c7-8b4a-712935d229f1\") " pod="openstack/memcached-0" Nov 26 00:43:28 crc kubenswrapper[4766]: I1126 00:43:28.149066 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 26 00:43:30 crc kubenswrapper[4766]: I1126 00:43:30.090518 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 26 00:43:30 crc kubenswrapper[4766]: I1126 00:43:30.092068 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 26 00:43:30 crc kubenswrapper[4766]: I1126 00:43:30.095246 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 26 00:43:30 crc kubenswrapper[4766]: I1126 00:43:30.101075 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-58h9w" Nov 26 00:43:30 crc kubenswrapper[4766]: I1126 00:43:30.269574 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sjwx\" (UniqueName: \"kubernetes.io/projected/aa692ff7-2356-4ff4-b9fe-08884e4081cf-kube-api-access-6sjwx\") pod \"kube-state-metrics-0\" (UID: \"aa692ff7-2356-4ff4-b9fe-08884e4081cf\") " pod="openstack/kube-state-metrics-0" Nov 26 00:43:30 crc kubenswrapper[4766]: I1126 00:43:30.372817 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sjwx\" (UniqueName: \"kubernetes.io/projected/aa692ff7-2356-4ff4-b9fe-08884e4081cf-kube-api-access-6sjwx\") pod \"kube-state-metrics-0\" (UID: \"aa692ff7-2356-4ff4-b9fe-08884e4081cf\") " pod="openstack/kube-state-metrics-0" Nov 26 00:43:30 crc kubenswrapper[4766]: I1126 00:43:30.397618 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sjwx\" (UniqueName: \"kubernetes.io/projected/aa692ff7-2356-4ff4-b9fe-08884e4081cf-kube-api-access-6sjwx\") pod \"kube-state-metrics-0\" (UID: \"aa692ff7-2356-4ff4-b9fe-08884e4081cf\") " pod="openstack/kube-state-metrics-0" Nov 26 00:43:30 crc kubenswrapper[4766]: I1126 00:43:30.420114 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 26 00:43:30 crc kubenswrapper[4766]: I1126 00:43:30.654009 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-cmzfk"] Nov 26 00:43:30 crc kubenswrapper[4766]: I1126 00:43:30.660467 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-cmzfk" Nov 26 00:43:30 crc kubenswrapper[4766]: I1126 00:43:30.671309 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards" Nov 26 00:43:30 crc kubenswrapper[4766]: I1126 00:43:30.684560 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/468c5f0c-30d9-4493-a87c-2f4482a3b4d3-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-cmzfk\" (UID: \"468c5f0c-30d9-4493-a87c-2f4482a3b4d3\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-cmzfk" Nov 26 00:43:30 crc kubenswrapper[4766]: I1126 00:43:30.684970 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6rvn\" (UniqueName: \"kubernetes.io/projected/468c5f0c-30d9-4493-a87c-2f4482a3b4d3-kube-api-access-w6rvn\") pod \"observability-ui-dashboards-7d5fb4cbfb-cmzfk\" (UID: \"468c5f0c-30d9-4493-a87c-2f4482a3b4d3\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-cmzfk" Nov 26 00:43:30 crc kubenswrapper[4766]: I1126 00:43:30.687974 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards-sa-dockercfg-lb8kj" Nov 26 00:43:30 crc kubenswrapper[4766]: I1126 00:43:30.698993 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-cmzfk"] Nov 26 00:43:30 crc kubenswrapper[4766]: I1126 00:43:30.797661 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/468c5f0c-30d9-4493-a87c-2f4482a3b4d3-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-cmzfk\" (UID: \"468c5f0c-30d9-4493-a87c-2f4482a3b4d3\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-cmzfk" Nov 26 00:43:30 crc kubenswrapper[4766]: I1126 00:43:30.797778 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6rvn\" (UniqueName: \"kubernetes.io/projected/468c5f0c-30d9-4493-a87c-2f4482a3b4d3-kube-api-access-w6rvn\") pod \"observability-ui-dashboards-7d5fb4cbfb-cmzfk\" (UID: \"468c5f0c-30d9-4493-a87c-2f4482a3b4d3\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-cmzfk" Nov 26 00:43:30 crc kubenswrapper[4766]: E1126 00:43:30.798628 4766 secret.go:188] Couldn't get secret openshift-operators/observability-ui-dashboards: secret "observability-ui-dashboards" not found Nov 26 00:43:30 crc kubenswrapper[4766]: E1126 00:43:30.798687 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/468c5f0c-30d9-4493-a87c-2f4482a3b4d3-serving-cert podName:468c5f0c-30d9-4493-a87c-2f4482a3b4d3 nodeName:}" failed. No retries permitted until 2025-11-26 00:43:31.298671602 +0000 UTC m=+1192.147442032 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/468c5f0c-30d9-4493-a87c-2f4482a3b4d3-serving-cert") pod "observability-ui-dashboards-7d5fb4cbfb-cmzfk" (UID: "468c5f0c-30d9-4493-a87c-2f4482a3b4d3") : secret "observability-ui-dashboards" not found Nov 26 00:43:30 crc kubenswrapper[4766]: I1126 00:43:30.851542 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6rvn\" (UniqueName: \"kubernetes.io/projected/468c5f0c-30d9-4493-a87c-2f4482a3b4d3-kube-api-access-w6rvn\") pod \"observability-ui-dashboards-7d5fb4cbfb-cmzfk\" (UID: \"468c5f0c-30d9-4493-a87c-2f4482a3b4d3\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-cmzfk" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.194343 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-58d4764cd7-9wcr5"] Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.195403 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.253226 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.259897 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.266294 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.266479 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.266594 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-k8x7n" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.266722 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.267519 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.267695 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-58d4764cd7-9wcr5"] Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.276187 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.289221 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.320641 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d39e7a54-3596-47fb-92b2-71b4b8c8109a-service-ca\") pod \"console-58d4764cd7-9wcr5\" (UID: \"d39e7a54-3596-47fb-92b2-71b4b8c8109a\") " pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.320711 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d39e7a54-3596-47fb-92b2-71b4b8c8109a-oauth-serving-cert\") pod \"console-58d4764cd7-9wcr5\" (UID: \"d39e7a54-3596-47fb-92b2-71b4b8c8109a\") " pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.320751 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d39e7a54-3596-47fb-92b2-71b4b8c8109a-trusted-ca-bundle\") pod \"console-58d4764cd7-9wcr5\" (UID: \"d39e7a54-3596-47fb-92b2-71b4b8c8109a\") " pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.320775 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d39e7a54-3596-47fb-92b2-71b4b8c8109a-console-oauth-config\") pod \"console-58d4764cd7-9wcr5\" (UID: \"d39e7a54-3596-47fb-92b2-71b4b8c8109a\") " pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.320793 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d39e7a54-3596-47fb-92b2-71b4b8c8109a-console-config\") pod \"console-58d4764cd7-9wcr5\" (UID: \"d39e7a54-3596-47fb-92b2-71b4b8c8109a\") " pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.321040 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sh6s\" (UniqueName: \"kubernetes.io/projected/d39e7a54-3596-47fb-92b2-71b4b8c8109a-kube-api-access-7sh6s\") pod \"console-58d4764cd7-9wcr5\" (UID: \"d39e7a54-3596-47fb-92b2-71b4b8c8109a\") " pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.321089 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/468c5f0c-30d9-4493-a87c-2f4482a3b4d3-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-cmzfk\" (UID: \"468c5f0c-30d9-4493-a87c-2f4482a3b4d3\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-cmzfk" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.321126 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d39e7a54-3596-47fb-92b2-71b4b8c8109a-console-serving-cert\") pod \"console-58d4764cd7-9wcr5\" (UID: \"d39e7a54-3596-47fb-92b2-71b4b8c8109a\") " pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.326190 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/468c5f0c-30d9-4493-a87c-2f4482a3b4d3-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-cmzfk\" (UID: \"468c5f0c-30d9-4493-a87c-2f4482a3b4d3\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-cmzfk" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.424673 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sh6s\" (UniqueName: \"kubernetes.io/projected/d39e7a54-3596-47fb-92b2-71b4b8c8109a-kube-api-access-7sh6s\") pod \"console-58d4764cd7-9wcr5\" (UID: \"d39e7a54-3596-47fb-92b2-71b4b8c8109a\") " pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.424741 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6e23ce44-c896-41b1-b427-1bc2db9955ca-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.424805 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6e23ce44-c896-41b1-b427-1bc2db9955ca-config\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.424834 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d39e7a54-3596-47fb-92b2-71b4b8c8109a-console-serving-cert\") pod \"console-58d4764cd7-9wcr5\" (UID: \"d39e7a54-3596-47fb-92b2-71b4b8c8109a\") " pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.424853 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6e23ce44-c896-41b1-b427-1bc2db9955ca-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.424899 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d39e7a54-3596-47fb-92b2-71b4b8c8109a-service-ca\") pod \"console-58d4764cd7-9wcr5\" (UID: \"d39e7a54-3596-47fb-92b2-71b4b8c8109a\") " pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.424926 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.424946 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d39e7a54-3596-47fb-92b2-71b4b8c8109a-oauth-serving-cert\") pod \"console-58d4764cd7-9wcr5\" (UID: \"d39e7a54-3596-47fb-92b2-71b4b8c8109a\") " pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.424961 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d39e7a54-3596-47fb-92b2-71b4b8c8109a-trusted-ca-bundle\") pod \"console-58d4764cd7-9wcr5\" (UID: \"d39e7a54-3596-47fb-92b2-71b4b8c8109a\") " pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.424982 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d39e7a54-3596-47fb-92b2-71b4b8c8109a-console-oauth-config\") pod \"console-58d4764cd7-9wcr5\" (UID: \"d39e7a54-3596-47fb-92b2-71b4b8c8109a\") " pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.424997 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d39e7a54-3596-47fb-92b2-71b4b8c8109a-console-config\") pod \"console-58d4764cd7-9wcr5\" (UID: \"d39e7a54-3596-47fb-92b2-71b4b8c8109a\") " pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.425019 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6e23ce44-c896-41b1-b427-1bc2db9955ca-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.425044 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6e23ce44-c896-41b1-b427-1bc2db9955ca-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.425074 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6e23ce44-c896-41b1-b427-1bc2db9955ca-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.425329 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7hjr\" (UniqueName: \"kubernetes.io/projected/6e23ce44-c896-41b1-b427-1bc2db9955ca-kube-api-access-g7hjr\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.427026 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d39e7a54-3596-47fb-92b2-71b4b8c8109a-service-ca\") pod \"console-58d4764cd7-9wcr5\" (UID: \"d39e7a54-3596-47fb-92b2-71b4b8c8109a\") " pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.427042 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d39e7a54-3596-47fb-92b2-71b4b8c8109a-trusted-ca-bundle\") pod \"console-58d4764cd7-9wcr5\" (UID: \"d39e7a54-3596-47fb-92b2-71b4b8c8109a\") " pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.427478 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d39e7a54-3596-47fb-92b2-71b4b8c8109a-oauth-serving-cert\") pod \"console-58d4764cd7-9wcr5\" (UID: \"d39e7a54-3596-47fb-92b2-71b4b8c8109a\") " pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.427795 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d39e7a54-3596-47fb-92b2-71b4b8c8109a-console-config\") pod \"console-58d4764cd7-9wcr5\" (UID: \"d39e7a54-3596-47fb-92b2-71b4b8c8109a\") " pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.430303 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d39e7a54-3596-47fb-92b2-71b4b8c8109a-console-serving-cert\") pod \"console-58d4764cd7-9wcr5\" (UID: \"d39e7a54-3596-47fb-92b2-71b4b8c8109a\") " pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.434466 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d39e7a54-3596-47fb-92b2-71b4b8c8109a-console-oauth-config\") pod \"console-58d4764cd7-9wcr5\" (UID: \"d39e7a54-3596-47fb-92b2-71b4b8c8109a\") " pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.443507 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sh6s\" (UniqueName: \"kubernetes.io/projected/d39e7a54-3596-47fb-92b2-71b4b8c8109a-kube-api-access-7sh6s\") pod \"console-58d4764cd7-9wcr5\" (UID: \"d39e7a54-3596-47fb-92b2-71b4b8c8109a\") " pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.519298 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.526555 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6e23ce44-c896-41b1-b427-1bc2db9955ca-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.526600 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6e23ce44-c896-41b1-b427-1bc2db9955ca-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.526623 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6e23ce44-c896-41b1-b427-1bc2db9955ca-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.526676 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7hjr\" (UniqueName: \"kubernetes.io/projected/6e23ce44-c896-41b1-b427-1bc2db9955ca-kube-api-access-g7hjr\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.526713 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6e23ce44-c896-41b1-b427-1bc2db9955ca-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.526754 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6e23ce44-c896-41b1-b427-1bc2db9955ca-config\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.526780 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6e23ce44-c896-41b1-b427-1bc2db9955ca-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.526831 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.528567 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6e23ce44-c896-41b1-b427-1bc2db9955ca-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.533147 4766 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.533184 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/149ea7c544c1e9392b1f012edc83d0a3042f9a6ea5986794ddffaab68b75c51a/globalmount\"" pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.534302 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6e23ce44-c896-41b1-b427-1bc2db9955ca-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.534849 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6e23ce44-c896-41b1-b427-1bc2db9955ca-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.536271 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6e23ce44-c896-41b1-b427-1bc2db9955ca-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.540960 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6e23ce44-c896-41b1-b427-1bc2db9955ca-config\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.542177 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6e23ce44-c896-41b1-b427-1bc2db9955ca-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.550453 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7hjr\" (UniqueName: \"kubernetes.io/projected/6e23ce44-c896-41b1-b427-1bc2db9955ca-kube-api-access-g7hjr\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.568356 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf\") pod \"prometheus-metric-storage-0\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.608419 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-cmzfk" Nov 26 00:43:31 crc kubenswrapper[4766]: I1126 00:43:31.683550 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.419014 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-p42tr"] Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.420488 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.429310 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.430567 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-lxqm8" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.430935 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.449779 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-p42tr"] Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.513308 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-6ghtc"] Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.515382 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.538861 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-6ghtc"] Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.558574 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/22168318-cef8-4363-8b3e-3042e82037ea-var-run-ovn\") pod \"ovn-controller-p42tr\" (UID: \"22168318-cef8-4363-8b3e-3042e82037ea\") " pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.558672 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/22168318-cef8-4363-8b3e-3042e82037ea-var-run\") pod \"ovn-controller-p42tr\" (UID: \"22168318-cef8-4363-8b3e-3042e82037ea\") " pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.558736 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/22168318-cef8-4363-8b3e-3042e82037ea-var-log-ovn\") pod \"ovn-controller-p42tr\" (UID: \"22168318-cef8-4363-8b3e-3042e82037ea\") " pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.558780 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gh67\" (UniqueName: \"kubernetes.io/projected/22168318-cef8-4363-8b3e-3042e82037ea-kube-api-access-9gh67\") pod \"ovn-controller-p42tr\" (UID: \"22168318-cef8-4363-8b3e-3042e82037ea\") " pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.558832 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/22168318-cef8-4363-8b3e-3042e82037ea-ovn-controller-tls-certs\") pod \"ovn-controller-p42tr\" (UID: \"22168318-cef8-4363-8b3e-3042e82037ea\") " pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.558865 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/22168318-cef8-4363-8b3e-3042e82037ea-scripts\") pod \"ovn-controller-p42tr\" (UID: \"22168318-cef8-4363-8b3e-3042e82037ea\") " pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.558921 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22168318-cef8-4363-8b3e-3042e82037ea-combined-ca-bundle\") pod \"ovn-controller-p42tr\" (UID: \"22168318-cef8-4363-8b3e-3042e82037ea\") " pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.660035 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71963f95-01c9-473a-be8a-7798f6038ffb-scripts\") pod \"ovn-controller-ovs-6ghtc\" (UID: \"71963f95-01c9-473a-be8a-7798f6038ffb\") " pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.660078 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/71963f95-01c9-473a-be8a-7798f6038ffb-var-run\") pod \"ovn-controller-ovs-6ghtc\" (UID: \"71963f95-01c9-473a-be8a-7798f6038ffb\") " pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.660114 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/22168318-cef8-4363-8b3e-3042e82037ea-var-run\") pod \"ovn-controller-p42tr\" (UID: \"22168318-cef8-4363-8b3e-3042e82037ea\") " pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.660275 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/22168318-cef8-4363-8b3e-3042e82037ea-var-log-ovn\") pod \"ovn-controller-p42tr\" (UID: \"22168318-cef8-4363-8b3e-3042e82037ea\") " pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.660363 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gh67\" (UniqueName: \"kubernetes.io/projected/22168318-cef8-4363-8b3e-3042e82037ea-kube-api-access-9gh67\") pod \"ovn-controller-p42tr\" (UID: \"22168318-cef8-4363-8b3e-3042e82037ea\") " pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.660401 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcc67\" (UniqueName: \"kubernetes.io/projected/71963f95-01c9-473a-be8a-7798f6038ffb-kube-api-access-jcc67\") pod \"ovn-controller-ovs-6ghtc\" (UID: \"71963f95-01c9-473a-be8a-7798f6038ffb\") " pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.660440 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/71963f95-01c9-473a-be8a-7798f6038ffb-var-lib\") pod \"ovn-controller-ovs-6ghtc\" (UID: \"71963f95-01c9-473a-be8a-7798f6038ffb\") " pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.660475 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/22168318-cef8-4363-8b3e-3042e82037ea-ovn-controller-tls-certs\") pod \"ovn-controller-p42tr\" (UID: \"22168318-cef8-4363-8b3e-3042e82037ea\") " pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.660541 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/22168318-cef8-4363-8b3e-3042e82037ea-scripts\") pod \"ovn-controller-p42tr\" (UID: \"22168318-cef8-4363-8b3e-3042e82037ea\") " pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.660573 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/71963f95-01c9-473a-be8a-7798f6038ffb-etc-ovs\") pod \"ovn-controller-ovs-6ghtc\" (UID: \"71963f95-01c9-473a-be8a-7798f6038ffb\") " pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.660671 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/71963f95-01c9-473a-be8a-7798f6038ffb-var-log\") pod \"ovn-controller-ovs-6ghtc\" (UID: \"71963f95-01c9-473a-be8a-7798f6038ffb\") " pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.660722 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/22168318-cef8-4363-8b3e-3042e82037ea-var-log-ovn\") pod \"ovn-controller-p42tr\" (UID: \"22168318-cef8-4363-8b3e-3042e82037ea\") " pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.660734 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22168318-cef8-4363-8b3e-3042e82037ea-combined-ca-bundle\") pod \"ovn-controller-p42tr\" (UID: \"22168318-cef8-4363-8b3e-3042e82037ea\") " pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.660824 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/22168318-cef8-4363-8b3e-3042e82037ea-var-run-ovn\") pod \"ovn-controller-p42tr\" (UID: \"22168318-cef8-4363-8b3e-3042e82037ea\") " pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.661181 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/22168318-cef8-4363-8b3e-3042e82037ea-var-run-ovn\") pod \"ovn-controller-p42tr\" (UID: \"22168318-cef8-4363-8b3e-3042e82037ea\") " pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.661311 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/22168318-cef8-4363-8b3e-3042e82037ea-var-run\") pod \"ovn-controller-p42tr\" (UID: \"22168318-cef8-4363-8b3e-3042e82037ea\") " pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.662299 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/22168318-cef8-4363-8b3e-3042e82037ea-scripts\") pod \"ovn-controller-p42tr\" (UID: \"22168318-cef8-4363-8b3e-3042e82037ea\") " pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.669416 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22168318-cef8-4363-8b3e-3042e82037ea-combined-ca-bundle\") pod \"ovn-controller-p42tr\" (UID: \"22168318-cef8-4363-8b3e-3042e82037ea\") " pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.669574 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/22168318-cef8-4363-8b3e-3042e82037ea-ovn-controller-tls-certs\") pod \"ovn-controller-p42tr\" (UID: \"22168318-cef8-4363-8b3e-3042e82037ea\") " pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.685605 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gh67\" (UniqueName: \"kubernetes.io/projected/22168318-cef8-4363-8b3e-3042e82037ea-kube-api-access-9gh67\") pod \"ovn-controller-p42tr\" (UID: \"22168318-cef8-4363-8b3e-3042e82037ea\") " pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.740878 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p42tr" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.762307 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcc67\" (UniqueName: \"kubernetes.io/projected/71963f95-01c9-473a-be8a-7798f6038ffb-kube-api-access-jcc67\") pod \"ovn-controller-ovs-6ghtc\" (UID: \"71963f95-01c9-473a-be8a-7798f6038ffb\") " pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.762359 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/71963f95-01c9-473a-be8a-7798f6038ffb-var-lib\") pod \"ovn-controller-ovs-6ghtc\" (UID: \"71963f95-01c9-473a-be8a-7798f6038ffb\") " pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.762406 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/71963f95-01c9-473a-be8a-7798f6038ffb-etc-ovs\") pod \"ovn-controller-ovs-6ghtc\" (UID: \"71963f95-01c9-473a-be8a-7798f6038ffb\") " pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.762446 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/71963f95-01c9-473a-be8a-7798f6038ffb-var-log\") pod \"ovn-controller-ovs-6ghtc\" (UID: \"71963f95-01c9-473a-be8a-7798f6038ffb\") " pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.762485 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71963f95-01c9-473a-be8a-7798f6038ffb-scripts\") pod \"ovn-controller-ovs-6ghtc\" (UID: \"71963f95-01c9-473a-be8a-7798f6038ffb\") " pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.762502 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/71963f95-01c9-473a-be8a-7798f6038ffb-var-run\") pod \"ovn-controller-ovs-6ghtc\" (UID: \"71963f95-01c9-473a-be8a-7798f6038ffb\") " pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.762759 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/71963f95-01c9-473a-be8a-7798f6038ffb-var-run\") pod \"ovn-controller-ovs-6ghtc\" (UID: \"71963f95-01c9-473a-be8a-7798f6038ffb\") " pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.762759 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/71963f95-01c9-473a-be8a-7798f6038ffb-var-log\") pod \"ovn-controller-ovs-6ghtc\" (UID: \"71963f95-01c9-473a-be8a-7798f6038ffb\") " pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.762779 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/71963f95-01c9-473a-be8a-7798f6038ffb-etc-ovs\") pod \"ovn-controller-ovs-6ghtc\" (UID: \"71963f95-01c9-473a-be8a-7798f6038ffb\") " pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.762751 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/71963f95-01c9-473a-be8a-7798f6038ffb-var-lib\") pod \"ovn-controller-ovs-6ghtc\" (UID: \"71963f95-01c9-473a-be8a-7798f6038ffb\") " pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.764420 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71963f95-01c9-473a-be8a-7798f6038ffb-scripts\") pod \"ovn-controller-ovs-6ghtc\" (UID: \"71963f95-01c9-473a-be8a-7798f6038ffb\") " pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.779867 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcc67\" (UniqueName: \"kubernetes.io/projected/71963f95-01c9-473a-be8a-7798f6038ffb-kube-api-access-jcc67\") pod \"ovn-controller-ovs-6ghtc\" (UID: \"71963f95-01c9-473a-be8a-7798f6038ffb\") " pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:43:33 crc kubenswrapper[4766]: I1126 00:43:33.836972 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.154060 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.156588 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.158318 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.158522 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-vjfll" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.158911 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.159084 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.161853 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.165977 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.272940 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d346209c-ec54-4719-9b56-7a079c603ebc-config\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.272998 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d346209c-ec54-4719-9b56-7a079c603ebc-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.273207 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d346209c-ec54-4719-9b56-7a079c603ebc-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.273320 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d346209c-ec54-4719-9b56-7a079c603ebc-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.273567 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98f8p\" (UniqueName: \"kubernetes.io/projected/d346209c-ec54-4719-9b56-7a079c603ebc-kube-api-access-98f8p\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.273604 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.273636 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d346209c-ec54-4719-9b56-7a079c603ebc-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.273699 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d346209c-ec54-4719-9b56-7a079c603ebc-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.375704 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d346209c-ec54-4719-9b56-7a079c603ebc-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.375772 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98f8p\" (UniqueName: \"kubernetes.io/projected/d346209c-ec54-4719-9b56-7a079c603ebc-kube-api-access-98f8p\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.375796 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.375815 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d346209c-ec54-4719-9b56-7a079c603ebc-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.375851 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d346209c-ec54-4719-9b56-7a079c603ebc-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.375895 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d346209c-ec54-4719-9b56-7a079c603ebc-config\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.375922 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d346209c-ec54-4719-9b56-7a079c603ebc-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.375968 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d346209c-ec54-4719-9b56-7a079c603ebc-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.376739 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.377177 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d346209c-ec54-4719-9b56-7a079c603ebc-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.377890 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d346209c-ec54-4719-9b56-7a079c603ebc-config\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.382321 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d346209c-ec54-4719-9b56-7a079c603ebc-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.382387 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d346209c-ec54-4719-9b56-7a079c603ebc-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.386051 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d346209c-ec54-4719-9b56-7a079c603ebc-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.388066 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d346209c-ec54-4719-9b56-7a079c603ebc-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.396148 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98f8p\" (UniqueName: \"kubernetes.io/projected/d346209c-ec54-4719-9b56-7a079c603ebc-kube-api-access-98f8p\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.403786 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d346209c-ec54-4719-9b56-7a079c603ebc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.430053 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 26 00:43:34 crc kubenswrapper[4766]: I1126 00:43:34.485641 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 26 00:43:36 crc kubenswrapper[4766]: I1126 00:43:36.883293 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 26 00:43:36 crc kubenswrapper[4766]: I1126 00:43:36.885121 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:36 crc kubenswrapper[4766]: I1126 00:43:36.889905 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-rjlnb" Nov 26 00:43:36 crc kubenswrapper[4766]: I1126 00:43:36.890122 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 26 00:43:36 crc kubenswrapper[4766]: I1126 00:43:36.890464 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 26 00:43:36 crc kubenswrapper[4766]: I1126 00:43:36.890508 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 26 00:43:36 crc kubenswrapper[4766]: I1126 00:43:36.899573 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 26 00:43:36 crc kubenswrapper[4766]: I1126 00:43:36.945891 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:36 crc kubenswrapper[4766]: I1126 00:43:36.945978 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/954395e4-178d-46ef-a695-8895ebe444d8-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:36 crc kubenswrapper[4766]: I1126 00:43:36.946011 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/954395e4-178d-46ef-a695-8895ebe444d8-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:36 crc kubenswrapper[4766]: I1126 00:43:36.946088 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/954395e4-178d-46ef-a695-8895ebe444d8-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:36 crc kubenswrapper[4766]: I1126 00:43:36.946110 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/954395e4-178d-46ef-a695-8895ebe444d8-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:36 crc kubenswrapper[4766]: I1126 00:43:36.946202 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/954395e4-178d-46ef-a695-8895ebe444d8-config\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:36 crc kubenswrapper[4766]: I1126 00:43:36.946284 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/954395e4-178d-46ef-a695-8895ebe444d8-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:36 crc kubenswrapper[4766]: I1126 00:43:36.946361 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmcxm\" (UniqueName: \"kubernetes.io/projected/954395e4-178d-46ef-a695-8895ebe444d8-kube-api-access-tmcxm\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:37 crc kubenswrapper[4766]: I1126 00:43:37.047733 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/954395e4-178d-46ef-a695-8895ebe444d8-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:37 crc kubenswrapper[4766]: I1126 00:43:37.047843 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmcxm\" (UniqueName: \"kubernetes.io/projected/954395e4-178d-46ef-a695-8895ebe444d8-kube-api-access-tmcxm\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:37 crc kubenswrapper[4766]: I1126 00:43:37.048223 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:37 crc kubenswrapper[4766]: I1126 00:43:37.048441 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:37 crc kubenswrapper[4766]: I1126 00:43:37.052156 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/954395e4-178d-46ef-a695-8895ebe444d8-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:37 crc kubenswrapper[4766]: I1126 00:43:37.060823 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/954395e4-178d-46ef-a695-8895ebe444d8-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:37 crc kubenswrapper[4766]: I1126 00:43:37.053272 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/954395e4-178d-46ef-a695-8895ebe444d8-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:37 crc kubenswrapper[4766]: I1126 00:43:37.060923 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/954395e4-178d-46ef-a695-8895ebe444d8-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:37 crc kubenswrapper[4766]: I1126 00:43:37.061363 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/954395e4-178d-46ef-a695-8895ebe444d8-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:37 crc kubenswrapper[4766]: I1126 00:43:37.061484 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/954395e4-178d-46ef-a695-8895ebe444d8-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:37 crc kubenswrapper[4766]: I1126 00:43:37.062154 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/954395e4-178d-46ef-a695-8895ebe444d8-config\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:37 crc kubenswrapper[4766]: I1126 00:43:37.063078 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/954395e4-178d-46ef-a695-8895ebe444d8-config\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:37 crc kubenswrapper[4766]: I1126 00:43:37.061387 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/954395e4-178d-46ef-a695-8895ebe444d8-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:37 crc kubenswrapper[4766]: I1126 00:43:37.065124 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmcxm\" (UniqueName: \"kubernetes.io/projected/954395e4-178d-46ef-a695-8895ebe444d8-kube-api-access-tmcxm\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:37 crc kubenswrapper[4766]: I1126 00:43:37.065590 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/954395e4-178d-46ef-a695-8895ebe444d8-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:37 crc kubenswrapper[4766]: I1126 00:43:37.068021 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/954395e4-178d-46ef-a695-8895ebe444d8-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:37 crc kubenswrapper[4766]: I1126 00:43:37.077740 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"954395e4-178d-46ef-a695-8895ebe444d8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:37 crc kubenswrapper[4766]: I1126 00:43:37.216969 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 26 00:43:39 crc kubenswrapper[4766]: W1126 00:43:39.883107 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0527e2fe_678e_4712_ac14_3f2e1f3f0fe0.slice/crio-40456bb04133e9b6fce53ed7c23091720a5c5fed300ddcb1d404ce30ba23c0af WatchSource:0}: Error finding container 40456bb04133e9b6fce53ed7c23091720a5c5fed300ddcb1d404ce30ba23c0af: Status 404 returned error can't find the container with id 40456bb04133e9b6fce53ed7c23091720a5c5fed300ddcb1d404ce30ba23c0af Nov 26 00:43:40 crc kubenswrapper[4766]: I1126 00:43:40.281727 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 26 00:43:40 crc kubenswrapper[4766]: I1126 00:43:40.534559 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0","Type":"ContainerStarted","Data":"40456bb04133e9b6fce53ed7c23091720a5c5fed300ddcb1d404ce30ba23c0af"} Nov 26 00:43:40 crc kubenswrapper[4766]: E1126 00:43:40.933503 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 26 00:43:40 crc kubenswrapper[4766]: E1126 00:43:40.933894 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7fsf7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-bsfs9_openstack(4187edff-cbce-40bc-b499-44eac6eb103e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 00:43:40 crc kubenswrapper[4766]: E1126 00:43:40.935043 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-bsfs9" podUID="4187edff-cbce-40bc-b499-44eac6eb103e" Nov 26 00:43:40 crc kubenswrapper[4766]: E1126 00:43:40.946370 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 26 00:43:40 crc kubenswrapper[4766]: E1126 00:43:40.946506 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-84hqx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-nnfjf_openstack(5633e35b-d324-4bf0-b039-d1a3969481bd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 00:43:40 crc kubenswrapper[4766]: E1126 00:43:40.947744 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-nnfjf" podUID="5633e35b-d324-4bf0-b039-d1a3969481bd" Nov 26 00:43:41 crc kubenswrapper[4766]: I1126 00:43:41.567535 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f3679589-d23a-48c7-8b4a-712935d229f1","Type":"ContainerStarted","Data":"5042c22d308d0df56c7e2c2b6c1b8d82fccc3081028f803f655180218fd72fa6"} Nov 26 00:43:41 crc kubenswrapper[4766]: I1126 00:43:41.569490 4766 generic.go:334] "Generic (PLEG): container finished" podID="cab5b48f-5422-4e6b-a207-c02958c69f79" containerID="29288c68d9d6afbc7a775b8862ae8d254065212b91c3585c0d5da7c9ce578886" exitCode=0 Nov 26 00:43:41 crc kubenswrapper[4766]: I1126 00:43:41.569542 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" event={"ID":"cab5b48f-5422-4e6b-a207-c02958c69f79","Type":"ContainerDied","Data":"29288c68d9d6afbc7a775b8862ae8d254065212b91c3585c0d5da7c9ce578886"} Nov 26 00:43:41 crc kubenswrapper[4766]: I1126 00:43:41.573447 4766 generic.go:334] "Generic (PLEG): container finished" podID="b064e590-4c62-4d62-841e-28a7d134632c" containerID="be727cd0c6c7e68fa18dc05725614418c2605ff5f44a3831b5630d5171c27263" exitCode=0 Nov 26 00:43:41 crc kubenswrapper[4766]: I1126 00:43:41.573776 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" event={"ID":"b064e590-4c62-4d62-841e-28a7d134632c","Type":"ContainerDied","Data":"be727cd0c6c7e68fa18dc05725614418c2605ff5f44a3831b5630d5171c27263"} Nov 26 00:43:41 crc kubenswrapper[4766]: I1126 00:43:41.619271 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 26 00:43:41 crc kubenswrapper[4766]: I1126 00:43:41.626887 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 26 00:43:41 crc kubenswrapper[4766]: E1126 00:43:41.776962 4766 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Nov 26 00:43:41 crc kubenswrapper[4766]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/cab5b48f-5422-4e6b-a207-c02958c69f79/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 26 00:43:41 crc kubenswrapper[4766]: > podSandboxID="327ee74afcb829fc22cadcd55f4b211ec2ef5da0b2d99562d6128964a792c7fd" Nov 26 00:43:41 crc kubenswrapper[4766]: E1126 00:43:41.777143 4766 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 26 00:43:41 crc kubenswrapper[4766]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zdq7v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-jb2rz_openstack(cab5b48f-5422-4e6b-a207-c02958c69f79): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/cab5b48f-5422-4e6b-a207-c02958c69f79/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 26 00:43:41 crc kubenswrapper[4766]: > logger="UnhandledError" Nov 26 00:43:41 crc kubenswrapper[4766]: E1126 00:43:41.778358 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/cab5b48f-5422-4e6b-a207-c02958c69f79/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" podUID="cab5b48f-5422-4e6b-a207-c02958c69f79" Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.311311 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-cmzfk"] Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.327813 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.336059 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.345129 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 26 00:43:42 crc kubenswrapper[4766]: W1126 00:43:42.348182 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod468c5f0c_30d9_4493_a87c_2f4482a3b4d3.slice/crio-bda2104d25a8d06b3f805af184a368f865f2121de2a003ca3b071bbaef436c9a WatchSource:0}: Error finding container bda2104d25a8d06b3f805af184a368f865f2121de2a003ca3b071bbaef436c9a: Status 404 returned error can't find the container with id bda2104d25a8d06b3f805af184a368f865f2121de2a003ca3b071bbaef436c9a Nov 26 00:43:42 crc kubenswrapper[4766]: W1126 00:43:42.349455 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod44f36c8b_51b8_4bce_afbb_1962eb317dec.slice/crio-cb3078a786e5c352db6c7ffdf839f7408b2a6d52e234d587a312eb4cf8b9d3fc WatchSource:0}: Error finding container cb3078a786e5c352db6c7ffdf839f7408b2a6d52e234d587a312eb4cf8b9d3fc: Status 404 returned error can't find the container with id cb3078a786e5c352db6c7ffdf839f7408b2a6d52e234d587a312eb4cf8b9d3fc Nov 26 00:43:42 crc kubenswrapper[4766]: W1126 00:43:42.359610 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e23ce44_c896_41b1_b427_1bc2db9955ca.slice/crio-3fe19dd68ce379294bd9201b944d402e76e55120cf5f90ad8c7531eef13219c4 WatchSource:0}: Error finding container 3fe19dd68ce379294bd9201b944d402e76e55120cf5f90ad8c7531eef13219c4: Status 404 returned error can't find the container with id 3fe19dd68ce379294bd9201b944d402e76e55120cf5f90ad8c7531eef13219c4 Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.418718 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-bsfs9" Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.423399 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-nnfjf" Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.465980 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4187edff-cbce-40bc-b499-44eac6eb103e-config\") pod \"4187edff-cbce-40bc-b499-44eac6eb103e\" (UID: \"4187edff-cbce-40bc-b499-44eac6eb103e\") " Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.466092 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84hqx\" (UniqueName: \"kubernetes.io/projected/5633e35b-d324-4bf0-b039-d1a3969481bd-kube-api-access-84hqx\") pod \"5633e35b-d324-4bf0-b039-d1a3969481bd\" (UID: \"5633e35b-d324-4bf0-b039-d1a3969481bd\") " Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.466212 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fsf7\" (UniqueName: \"kubernetes.io/projected/4187edff-cbce-40bc-b499-44eac6eb103e-kube-api-access-7fsf7\") pod \"4187edff-cbce-40bc-b499-44eac6eb103e\" (UID: \"4187edff-cbce-40bc-b499-44eac6eb103e\") " Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.466236 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5633e35b-d324-4bf0-b039-d1a3969481bd-config\") pod \"5633e35b-d324-4bf0-b039-d1a3969481bd\" (UID: \"5633e35b-d324-4bf0-b039-d1a3969481bd\") " Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.466265 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4187edff-cbce-40bc-b499-44eac6eb103e-dns-svc\") pod \"4187edff-cbce-40bc-b499-44eac6eb103e\" (UID: \"4187edff-cbce-40bc-b499-44eac6eb103e\") " Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.467370 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4187edff-cbce-40bc-b499-44eac6eb103e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4187edff-cbce-40bc-b499-44eac6eb103e" (UID: "4187edff-cbce-40bc-b499-44eac6eb103e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.467791 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4187edff-cbce-40bc-b499-44eac6eb103e-config" (OuterVolumeSpecName: "config") pod "4187edff-cbce-40bc-b499-44eac6eb103e" (UID: "4187edff-cbce-40bc-b499-44eac6eb103e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.471207 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5633e35b-d324-4bf0-b039-d1a3969481bd-config" (OuterVolumeSpecName: "config") pod "5633e35b-d324-4bf0-b039-d1a3969481bd" (UID: "5633e35b-d324-4bf0-b039-d1a3969481bd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.473843 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5633e35b-d324-4bf0-b039-d1a3969481bd-kube-api-access-84hqx" (OuterVolumeSpecName: "kube-api-access-84hqx") pod "5633e35b-d324-4bf0-b039-d1a3969481bd" (UID: "5633e35b-d324-4bf0-b039-d1a3969481bd"). InnerVolumeSpecName "kube-api-access-84hqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.477223 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-58d4764cd7-9wcr5"] Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.477391 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4187edff-cbce-40bc-b499-44eac6eb103e-kube-api-access-7fsf7" (OuterVolumeSpecName: "kube-api-access-7fsf7") pod "4187edff-cbce-40bc-b499-44eac6eb103e" (UID: "4187edff-cbce-40bc-b499-44eac6eb103e"). InnerVolumeSpecName "kube-api-access-7fsf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:43:42 crc kubenswrapper[4766]: W1126 00:43:42.493627 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd39e7a54_3596_47fb_92b2_71b4b8c8109a.slice/crio-9321e3fd494c5485e1aed436db276ac70b9896d4bd8c0d98921fb0353fd1167c WatchSource:0}: Error finding container 9321e3fd494c5485e1aed436db276ac70b9896d4bd8c0d98921fb0353fd1167c: Status 404 returned error can't find the container with id 9321e3fd494c5485e1aed436db276ac70b9896d4bd8c0d98921fb0353fd1167c Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.514784 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-p42tr"] Nov 26 00:43:42 crc kubenswrapper[4766]: W1126 00:43:42.551191 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22168318_cef8_4363_8b3e_3042e82037ea.slice/crio-a736c75d916412b0667b20dfbc4c6532e271ef23e2b0327d9c4a9c7ac051f45f WatchSource:0}: Error finding container a736c75d916412b0667b20dfbc4c6532e271ef23e2b0327d9c4a9c7ac051f45f: Status 404 returned error can't find the container with id a736c75d916412b0667b20dfbc4c6532e271ef23e2b0327d9c4a9c7ac051f45f Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.568618 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fsf7\" (UniqueName: \"kubernetes.io/projected/4187edff-cbce-40bc-b499-44eac6eb103e-kube-api-access-7fsf7\") on node \"crc\" DevicePath \"\"" Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.568675 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5633e35b-d324-4bf0-b039-d1a3969481bd-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.568688 4766 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4187edff-cbce-40bc-b499-44eac6eb103e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.568700 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4187edff-cbce-40bc-b499-44eac6eb103e-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.568711 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84hqx\" (UniqueName: \"kubernetes.io/projected/5633e35b-d324-4bf0-b039-d1a3969481bd-kube-api-access-84hqx\") on node \"crc\" DevicePath \"\"" Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.582927 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-cmzfk" event={"ID":"468c5f0c-30d9-4493-a87c-2f4482a3b4d3","Type":"ContainerStarted","Data":"bda2104d25a8d06b3f805af184a368f865f2121de2a003ca3b071bbaef436c9a"} Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.584684 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"aa692ff7-2356-4ff4-b9fe-08884e4081cf","Type":"ContainerStarted","Data":"4d43c11d7971dbe73139e7fa058242d5076d9ac157ea01e58c08708156ac63f7"} Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.587685 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8a944def-677c-4bdc-a7df-2586d0c7937f","Type":"ContainerStarted","Data":"ea70652be2b194e5c24fcf97542f3dab2c21f84e848aa50dd6610284c7b5f461"} Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.591072 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-58d4764cd7-9wcr5" event={"ID":"d39e7a54-3596-47fb-92b2-71b4b8c8109a","Type":"ContainerStarted","Data":"9321e3fd494c5485e1aed436db276ac70b9896d4bd8c0d98921fb0353fd1167c"} Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.592548 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-bsfs9" event={"ID":"4187edff-cbce-40bc-b499-44eac6eb103e","Type":"ContainerDied","Data":"965eb157337206e3b32c4cc1f26746d006c192144ac9ef5f2ffa8243963a3b38"} Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.592714 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-bsfs9" Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.604363 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"44f36c8b-51b8-4bce-afbb-1962eb317dec","Type":"ContainerStarted","Data":"cb3078a786e5c352db6c7ffdf839f7408b2a6d52e234d587a312eb4cf8b9d3fc"} Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.606331 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p42tr" event={"ID":"22168318-cef8-4363-8b3e-3042e82037ea","Type":"ContainerStarted","Data":"a736c75d916412b0667b20dfbc4c6532e271ef23e2b0327d9c4a9c7ac051f45f"} Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.609140 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" event={"ID":"b064e590-4c62-4d62-841e-28a7d134632c","Type":"ContainerStarted","Data":"02162a110dc761842427e2838e3c4b9002ed51c5213c5d3293a64dcf21c5fb46"} Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.609280 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.611371 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-nnfjf" Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.611408 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-nnfjf" event={"ID":"5633e35b-d324-4bf0-b039-d1a3969481bd","Type":"ContainerDied","Data":"3f11a5510d7a9c7ef4b0640fa9f37fa8625ad473e561269a96df1393b97acbbe"} Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.616214 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"54c97191-0bfc-4a65-8168-10833aab2814","Type":"ContainerStarted","Data":"8a9fad29fd5736ee98d3c3ac5c5410f7f813ccf14ddc71fa1fc619273050f9cc"} Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.618293 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6e23ce44-c896-41b1-b427-1bc2db9955ca","Type":"ContainerStarted","Data":"3fe19dd68ce379294bd9201b944d402e76e55120cf5f90ad8c7531eef13219c4"} Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.662737 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" podStartSLOduration=2.828296304 podStartE2EDuration="19.662702373s" podCreationTimestamp="2025-11-26 00:43:23 +0000 UTC" firstStartedPulling="2025-11-26 00:43:24.255039078 +0000 UTC m=+1185.103809508" lastFinishedPulling="2025-11-26 00:43:41.089445157 +0000 UTC m=+1201.938215577" observedRunningTime="2025-11-26 00:43:42.640708691 +0000 UTC m=+1203.489479121" watchObservedRunningTime="2025-11-26 00:43:42.662702373 +0000 UTC m=+1203.511472793" Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.714472 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nnfjf"] Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.735867 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nnfjf"] Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.761344 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bsfs9"] Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.779871 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bsfs9"] Nov 26 00:43:42 crc kubenswrapper[4766]: I1126 00:43:42.981594 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 26 00:43:43 crc kubenswrapper[4766]: I1126 00:43:43.627538 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-58d4764cd7-9wcr5" event={"ID":"d39e7a54-3596-47fb-92b2-71b4b8c8109a","Type":"ContainerStarted","Data":"5b07db0914295b6bd1a74be1aad3faf8e30d59962d5d155621b0ce3e754263b0"} Nov 26 00:43:43 crc kubenswrapper[4766]: I1126 00:43:43.636086 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" event={"ID":"cab5b48f-5422-4e6b-a207-c02958c69f79","Type":"ContainerStarted","Data":"05f3cf7c20b88700a8f777fd941907ae207221807e9cb63e26028b08fd5af782"} Nov 26 00:43:43 crc kubenswrapper[4766]: I1126 00:43:43.636341 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" Nov 26 00:43:43 crc kubenswrapper[4766]: I1126 00:43:43.645570 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-58d4764cd7-9wcr5" podStartSLOduration=12.64554901 podStartE2EDuration="12.64554901s" podCreationTimestamp="2025-11-26 00:43:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:43:43.644894554 +0000 UTC m=+1204.493664984" watchObservedRunningTime="2025-11-26 00:43:43.64554901 +0000 UTC m=+1204.494319440" Nov 26 00:43:43 crc kubenswrapper[4766]: I1126 00:43:43.668662 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" podStartSLOduration=3.639482056 podStartE2EDuration="20.668629599s" podCreationTimestamp="2025-11-26 00:43:23 +0000 UTC" firstStartedPulling="2025-11-26 00:43:24.034043996 +0000 UTC m=+1184.882814436" lastFinishedPulling="2025-11-26 00:43:41.063191549 +0000 UTC m=+1201.911961979" observedRunningTime="2025-11-26 00:43:43.665606473 +0000 UTC m=+1204.514376903" watchObservedRunningTime="2025-11-26 00:43:43.668629599 +0000 UTC m=+1204.517400029" Nov 26 00:43:43 crc kubenswrapper[4766]: I1126 00:43:43.858145 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4187edff-cbce-40bc-b499-44eac6eb103e" path="/var/lib/kubelet/pods/4187edff-cbce-40bc-b499-44eac6eb103e/volumes" Nov 26 00:43:43 crc kubenswrapper[4766]: I1126 00:43:43.858593 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5633e35b-d324-4bf0-b039-d1a3969481bd" path="/var/lib/kubelet/pods/5633e35b-d324-4bf0-b039-d1a3969481bd/volumes" Nov 26 00:43:43 crc kubenswrapper[4766]: I1126 00:43:43.862486 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 26 00:43:43 crc kubenswrapper[4766]: I1126 00:43:43.982380 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-6ghtc"] Nov 26 00:43:45 crc kubenswrapper[4766]: W1126 00:43:45.048393 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod954395e4_178d_46ef_a695_8895ebe444d8.slice/crio-3f9e7b9e73788111aad9c7c0fa35119f5b0e28dc4b1451bf86a9c680c80e5e72 WatchSource:0}: Error finding container 3f9e7b9e73788111aad9c7c0fa35119f5b0e28dc4b1451bf86a9c680c80e5e72: Status 404 returned error can't find the container with id 3f9e7b9e73788111aad9c7c0fa35119f5b0e28dc4b1451bf86a9c680c80e5e72 Nov 26 00:43:45 crc kubenswrapper[4766]: I1126 00:43:45.653479 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6ghtc" event={"ID":"71963f95-01c9-473a-be8a-7798f6038ffb","Type":"ContainerStarted","Data":"55fea22d4df35bd6c98e0b0d9ead8ab43976f173d40d4f7507852047c5b03130"} Nov 26 00:43:45 crc kubenswrapper[4766]: I1126 00:43:45.655802 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"954395e4-178d-46ef-a695-8895ebe444d8","Type":"ContainerStarted","Data":"3f9e7b9e73788111aad9c7c0fa35119f5b0e28dc4b1451bf86a9c680c80e5e72"} Nov 26 00:43:47 crc kubenswrapper[4766]: I1126 00:43:47.081990 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 00:43:47 crc kubenswrapper[4766]: I1126 00:43:47.674014 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d346209c-ec54-4719-9b56-7a079c603ebc","Type":"ContainerStarted","Data":"6c36a9460919ebc0fe71abcc860929a7efe161f5d72d52276eea801bb64e7002"} Nov 26 00:43:48 crc kubenswrapper[4766]: I1126 00:43:48.477798 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" Nov 26 00:43:48 crc kubenswrapper[4766]: I1126 00:43:48.769803 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" Nov 26 00:43:48 crc kubenswrapper[4766]: I1126 00:43:48.820316 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jb2rz"] Nov 26 00:43:48 crc kubenswrapper[4766]: I1126 00:43:48.820545 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" podUID="cab5b48f-5422-4e6b-a207-c02958c69f79" containerName="dnsmasq-dns" containerID="cri-o://05f3cf7c20b88700a8f777fd941907ae207221807e9cb63e26028b08fd5af782" gracePeriod=10 Nov 26 00:43:49 crc kubenswrapper[4766]: I1126 00:43:49.702037 4766 generic.go:334] "Generic (PLEG): container finished" podID="cab5b48f-5422-4e6b-a207-c02958c69f79" containerID="05f3cf7c20b88700a8f777fd941907ae207221807e9cb63e26028b08fd5af782" exitCode=0 Nov 26 00:43:49 crc kubenswrapper[4766]: I1126 00:43:49.702092 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" event={"ID":"cab5b48f-5422-4e6b-a207-c02958c69f79","Type":"ContainerDied","Data":"05f3cf7c20b88700a8f777fd941907ae207221807e9cb63e26028b08fd5af782"} Nov 26 00:43:51 crc kubenswrapper[4766]: I1126 00:43:51.520933 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:51 crc kubenswrapper[4766]: I1126 00:43:51.521257 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:51 crc kubenswrapper[4766]: I1126 00:43:51.525994 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:51 crc kubenswrapper[4766]: I1126 00:43:51.721559 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-58d4764cd7-9wcr5" Nov 26 00:43:51 crc kubenswrapper[4766]: I1126 00:43:51.773195 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-5bc7576dcd-sqbpf"] Nov 26 00:43:54 crc kubenswrapper[4766]: I1126 00:43:54.318518 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" Nov 26 00:43:54 crc kubenswrapper[4766]: I1126 00:43:54.431309 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cab5b48f-5422-4e6b-a207-c02958c69f79-dns-svc\") pod \"cab5b48f-5422-4e6b-a207-c02958c69f79\" (UID: \"cab5b48f-5422-4e6b-a207-c02958c69f79\") " Nov 26 00:43:54 crc kubenswrapper[4766]: I1126 00:43:54.431381 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cab5b48f-5422-4e6b-a207-c02958c69f79-config\") pod \"cab5b48f-5422-4e6b-a207-c02958c69f79\" (UID: \"cab5b48f-5422-4e6b-a207-c02958c69f79\") " Nov 26 00:43:54 crc kubenswrapper[4766]: I1126 00:43:54.431507 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdq7v\" (UniqueName: \"kubernetes.io/projected/cab5b48f-5422-4e6b-a207-c02958c69f79-kube-api-access-zdq7v\") pod \"cab5b48f-5422-4e6b-a207-c02958c69f79\" (UID: \"cab5b48f-5422-4e6b-a207-c02958c69f79\") " Nov 26 00:43:54 crc kubenswrapper[4766]: I1126 00:43:54.453043 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cab5b48f-5422-4e6b-a207-c02958c69f79-kube-api-access-zdq7v" (OuterVolumeSpecName: "kube-api-access-zdq7v") pod "cab5b48f-5422-4e6b-a207-c02958c69f79" (UID: "cab5b48f-5422-4e6b-a207-c02958c69f79"). InnerVolumeSpecName "kube-api-access-zdq7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:43:54 crc kubenswrapper[4766]: I1126 00:43:54.475966 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cab5b48f-5422-4e6b-a207-c02958c69f79-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cab5b48f-5422-4e6b-a207-c02958c69f79" (UID: "cab5b48f-5422-4e6b-a207-c02958c69f79"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:43:54 crc kubenswrapper[4766]: I1126 00:43:54.479615 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cab5b48f-5422-4e6b-a207-c02958c69f79-config" (OuterVolumeSpecName: "config") pod "cab5b48f-5422-4e6b-a207-c02958c69f79" (UID: "cab5b48f-5422-4e6b-a207-c02958c69f79"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:43:54 crc kubenswrapper[4766]: I1126 00:43:54.536940 4766 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cab5b48f-5422-4e6b-a207-c02958c69f79-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 00:43:54 crc kubenswrapper[4766]: I1126 00:43:54.536974 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cab5b48f-5422-4e6b-a207-c02958c69f79-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:43:54 crc kubenswrapper[4766]: I1126 00:43:54.536985 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdq7v\" (UniqueName: \"kubernetes.io/projected/cab5b48f-5422-4e6b-a207-c02958c69f79-kube-api-access-zdq7v\") on node \"crc\" DevicePath \"\"" Nov 26 00:43:54 crc kubenswrapper[4766]: I1126 00:43:54.742367 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" event={"ID":"cab5b48f-5422-4e6b-a207-c02958c69f79","Type":"ContainerDied","Data":"327ee74afcb829fc22cadcd55f4b211ec2ef5da0b2d99562d6128964a792c7fd"} Nov 26 00:43:54 crc kubenswrapper[4766]: I1126 00:43:54.742441 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" Nov 26 00:43:54 crc kubenswrapper[4766]: I1126 00:43:54.742461 4766 scope.go:117] "RemoveContainer" containerID="05f3cf7c20b88700a8f777fd941907ae207221807e9cb63e26028b08fd5af782" Nov 26 00:43:54 crc kubenswrapper[4766]: I1126 00:43:54.778234 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jb2rz"] Nov 26 00:43:54 crc kubenswrapper[4766]: I1126 00:43:54.785310 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jb2rz"] Nov 26 00:43:55 crc kubenswrapper[4766]: I1126 00:43:55.247097 4766 scope.go:117] "RemoveContainer" containerID="29288c68d9d6afbc7a775b8862ae8d254065212b91c3585c0d5da7c9ce578886" Nov 26 00:43:55 crc kubenswrapper[4766]: I1126 00:43:55.838079 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cab5b48f-5422-4e6b-a207-c02958c69f79" path="/var/lib/kubelet/pods/cab5b48f-5422-4e6b-a207-c02958c69f79/volumes" Nov 26 00:43:56 crc kubenswrapper[4766]: I1126 00:43:56.792566 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f3679589-d23a-48c7-8b4a-712935d229f1","Type":"ContainerStarted","Data":"b90a67ddf3567b56532a5537ebdf75eb86cd6960ce20e551a3422624065da892"} Nov 26 00:43:56 crc kubenswrapper[4766]: I1126 00:43:56.793009 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 26 00:43:56 crc kubenswrapper[4766]: I1126 00:43:56.799709 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"44f36c8b-51b8-4bce-afbb-1962eb317dec","Type":"ContainerStarted","Data":"923681756e1854ea0c12a9b2793aa6e5bba85d62da3fde1ee778ab1ae93829cf"} Nov 26 00:43:56 crc kubenswrapper[4766]: I1126 00:43:56.804593 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"54c97191-0bfc-4a65-8168-10833aab2814","Type":"ContainerStarted","Data":"fd95c4ef8cd316d6ba0ccd2a83dbff58b228eb06197012de4ff863af48e90124"} Nov 26 00:43:56 crc kubenswrapper[4766]: I1126 00:43:56.806742 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8a944def-677c-4bdc-a7df-2586d0c7937f","Type":"ContainerStarted","Data":"ce25f81214edc08620b9866c07d4fdf518d951e03f1f623a3eeb20ca690ef230"} Nov 26 00:43:56 crc kubenswrapper[4766]: I1126 00:43:56.818020 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=22.854671905 podStartE2EDuration="29.818002564s" podCreationTimestamp="2025-11-26 00:43:27 +0000 UTC" firstStartedPulling="2025-11-26 00:43:40.868528737 +0000 UTC m=+1201.717299177" lastFinishedPulling="2025-11-26 00:43:47.831859406 +0000 UTC m=+1208.680629836" observedRunningTime="2025-11-26 00:43:56.815271046 +0000 UTC m=+1217.664041466" watchObservedRunningTime="2025-11-26 00:43:56.818002564 +0000 UTC m=+1217.666772994" Nov 26 00:43:56 crc kubenswrapper[4766]: I1126 00:43:56.871574 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-qnhgk"] Nov 26 00:43:56 crc kubenswrapper[4766]: E1126 00:43:56.873943 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cab5b48f-5422-4e6b-a207-c02958c69f79" containerName="init" Nov 26 00:43:56 crc kubenswrapper[4766]: I1126 00:43:56.877670 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="cab5b48f-5422-4e6b-a207-c02958c69f79" containerName="init" Nov 26 00:43:56 crc kubenswrapper[4766]: E1126 00:43:56.877723 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cab5b48f-5422-4e6b-a207-c02958c69f79" containerName="dnsmasq-dns" Nov 26 00:43:56 crc kubenswrapper[4766]: I1126 00:43:56.877730 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="cab5b48f-5422-4e6b-a207-c02958c69f79" containerName="dnsmasq-dns" Nov 26 00:43:56 crc kubenswrapper[4766]: I1126 00:43:56.881049 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="cab5b48f-5422-4e6b-a207-c02958c69f79" containerName="dnsmasq-dns" Nov 26 00:43:56 crc kubenswrapper[4766]: I1126 00:43:56.883551 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-qnhgk" Nov 26 00:43:56 crc kubenswrapper[4766]: I1126 00:43:56.886914 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 26 00:43:56 crc kubenswrapper[4766]: I1126 00:43:56.934490 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-qnhgk"] Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.028245 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/342b8104-66eb-4fe1-88ad-ef5afc617d65-ovs-rundir\") pod \"ovn-controller-metrics-qnhgk\" (UID: \"342b8104-66eb-4fe1-88ad-ef5afc617d65\") " pod="openstack/ovn-controller-metrics-qnhgk" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.028336 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/342b8104-66eb-4fe1-88ad-ef5afc617d65-config\") pod \"ovn-controller-metrics-qnhgk\" (UID: \"342b8104-66eb-4fe1-88ad-ef5afc617d65\") " pod="openstack/ovn-controller-metrics-qnhgk" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.028369 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/342b8104-66eb-4fe1-88ad-ef5afc617d65-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-qnhgk\" (UID: \"342b8104-66eb-4fe1-88ad-ef5afc617d65\") " pod="openstack/ovn-controller-metrics-qnhgk" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.028413 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/342b8104-66eb-4fe1-88ad-ef5afc617d65-combined-ca-bundle\") pod \"ovn-controller-metrics-qnhgk\" (UID: \"342b8104-66eb-4fe1-88ad-ef5afc617d65\") " pod="openstack/ovn-controller-metrics-qnhgk" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.028447 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh6f5\" (UniqueName: \"kubernetes.io/projected/342b8104-66eb-4fe1-88ad-ef5afc617d65-kube-api-access-rh6f5\") pod \"ovn-controller-metrics-qnhgk\" (UID: \"342b8104-66eb-4fe1-88ad-ef5afc617d65\") " pod="openstack/ovn-controller-metrics-qnhgk" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.028616 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/342b8104-66eb-4fe1-88ad-ef5afc617d65-ovn-rundir\") pod \"ovn-controller-metrics-qnhgk\" (UID: \"342b8104-66eb-4fe1-88ad-ef5afc617d65\") " pod="openstack/ovn-controller-metrics-qnhgk" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.031007 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-xkcs4"] Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.050908 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-xkcs4" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.055807 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.057633 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-xkcs4"] Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.130087 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/342b8104-66eb-4fe1-88ad-ef5afc617d65-ovn-rundir\") pod \"ovn-controller-metrics-qnhgk\" (UID: \"342b8104-66eb-4fe1-88ad-ef5afc617d65\") " pod="openstack/ovn-controller-metrics-qnhgk" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.130458 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/342b8104-66eb-4fe1-88ad-ef5afc617d65-ovs-rundir\") pod \"ovn-controller-metrics-qnhgk\" (UID: \"342b8104-66eb-4fe1-88ad-ef5afc617d65\") " pod="openstack/ovn-controller-metrics-qnhgk" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.130962 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/342b8104-66eb-4fe1-88ad-ef5afc617d65-ovs-rundir\") pod \"ovn-controller-metrics-qnhgk\" (UID: \"342b8104-66eb-4fe1-88ad-ef5afc617d65\") " pod="openstack/ovn-controller-metrics-qnhgk" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.131102 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/342b8104-66eb-4fe1-88ad-ef5afc617d65-config\") pod \"ovn-controller-metrics-qnhgk\" (UID: \"342b8104-66eb-4fe1-88ad-ef5afc617d65\") " pod="openstack/ovn-controller-metrics-qnhgk" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.131243 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/342b8104-66eb-4fe1-88ad-ef5afc617d65-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-qnhgk\" (UID: \"342b8104-66eb-4fe1-88ad-ef5afc617d65\") " pod="openstack/ovn-controller-metrics-qnhgk" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.131364 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/683ec07f-9451-46ca-b41f-fe332bf2284c-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-xkcs4\" (UID: \"683ec07f-9451-46ca-b41f-fe332bf2284c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xkcs4" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.131479 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlqwd\" (UniqueName: \"kubernetes.io/projected/683ec07f-9451-46ca-b41f-fe332bf2284c-kube-api-access-hlqwd\") pod \"dnsmasq-dns-5bf47b49b7-xkcs4\" (UID: \"683ec07f-9451-46ca-b41f-fe332bf2284c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xkcs4" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.131617 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/342b8104-66eb-4fe1-88ad-ef5afc617d65-combined-ca-bundle\") pod \"ovn-controller-metrics-qnhgk\" (UID: \"342b8104-66eb-4fe1-88ad-ef5afc617d65\") " pod="openstack/ovn-controller-metrics-qnhgk" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.131736 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/683ec07f-9451-46ca-b41f-fe332bf2284c-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-xkcs4\" (UID: \"683ec07f-9451-46ca-b41f-fe332bf2284c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xkcs4" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.131862 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh6f5\" (UniqueName: \"kubernetes.io/projected/342b8104-66eb-4fe1-88ad-ef5afc617d65-kube-api-access-rh6f5\") pod \"ovn-controller-metrics-qnhgk\" (UID: \"342b8104-66eb-4fe1-88ad-ef5afc617d65\") " pod="openstack/ovn-controller-metrics-qnhgk" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.131982 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/683ec07f-9451-46ca-b41f-fe332bf2284c-config\") pod \"dnsmasq-dns-5bf47b49b7-xkcs4\" (UID: \"683ec07f-9451-46ca-b41f-fe332bf2284c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xkcs4" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.132801 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/342b8104-66eb-4fe1-88ad-ef5afc617d65-config\") pod \"ovn-controller-metrics-qnhgk\" (UID: \"342b8104-66eb-4fe1-88ad-ef5afc617d65\") " pod="openstack/ovn-controller-metrics-qnhgk" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.132899 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/342b8104-66eb-4fe1-88ad-ef5afc617d65-ovn-rundir\") pod \"ovn-controller-metrics-qnhgk\" (UID: \"342b8104-66eb-4fe1-88ad-ef5afc617d65\") " pod="openstack/ovn-controller-metrics-qnhgk" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.139426 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/342b8104-66eb-4fe1-88ad-ef5afc617d65-combined-ca-bundle\") pod \"ovn-controller-metrics-qnhgk\" (UID: \"342b8104-66eb-4fe1-88ad-ef5afc617d65\") " pod="openstack/ovn-controller-metrics-qnhgk" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.149735 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/342b8104-66eb-4fe1-88ad-ef5afc617d65-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-qnhgk\" (UID: \"342b8104-66eb-4fe1-88ad-ef5afc617d65\") " pod="openstack/ovn-controller-metrics-qnhgk" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.168870 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh6f5\" (UniqueName: \"kubernetes.io/projected/342b8104-66eb-4fe1-88ad-ef5afc617d65-kube-api-access-rh6f5\") pod \"ovn-controller-metrics-qnhgk\" (UID: \"342b8104-66eb-4fe1-88ad-ef5afc617d65\") " pod="openstack/ovn-controller-metrics-qnhgk" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.203331 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-xkcs4"] Nov 26 00:43:57 crc kubenswrapper[4766]: E1126 00:43:57.204614 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-hlqwd ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-5bf47b49b7-xkcs4" podUID="683ec07f-9451-46ca-b41f-fe332bf2284c" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.205137 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-qnhgk" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.233173 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/683ec07f-9451-46ca-b41f-fe332bf2284c-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-xkcs4\" (UID: \"683ec07f-9451-46ca-b41f-fe332bf2284c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xkcs4" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.233249 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlqwd\" (UniqueName: \"kubernetes.io/projected/683ec07f-9451-46ca-b41f-fe332bf2284c-kube-api-access-hlqwd\") pod \"dnsmasq-dns-5bf47b49b7-xkcs4\" (UID: \"683ec07f-9451-46ca-b41f-fe332bf2284c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xkcs4" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.233304 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/683ec07f-9451-46ca-b41f-fe332bf2284c-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-xkcs4\" (UID: \"683ec07f-9451-46ca-b41f-fe332bf2284c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xkcs4" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.233342 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/683ec07f-9451-46ca-b41f-fe332bf2284c-config\") pod \"dnsmasq-dns-5bf47b49b7-xkcs4\" (UID: \"683ec07f-9451-46ca-b41f-fe332bf2284c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xkcs4" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.234465 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/683ec07f-9451-46ca-b41f-fe332bf2284c-config\") pod \"dnsmasq-dns-5bf47b49b7-xkcs4\" (UID: \"683ec07f-9451-46ca-b41f-fe332bf2284c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xkcs4" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.235116 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/683ec07f-9451-46ca-b41f-fe332bf2284c-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-xkcs4\" (UID: \"683ec07f-9451-46ca-b41f-fe332bf2284c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xkcs4" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.236128 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/683ec07f-9451-46ca-b41f-fe332bf2284c-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-xkcs4\" (UID: \"683ec07f-9451-46ca-b41f-fe332bf2284c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xkcs4" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.239454 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-7mdpn"] Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.241423 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-7mdpn" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.251856 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.255042 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-7mdpn"] Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.313356 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlqwd\" (UniqueName: \"kubernetes.io/projected/683ec07f-9451-46ca-b41f-fe332bf2284c-kube-api-access-hlqwd\") pod \"dnsmasq-dns-5bf47b49b7-xkcs4\" (UID: \"683ec07f-9451-46ca-b41f-fe332bf2284c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xkcs4" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.334602 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-config\") pod \"dnsmasq-dns-8554648995-7mdpn\" (UID: \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\") " pod="openstack/dnsmasq-dns-8554648995-7mdpn" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.334680 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9twth\" (UniqueName: \"kubernetes.io/projected/de34e733-406f-4d3c-8c30-e36c41fdf3a1-kube-api-access-9twth\") pod \"dnsmasq-dns-8554648995-7mdpn\" (UID: \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\") " pod="openstack/dnsmasq-dns-8554648995-7mdpn" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.334782 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-dns-svc\") pod \"dnsmasq-dns-8554648995-7mdpn\" (UID: \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\") " pod="openstack/dnsmasq-dns-8554648995-7mdpn" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.334819 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-7mdpn\" (UID: \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\") " pod="openstack/dnsmasq-dns-8554648995-7mdpn" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.334964 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-7mdpn\" (UID: \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\") " pod="openstack/dnsmasq-dns-8554648995-7mdpn" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.436139 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-dns-svc\") pod \"dnsmasq-dns-8554648995-7mdpn\" (UID: \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\") " pod="openstack/dnsmasq-dns-8554648995-7mdpn" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.436206 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-7mdpn\" (UID: \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\") " pod="openstack/dnsmasq-dns-8554648995-7mdpn" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.436319 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-7mdpn\" (UID: \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\") " pod="openstack/dnsmasq-dns-8554648995-7mdpn" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.436353 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-config\") pod \"dnsmasq-dns-8554648995-7mdpn\" (UID: \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\") " pod="openstack/dnsmasq-dns-8554648995-7mdpn" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.436383 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9twth\" (UniqueName: \"kubernetes.io/projected/de34e733-406f-4d3c-8c30-e36c41fdf3a1-kube-api-access-9twth\") pod \"dnsmasq-dns-8554648995-7mdpn\" (UID: \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\") " pod="openstack/dnsmasq-dns-8554648995-7mdpn" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.437222 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-dns-svc\") pod \"dnsmasq-dns-8554648995-7mdpn\" (UID: \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\") " pod="openstack/dnsmasq-dns-8554648995-7mdpn" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.437361 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-config\") pod \"dnsmasq-dns-8554648995-7mdpn\" (UID: \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\") " pod="openstack/dnsmasq-dns-8554648995-7mdpn" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.437408 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-7mdpn\" (UID: \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\") " pod="openstack/dnsmasq-dns-8554648995-7mdpn" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.437439 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-7mdpn\" (UID: \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\") " pod="openstack/dnsmasq-dns-8554648995-7mdpn" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.509524 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9twth\" (UniqueName: \"kubernetes.io/projected/de34e733-406f-4d3c-8c30-e36c41fdf3a1-kube-api-access-9twth\") pod \"dnsmasq-dns-8554648995-7mdpn\" (UID: \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\") " pod="openstack/dnsmasq-dns-8554648995-7mdpn" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.706236 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-7mdpn" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.823211 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"aa692ff7-2356-4ff4-b9fe-08884e4081cf","Type":"ContainerStarted","Data":"ec396350a1da311b232393c00a8651565f79130843bddb720cbd326d3df4f5dd"} Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.823548 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.825227 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"954395e4-178d-46ef-a695-8895ebe444d8","Type":"ContainerStarted","Data":"5064f43cdc6d13260e1178614c650dc41ca14f760d8d940c3c0641039ba164a7"} Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.837683 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6ghtc" event={"ID":"71963f95-01c9-473a-be8a-7798f6038ffb","Type":"ContainerStarted","Data":"efe7bcb5986b9144f49b78c91039396f4936c0c0228a4d1a2097f46faaf5baf7"} Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.837713 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-cmzfk" event={"ID":"468c5f0c-30d9-4493-a87c-2f4482a3b4d3","Type":"ContainerStarted","Data":"bcf5722360635ac4226fb33166a9a0f730c9f769eb32bce3d1be1f97ffc70eae"} Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.841758 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-xkcs4" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.847513 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0","Type":"ContainerStarted","Data":"84d5f79a02963d431e4b2b22e8c479a860d1676c7ccd73a598d1c7c0726230be"} Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.862521 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-xkcs4" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.867161 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=13.895521309 podStartE2EDuration="27.867134824s" podCreationTimestamp="2025-11-26 00:43:30 +0000 UTC" firstStartedPulling="2025-11-26 00:43:42.345503908 +0000 UTC m=+1203.194274338" lastFinishedPulling="2025-11-26 00:43:56.317117423 +0000 UTC m=+1217.165887853" observedRunningTime="2025-11-26 00:43:57.852578539 +0000 UTC m=+1218.701348969" watchObservedRunningTime="2025-11-26 00:43:57.867134824 +0000 UTC m=+1218.715905254" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.927188 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-qnhgk"] Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.934078 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-cmzfk" podStartSLOduration=15.208841355 podStartE2EDuration="27.934062643s" podCreationTimestamp="2025-11-26 00:43:30 +0000 UTC" firstStartedPulling="2025-11-26 00:43:42.350875532 +0000 UTC m=+1203.199645962" lastFinishedPulling="2025-11-26 00:43:55.07609683 +0000 UTC m=+1215.924867250" observedRunningTime="2025-11-26 00:43:57.908715807 +0000 UTC m=+1218.757486257" watchObservedRunningTime="2025-11-26 00:43:57.934062643 +0000 UTC m=+1218.782833073" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.953701 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlqwd\" (UniqueName: \"kubernetes.io/projected/683ec07f-9451-46ca-b41f-fe332bf2284c-kube-api-access-hlqwd\") pod \"683ec07f-9451-46ca-b41f-fe332bf2284c\" (UID: \"683ec07f-9451-46ca-b41f-fe332bf2284c\") " Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.953801 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/683ec07f-9451-46ca-b41f-fe332bf2284c-ovsdbserver-nb\") pod \"683ec07f-9451-46ca-b41f-fe332bf2284c\" (UID: \"683ec07f-9451-46ca-b41f-fe332bf2284c\") " Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.953879 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/683ec07f-9451-46ca-b41f-fe332bf2284c-dns-svc\") pod \"683ec07f-9451-46ca-b41f-fe332bf2284c\" (UID: \"683ec07f-9451-46ca-b41f-fe332bf2284c\") " Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.953972 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/683ec07f-9451-46ca-b41f-fe332bf2284c-config\") pod \"683ec07f-9451-46ca-b41f-fe332bf2284c\" (UID: \"683ec07f-9451-46ca-b41f-fe332bf2284c\") " Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.954424 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/683ec07f-9451-46ca-b41f-fe332bf2284c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "683ec07f-9451-46ca-b41f-fe332bf2284c" (UID: "683ec07f-9451-46ca-b41f-fe332bf2284c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.954850 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/683ec07f-9451-46ca-b41f-fe332bf2284c-config" (OuterVolumeSpecName: "config") pod "683ec07f-9451-46ca-b41f-fe332bf2284c" (UID: "683ec07f-9451-46ca-b41f-fe332bf2284c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.955595 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/683ec07f-9451-46ca-b41f-fe332bf2284c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "683ec07f-9451-46ca-b41f-fe332bf2284c" (UID: "683ec07f-9451-46ca-b41f-fe332bf2284c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.958439 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/683ec07f-9451-46ca-b41f-fe332bf2284c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.958528 4766 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/683ec07f-9451-46ca-b41f-fe332bf2284c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.958540 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/683ec07f-9451-46ca-b41f-fe332bf2284c-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:43:57 crc kubenswrapper[4766]: I1126 00:43:57.967602 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/683ec07f-9451-46ca-b41f-fe332bf2284c-kube-api-access-hlqwd" (OuterVolumeSpecName: "kube-api-access-hlqwd") pod "683ec07f-9451-46ca-b41f-fe332bf2284c" (UID: "683ec07f-9451-46ca-b41f-fe332bf2284c"). InnerVolumeSpecName "kube-api-access-hlqwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:43:58 crc kubenswrapper[4766]: I1126 00:43:58.060256 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlqwd\" (UniqueName: \"kubernetes.io/projected/683ec07f-9451-46ca-b41f-fe332bf2284c-kube-api-access-hlqwd\") on node \"crc\" DevicePath \"\"" Nov 26 00:43:58 crc kubenswrapper[4766]: I1126 00:43:58.254588 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-7mdpn"] Nov 26 00:43:58 crc kubenswrapper[4766]: W1126 00:43:58.270478 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde34e733_406f_4d3c_8c30_e36c41fdf3a1.slice/crio-67fc6f75f029111fb0ed46f71b4741e95e39b3b9e798d51e9e4c8224b2a2c4af WatchSource:0}: Error finding container 67fc6f75f029111fb0ed46f71b4741e95e39b3b9e798d51e9e4c8224b2a2c4af: Status 404 returned error can't find the container with id 67fc6f75f029111fb0ed46f71b4741e95e39b3b9e798d51e9e4c8224b2a2c4af Nov 26 00:43:58 crc kubenswrapper[4766]: I1126 00:43:58.476807 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-666b6646f7-jb2rz" podUID="cab5b48f-5422-4e6b-a207-c02958c69f79" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.118:5353: i/o timeout" Nov 26 00:43:58 crc kubenswrapper[4766]: I1126 00:43:58.858374 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p42tr" event={"ID":"22168318-cef8-4363-8b3e-3042e82037ea","Type":"ContainerStarted","Data":"a5be603e4720129211f767e661cdcfbff384e3c60d50d62968e8b5d50377db36"} Nov 26 00:43:58 crc kubenswrapper[4766]: I1126 00:43:58.858445 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-p42tr" Nov 26 00:43:58 crc kubenswrapper[4766]: I1126 00:43:58.862836 4766 generic.go:334] "Generic (PLEG): container finished" podID="71963f95-01c9-473a-be8a-7798f6038ffb" containerID="efe7bcb5986b9144f49b78c91039396f4936c0c0228a4d1a2097f46faaf5baf7" exitCode=0 Nov 26 00:43:58 crc kubenswrapper[4766]: I1126 00:43:58.863627 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6ghtc" event={"ID":"71963f95-01c9-473a-be8a-7798f6038ffb","Type":"ContainerDied","Data":"efe7bcb5986b9144f49b78c91039396f4936c0c0228a4d1a2097f46faaf5baf7"} Nov 26 00:43:58 crc kubenswrapper[4766]: I1126 00:43:58.872000 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d346209c-ec54-4719-9b56-7a079c603ebc","Type":"ContainerStarted","Data":"1b4ccddabed312260cdd833baeb0a48bf2126a76463faecc36772c3504bb2ed3"} Nov 26 00:43:58 crc kubenswrapper[4766]: I1126 00:43:58.878448 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-qnhgk" event={"ID":"342b8104-66eb-4fe1-88ad-ef5afc617d65","Type":"ContainerStarted","Data":"0d74f230972d2946cb0bb18fd4281ba9da4b70bc8f4471c8162fe21498679842"} Nov 26 00:43:58 crc kubenswrapper[4766]: I1126 00:43:58.879481 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-p42tr" podStartSLOduration=12.91958413 podStartE2EDuration="25.879462552s" podCreationTimestamp="2025-11-26 00:43:33 +0000 UTC" firstStartedPulling="2025-11-26 00:43:42.555006582 +0000 UTC m=+1203.403777012" lastFinishedPulling="2025-11-26 00:43:55.514884994 +0000 UTC m=+1216.363655434" observedRunningTime="2025-11-26 00:43:58.875280387 +0000 UTC m=+1219.724050817" watchObservedRunningTime="2025-11-26 00:43:58.879462552 +0000 UTC m=+1219.728232982" Nov 26 00:43:58 crc kubenswrapper[4766]: I1126 00:43:58.885610 4766 generic.go:334] "Generic (PLEG): container finished" podID="de34e733-406f-4d3c-8c30-e36c41fdf3a1" containerID="88feb284c7e95b21c8d7ffe541f9c8bc6c46f6e259cbbe36ad1dbc891ae8723e" exitCode=0 Nov 26 00:43:58 crc kubenswrapper[4766]: I1126 00:43:58.885686 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-7mdpn" event={"ID":"de34e733-406f-4d3c-8c30-e36c41fdf3a1","Type":"ContainerDied","Data":"88feb284c7e95b21c8d7ffe541f9c8bc6c46f6e259cbbe36ad1dbc891ae8723e"} Nov 26 00:43:58 crc kubenswrapper[4766]: I1126 00:43:58.885950 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-7mdpn" event={"ID":"de34e733-406f-4d3c-8c30-e36c41fdf3a1","Type":"ContainerStarted","Data":"67fc6f75f029111fb0ed46f71b4741e95e39b3b9e798d51e9e4c8224b2a2c4af"} Nov 26 00:43:58 crc kubenswrapper[4766]: I1126 00:43:58.886327 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-xkcs4" Nov 26 00:43:58 crc kubenswrapper[4766]: I1126 00:43:58.981546 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-xkcs4"] Nov 26 00:43:59 crc kubenswrapper[4766]: I1126 00:43:59.002545 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-xkcs4"] Nov 26 00:43:59 crc kubenswrapper[4766]: I1126 00:43:59.840434 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="683ec07f-9451-46ca-b41f-fe332bf2284c" path="/var/lib/kubelet/pods/683ec07f-9451-46ca-b41f-fe332bf2284c/volumes" Nov 26 00:43:59 crc kubenswrapper[4766]: I1126 00:43:59.897761 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-7mdpn" event={"ID":"de34e733-406f-4d3c-8c30-e36c41fdf3a1","Type":"ContainerStarted","Data":"a7d6d18d696bd29e80eb1e79e59a4744e8f12f5145baf077cd2db32620687e1a"} Nov 26 00:43:59 crc kubenswrapper[4766]: I1126 00:43:59.897862 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-7mdpn" Nov 26 00:43:59 crc kubenswrapper[4766]: I1126 00:43:59.905922 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6e23ce44-c896-41b1-b427-1bc2db9955ca","Type":"ContainerStarted","Data":"a70fc172683be2ed9e0b00864643e55a3d8ede537b6d0359d28f439575913435"} Nov 26 00:43:59 crc kubenswrapper[4766]: I1126 00:43:59.908225 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6ghtc" event={"ID":"71963f95-01c9-473a-be8a-7798f6038ffb","Type":"ContainerStarted","Data":"d5cba9a3600b293c079aee991364b4cf3e32da6c7000b59ae7fb7c711f03d9bc"} Nov 26 00:43:59 crc kubenswrapper[4766]: I1126 00:43:59.908270 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6ghtc" event={"ID":"71963f95-01c9-473a-be8a-7798f6038ffb","Type":"ContainerStarted","Data":"2a4a6e0916a9a36d904cb743811cd75c304a75ce00666e0088209a9b9abd6537"} Nov 26 00:43:59 crc kubenswrapper[4766]: I1126 00:43:59.908452 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:43:59 crc kubenswrapper[4766]: I1126 00:43:59.925602 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-7mdpn" podStartSLOduration=2.925579417 podStartE2EDuration="2.925579417s" podCreationTimestamp="2025-11-26 00:43:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:43:59.91811174 +0000 UTC m=+1220.766882180" watchObservedRunningTime="2025-11-26 00:43:59.925579417 +0000 UTC m=+1220.774349847" Nov 26 00:43:59 crc kubenswrapper[4766]: I1126 00:43:59.941968 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-6ghtc" podStartSLOduration=16.7527325 podStartE2EDuration="26.941950018s" podCreationTimestamp="2025-11-26 00:43:33 +0000 UTC" firstStartedPulling="2025-11-26 00:43:45.047599102 +0000 UTC m=+1205.896369532" lastFinishedPulling="2025-11-26 00:43:55.23681662 +0000 UTC m=+1216.085587050" observedRunningTime="2025-11-26 00:43:59.938828919 +0000 UTC m=+1220.787599359" watchObservedRunningTime="2025-11-26 00:43:59.941950018 +0000 UTC m=+1220.790720458" Nov 26 00:44:00 crc kubenswrapper[4766]: I1126 00:44:00.918534 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:44:01 crc kubenswrapper[4766]: I1126 00:44:01.931468 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-qnhgk" event={"ID":"342b8104-66eb-4fe1-88ad-ef5afc617d65","Type":"ContainerStarted","Data":"0d665643b748bc66e6b10273f1b686e075db393410e1f4e82867193099bf847a"} Nov 26 00:44:01 crc kubenswrapper[4766]: I1126 00:44:01.933792 4766 generic.go:334] "Generic (PLEG): container finished" podID="44f36c8b-51b8-4bce-afbb-1962eb317dec" containerID="923681756e1854ea0c12a9b2793aa6e5bba85d62da3fde1ee778ab1ae93829cf" exitCode=0 Nov 26 00:44:01 crc kubenswrapper[4766]: I1126 00:44:01.933835 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"44f36c8b-51b8-4bce-afbb-1962eb317dec","Type":"ContainerDied","Data":"923681756e1854ea0c12a9b2793aa6e5bba85d62da3fde1ee778ab1ae93829cf"} Nov 26 00:44:01 crc kubenswrapper[4766]: I1126 00:44:01.935615 4766 generic.go:334] "Generic (PLEG): container finished" podID="54c97191-0bfc-4a65-8168-10833aab2814" containerID="fd95c4ef8cd316d6ba0ccd2a83dbff58b228eb06197012de4ff863af48e90124" exitCode=0 Nov 26 00:44:01 crc kubenswrapper[4766]: I1126 00:44:01.935679 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"54c97191-0bfc-4a65-8168-10833aab2814","Type":"ContainerDied","Data":"fd95c4ef8cd316d6ba0ccd2a83dbff58b228eb06197012de4ff863af48e90124"} Nov 26 00:44:01 crc kubenswrapper[4766]: I1126 00:44:01.937807 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d346209c-ec54-4719-9b56-7a079c603ebc","Type":"ContainerStarted","Data":"e451168f528face84285d926b6268172a8d4a0c09944921f9ec1cfb8c6cf4604"} Nov 26 00:44:01 crc kubenswrapper[4766]: I1126 00:44:01.939296 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"954395e4-178d-46ef-a695-8895ebe444d8","Type":"ContainerStarted","Data":"bc32fb8a6689aedcfd139836bd59f83fdd2adf505c6f38af16211b9a2a02f966"} Nov 26 00:44:02 crc kubenswrapper[4766]: I1126 00:44:02.076868 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-qnhgk" podStartSLOduration=2.507442082 podStartE2EDuration="6.076847297s" podCreationTimestamp="2025-11-26 00:43:56 +0000 UTC" firstStartedPulling="2025-11-26 00:43:57.932110224 +0000 UTC m=+1218.780880654" lastFinishedPulling="2025-11-26 00:44:01.501515409 +0000 UTC m=+1222.350285869" observedRunningTime="2025-11-26 00:44:02.016309239 +0000 UTC m=+1222.865079669" watchObservedRunningTime="2025-11-26 00:44:02.076847297 +0000 UTC m=+1222.925617727" Nov 26 00:44:02 crc kubenswrapper[4766]: I1126 00:44:02.077675 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=14.668297644999999 podStartE2EDuration="29.077667398s" podCreationTimestamp="2025-11-26 00:43:33 +0000 UTC" firstStartedPulling="2025-11-26 00:43:47.081540139 +0000 UTC m=+1207.930310579" lastFinishedPulling="2025-11-26 00:44:01.490909902 +0000 UTC m=+1222.339680332" observedRunningTime="2025-11-26 00:44:02.061553353 +0000 UTC m=+1222.910323783" watchObservedRunningTime="2025-11-26 00:44:02.077667398 +0000 UTC m=+1222.926437828" Nov 26 00:44:02 crc kubenswrapper[4766]: I1126 00:44:02.126201 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=10.686872093 podStartE2EDuration="27.126178594s" podCreationTimestamp="2025-11-26 00:43:35 +0000 UTC" firstStartedPulling="2025-11-26 00:43:45.05032931 +0000 UTC m=+1205.899099740" lastFinishedPulling="2025-11-26 00:44:01.489635771 +0000 UTC m=+1222.338406241" observedRunningTime="2025-11-26 00:44:02.111757612 +0000 UTC m=+1222.960528062" watchObservedRunningTime="2025-11-26 00:44:02.126178594 +0000 UTC m=+1222.974949024" Nov 26 00:44:02 crc kubenswrapper[4766]: I1126 00:44:02.217924 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 26 00:44:02 crc kubenswrapper[4766]: I1126 00:44:02.957885 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"44f36c8b-51b8-4bce-afbb-1962eb317dec","Type":"ContainerStarted","Data":"1f361d6021276d295eb2ac91ab2c1a9c939e38b2a2d5153e5885f6521b707568"} Nov 26 00:44:02 crc kubenswrapper[4766]: I1126 00:44:02.962875 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"54c97191-0bfc-4a65-8168-10833aab2814","Type":"ContainerStarted","Data":"df867121088fdc1e05c03a63a46fbab6f3807946b078cb1c8a60b9481b47809f"} Nov 26 00:44:02 crc kubenswrapper[4766]: I1126 00:44:02.988697 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=31.511621139 podStartE2EDuration="36.988675064s" podCreationTimestamp="2025-11-26 00:43:26 +0000 UTC" firstStartedPulling="2025-11-26 00:43:42.351431626 +0000 UTC m=+1203.200202056" lastFinishedPulling="2025-11-26 00:43:47.828485551 +0000 UTC m=+1208.677255981" observedRunningTime="2025-11-26 00:44:02.982157761 +0000 UTC m=+1223.830928231" watchObservedRunningTime="2025-11-26 00:44:02.988675064 +0000 UTC m=+1223.837445504" Nov 26 00:44:03 crc kubenswrapper[4766]: I1126 00:44:03.002836 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=32.302264859 podStartE2EDuration="38.002820449s" podCreationTimestamp="2025-11-26 00:43:25 +0000 UTC" firstStartedPulling="2025-11-26 00:43:41.638008495 +0000 UTC m=+1202.486778935" lastFinishedPulling="2025-11-26 00:43:47.338564095 +0000 UTC m=+1208.187334525" observedRunningTime="2025-11-26 00:44:03.000431149 +0000 UTC m=+1223.849201589" watchObservedRunningTime="2025-11-26 00:44:03.002820449 +0000 UTC m=+1223.851590879" Nov 26 00:44:03 crc kubenswrapper[4766]: I1126 00:44:03.150885 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 26 00:44:04 crc kubenswrapper[4766]: I1126 00:44:04.218616 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 26 00:44:04 crc kubenswrapper[4766]: I1126 00:44:04.254003 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 26 00:44:04 crc kubenswrapper[4766]: I1126 00:44:04.486689 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 26 00:44:04 crc kubenswrapper[4766]: I1126 00:44:04.487040 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 26 00:44:04 crc kubenswrapper[4766]: I1126 00:44:04.545170 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.017976 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.025938 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.272497 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.274766 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.286550 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.286550 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.287308 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-d7c28" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.287428 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.298292 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.305935 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/64b18a9d-088b-4a4a-9c1a-4b63cbb1d642-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642\") " pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.305986 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/64b18a9d-088b-4a4a-9c1a-4b63cbb1d642-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642\") " pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.306058 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/64b18a9d-088b-4a4a-9c1a-4b63cbb1d642-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642\") " pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.306079 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdh95\" (UniqueName: \"kubernetes.io/projected/64b18a9d-088b-4a4a-9c1a-4b63cbb1d642-kube-api-access-fdh95\") pod \"ovn-northd-0\" (UID: \"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642\") " pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.306118 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64b18a9d-088b-4a4a-9c1a-4b63cbb1d642-scripts\") pod \"ovn-northd-0\" (UID: \"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642\") " pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.306182 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64b18a9d-088b-4a4a-9c1a-4b63cbb1d642-config\") pod \"ovn-northd-0\" (UID: \"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642\") " pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.306258 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64b18a9d-088b-4a4a-9c1a-4b63cbb1d642-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642\") " pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.408061 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64b18a9d-088b-4a4a-9c1a-4b63cbb1d642-config\") pod \"ovn-northd-0\" (UID: \"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642\") " pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.408208 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64b18a9d-088b-4a4a-9c1a-4b63cbb1d642-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642\") " pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.408297 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/64b18a9d-088b-4a4a-9c1a-4b63cbb1d642-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642\") " pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.408325 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/64b18a9d-088b-4a4a-9c1a-4b63cbb1d642-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642\") " pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.408402 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/64b18a9d-088b-4a4a-9c1a-4b63cbb1d642-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642\") " pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.408424 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdh95\" (UniqueName: \"kubernetes.io/projected/64b18a9d-088b-4a4a-9c1a-4b63cbb1d642-kube-api-access-fdh95\") pod \"ovn-northd-0\" (UID: \"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642\") " pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.408471 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64b18a9d-088b-4a4a-9c1a-4b63cbb1d642-scripts\") pod \"ovn-northd-0\" (UID: \"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642\") " pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.408853 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64b18a9d-088b-4a4a-9c1a-4b63cbb1d642-config\") pod \"ovn-northd-0\" (UID: \"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642\") " pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.408864 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/64b18a9d-088b-4a4a-9c1a-4b63cbb1d642-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642\") " pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.409299 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64b18a9d-088b-4a4a-9c1a-4b63cbb1d642-scripts\") pod \"ovn-northd-0\" (UID: \"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642\") " pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.415331 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/64b18a9d-088b-4a4a-9c1a-4b63cbb1d642-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642\") " pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.415752 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64b18a9d-088b-4a4a-9c1a-4b63cbb1d642-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642\") " pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.431423 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/64b18a9d-088b-4a4a-9c1a-4b63cbb1d642-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642\") " pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.431619 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdh95\" (UniqueName: \"kubernetes.io/projected/64b18a9d-088b-4a4a-9c1a-4b63cbb1d642-kube-api-access-fdh95\") pod \"ovn-northd-0\" (UID: \"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642\") " pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.597789 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 26 00:44:05 crc kubenswrapper[4766]: E1126 00:44:05.906308 4766 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.32:33508->38.102.83.32:36993: read tcp 38.102.83.32:33508->38.102.83.32:36993: read: connection reset by peer Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.987197 4766 generic.go:334] "Generic (PLEG): container finished" podID="6e23ce44-c896-41b1-b427-1bc2db9955ca" containerID="a70fc172683be2ed9e0b00864643e55a3d8ede537b6d0359d28f439575913435" exitCode=0 Nov 26 00:44:05 crc kubenswrapper[4766]: I1126 00:44:05.987398 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6e23ce44-c896-41b1-b427-1bc2db9955ca","Type":"ContainerDied","Data":"a70fc172683be2ed9e0b00864643e55a3d8ede537b6d0359d28f439575913435"} Nov 26 00:44:06 crc kubenswrapper[4766]: I1126 00:44:06.081038 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 26 00:44:06 crc kubenswrapper[4766]: W1126 00:44:06.081799 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64b18a9d_088b_4a4a_9c1a_4b63cbb1d642.slice/crio-87dfa44435a140dc02b52a4fb6c4bcbc83678b7b7ec2518efce9e3ffda4e3dcf WatchSource:0}: Error finding container 87dfa44435a140dc02b52a4fb6c4bcbc83678b7b7ec2518efce9e3ffda4e3dcf: Status 404 returned error can't find the container with id 87dfa44435a140dc02b52a4fb6c4bcbc83678b7b7ec2518efce9e3ffda4e3dcf Nov 26 00:44:06 crc kubenswrapper[4766]: I1126 00:44:06.480030 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 26 00:44:06 crc kubenswrapper[4766]: I1126 00:44:06.480336 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 26 00:44:07 crc kubenswrapper[4766]: I1126 00:44:07.000775 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642","Type":"ContainerStarted","Data":"87dfa44435a140dc02b52a4fb6c4bcbc83678b7b7ec2518efce9e3ffda4e3dcf"} Nov 26 00:44:07 crc kubenswrapper[4766]: I1126 00:44:07.175301 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 26 00:44:07 crc kubenswrapper[4766]: I1126 00:44:07.274910 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 26 00:44:07 crc kubenswrapper[4766]: I1126 00:44:07.707806 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-7mdpn" Nov 26 00:44:07 crc kubenswrapper[4766]: I1126 00:44:07.760315 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-chhxg"] Nov 26 00:44:07 crc kubenswrapper[4766]: I1126 00:44:07.760544 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" podUID="b064e590-4c62-4d62-841e-28a7d134632c" containerName="dnsmasq-dns" containerID="cri-o://02162a110dc761842427e2838e3c4b9002ed51c5213c5d3293a64dcf21c5fb46" gracePeriod=10 Nov 26 00:44:07 crc kubenswrapper[4766]: I1126 00:44:07.794850 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-b1a9-account-create-update-tzrh4"] Nov 26 00:44:07 crc kubenswrapper[4766]: I1126 00:44:07.795926 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b1a9-account-create-update-tzrh4" Nov 26 00:44:07 crc kubenswrapper[4766]: I1126 00:44:07.806058 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 26 00:44:07 crc kubenswrapper[4766]: I1126 00:44:07.818762 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 26 00:44:07 crc kubenswrapper[4766]: I1126 00:44:07.818800 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 26 00:44:07 crc kubenswrapper[4766]: I1126 00:44:07.846294 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b1a9-account-create-update-tzrh4"] Nov 26 00:44:07 crc kubenswrapper[4766]: I1126 00:44:07.906704 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-zn96g"] Nov 26 00:44:07 crc kubenswrapper[4766]: I1126 00:44:07.907877 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-zn96g" Nov 26 00:44:07 crc kubenswrapper[4766]: I1126 00:44:07.926191 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-zn96g"] Nov 26 00:44:07 crc kubenswrapper[4766]: I1126 00:44:07.981207 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcf23dba-a1e2-41be-b115-777bda2d8dc4-operator-scripts\") pod \"keystone-b1a9-account-create-update-tzrh4\" (UID: \"bcf23dba-a1e2-41be-b115-777bda2d8dc4\") " pod="openstack/keystone-b1a9-account-create-update-tzrh4" Nov 26 00:44:07 crc kubenswrapper[4766]: I1126 00:44:07.981303 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7qsg\" (UniqueName: \"kubernetes.io/projected/bcf23dba-a1e2-41be-b115-777bda2d8dc4-kube-api-access-l7qsg\") pod \"keystone-b1a9-account-create-update-tzrh4\" (UID: \"bcf23dba-a1e2-41be-b115-777bda2d8dc4\") " pod="openstack/keystone-b1a9-account-create-update-tzrh4" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.029550 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642","Type":"ContainerStarted","Data":"149664b703f5a0534dafb2fbc1d32bce495dbfb1ab8640efc9614eae2bc23dd7"} Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.029603 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"64b18a9d-088b-4a4a-9c1a-4b63cbb1d642","Type":"ContainerStarted","Data":"e12fbe00d15d511c4be694d2fe9f2d5189852c22333f06912a96073b2fcf241e"} Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.030500 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.032130 4766 generic.go:334] "Generic (PLEG): container finished" podID="b064e590-4c62-4d62-841e-28a7d134632c" containerID="02162a110dc761842427e2838e3c4b9002ed51c5213c5d3293a64dcf21c5fb46" exitCode=0 Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.032608 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" event={"ID":"b064e590-4c62-4d62-841e-28a7d134632c","Type":"ContainerDied","Data":"02162a110dc761842427e2838e3c4b9002ed51c5213c5d3293a64dcf21c5fb46"} Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.055095 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.002329442 podStartE2EDuration="3.055076044s" podCreationTimestamp="2025-11-26 00:44:05 +0000 UTC" firstStartedPulling="2025-11-26 00:44:06.083873839 +0000 UTC m=+1226.932644269" lastFinishedPulling="2025-11-26 00:44:07.136620441 +0000 UTC m=+1227.985390871" observedRunningTime="2025-11-26 00:44:08.052430278 +0000 UTC m=+1228.901200708" watchObservedRunningTime="2025-11-26 00:44:08.055076044 +0000 UTC m=+1228.903846644" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.079676 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-xx494"] Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.082455 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-xx494" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.083015 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2ec3192-d2ac-4f92-abd2-4c20548dceb3-operator-scripts\") pod \"keystone-db-create-zn96g\" (UID: \"c2ec3192-d2ac-4f92-abd2-4c20548dceb3\") " pod="openstack/keystone-db-create-zn96g" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.083102 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw9qj\" (UniqueName: \"kubernetes.io/projected/c2ec3192-d2ac-4f92-abd2-4c20548dceb3-kube-api-access-xw9qj\") pod \"keystone-db-create-zn96g\" (UID: \"c2ec3192-d2ac-4f92-abd2-4c20548dceb3\") " pod="openstack/keystone-db-create-zn96g" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.083158 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcf23dba-a1e2-41be-b115-777bda2d8dc4-operator-scripts\") pod \"keystone-b1a9-account-create-update-tzrh4\" (UID: \"bcf23dba-a1e2-41be-b115-777bda2d8dc4\") " pod="openstack/keystone-b1a9-account-create-update-tzrh4" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.083402 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7qsg\" (UniqueName: \"kubernetes.io/projected/bcf23dba-a1e2-41be-b115-777bda2d8dc4-kube-api-access-l7qsg\") pod \"keystone-b1a9-account-create-update-tzrh4\" (UID: \"bcf23dba-a1e2-41be-b115-777bda2d8dc4\") " pod="openstack/keystone-b1a9-account-create-update-tzrh4" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.085052 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcf23dba-a1e2-41be-b115-777bda2d8dc4-operator-scripts\") pod \"keystone-b1a9-account-create-update-tzrh4\" (UID: \"bcf23dba-a1e2-41be-b115-777bda2d8dc4\") " pod="openstack/keystone-b1a9-account-create-update-tzrh4" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.103529 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-fbd2-account-create-update-pcmbx"] Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.105002 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-fbd2-account-create-update-pcmbx" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.107782 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.111854 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7qsg\" (UniqueName: \"kubernetes.io/projected/bcf23dba-a1e2-41be-b115-777bda2d8dc4-kube-api-access-l7qsg\") pod \"keystone-b1a9-account-create-update-tzrh4\" (UID: \"bcf23dba-a1e2-41be-b115-777bda2d8dc4\") " pod="openstack/keystone-b1a9-account-create-update-tzrh4" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.126143 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.140472 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-xx494"] Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.171770 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-fbd2-account-create-update-pcmbx"] Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.172120 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b1a9-account-create-update-tzrh4" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.185199 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2ec3192-d2ac-4f92-abd2-4c20548dceb3-operator-scripts\") pod \"keystone-db-create-zn96g\" (UID: \"c2ec3192-d2ac-4f92-abd2-4c20548dceb3\") " pod="openstack/keystone-db-create-zn96g" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.185274 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrrdn\" (UniqueName: \"kubernetes.io/projected/34ebe8c7-f89d-478a-ba71-4160910955c6-kube-api-access-lrrdn\") pod \"placement-db-create-xx494\" (UID: \"34ebe8c7-f89d-478a-ba71-4160910955c6\") " pod="openstack/placement-db-create-xx494" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.185356 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw9qj\" (UniqueName: \"kubernetes.io/projected/c2ec3192-d2ac-4f92-abd2-4c20548dceb3-kube-api-access-xw9qj\") pod \"keystone-db-create-zn96g\" (UID: \"c2ec3192-d2ac-4f92-abd2-4c20548dceb3\") " pod="openstack/keystone-db-create-zn96g" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.185472 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34ebe8c7-f89d-478a-ba71-4160910955c6-operator-scripts\") pod \"placement-db-create-xx494\" (UID: \"34ebe8c7-f89d-478a-ba71-4160910955c6\") " pod="openstack/placement-db-create-xx494" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.186753 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2ec3192-d2ac-4f92-abd2-4c20548dceb3-operator-scripts\") pod \"keystone-db-create-zn96g\" (UID: \"c2ec3192-d2ac-4f92-abd2-4c20548dceb3\") " pod="openstack/keystone-db-create-zn96g" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.211071 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw9qj\" (UniqueName: \"kubernetes.io/projected/c2ec3192-d2ac-4f92-abd2-4c20548dceb3-kube-api-access-xw9qj\") pod \"keystone-db-create-zn96g\" (UID: \"c2ec3192-d2ac-4f92-abd2-4c20548dceb3\") " pod="openstack/keystone-db-create-zn96g" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.231414 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.288044 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34ebe8c7-f89d-478a-ba71-4160910955c6-operator-scripts\") pod \"placement-db-create-xx494\" (UID: \"34ebe8c7-f89d-478a-ba71-4160910955c6\") " pod="openstack/placement-db-create-xx494" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.288212 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrrdn\" (UniqueName: \"kubernetes.io/projected/34ebe8c7-f89d-478a-ba71-4160910955c6-kube-api-access-lrrdn\") pod \"placement-db-create-xx494\" (UID: \"34ebe8c7-f89d-478a-ba71-4160910955c6\") " pod="openstack/placement-db-create-xx494" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.288262 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2kk2\" (UniqueName: \"kubernetes.io/projected/095a8e06-fc50-4b1c-b27d-9efb621c523a-kube-api-access-w2kk2\") pod \"placement-fbd2-account-create-update-pcmbx\" (UID: \"095a8e06-fc50-4b1c-b27d-9efb621c523a\") " pod="openstack/placement-fbd2-account-create-update-pcmbx" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.288318 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/095a8e06-fc50-4b1c-b27d-9efb621c523a-operator-scripts\") pod \"placement-fbd2-account-create-update-pcmbx\" (UID: \"095a8e06-fc50-4b1c-b27d-9efb621c523a\") " pod="openstack/placement-fbd2-account-create-update-pcmbx" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.290521 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34ebe8c7-f89d-478a-ba71-4160910955c6-operator-scripts\") pod \"placement-db-create-xx494\" (UID: \"34ebe8c7-f89d-478a-ba71-4160910955c6\") " pod="openstack/placement-db-create-xx494" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.316523 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrrdn\" (UniqueName: \"kubernetes.io/projected/34ebe8c7-f89d-478a-ba71-4160910955c6-kube-api-access-lrrdn\") pod \"placement-db-create-xx494\" (UID: \"34ebe8c7-f89d-478a-ba71-4160910955c6\") " pod="openstack/placement-db-create-xx494" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.322181 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-zn96g" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.390382 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2kk2\" (UniqueName: \"kubernetes.io/projected/095a8e06-fc50-4b1c-b27d-9efb621c523a-kube-api-access-w2kk2\") pod \"placement-fbd2-account-create-update-pcmbx\" (UID: \"095a8e06-fc50-4b1c-b27d-9efb621c523a\") " pod="openstack/placement-fbd2-account-create-update-pcmbx" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.390437 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/095a8e06-fc50-4b1c-b27d-9efb621c523a-operator-scripts\") pod \"placement-fbd2-account-create-update-pcmbx\" (UID: \"095a8e06-fc50-4b1c-b27d-9efb621c523a\") " pod="openstack/placement-fbd2-account-create-update-pcmbx" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.391751 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/095a8e06-fc50-4b1c-b27d-9efb621c523a-operator-scripts\") pod \"placement-fbd2-account-create-update-pcmbx\" (UID: \"095a8e06-fc50-4b1c-b27d-9efb621c523a\") " pod="openstack/placement-fbd2-account-create-update-pcmbx" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.401641 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-xx494" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.420259 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2kk2\" (UniqueName: \"kubernetes.io/projected/095a8e06-fc50-4b1c-b27d-9efb621c523a-kube-api-access-w2kk2\") pod \"placement-fbd2-account-create-update-pcmbx\" (UID: \"095a8e06-fc50-4b1c-b27d-9efb621c523a\") " pod="openstack/placement-fbd2-account-create-update-pcmbx" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.450065 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-fbd2-account-create-update-pcmbx" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.683665 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b1a9-account-create-update-tzrh4"] Nov 26 00:44:08 crc kubenswrapper[4766]: W1126 00:44:08.721041 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbcf23dba_a1e2_41be_b115_777bda2d8dc4.slice/crio-88dd922e495e0e5db8a846880b8c739507987a6d0922ec172adf6638ceed826a WatchSource:0}: Error finding container 88dd922e495e0e5db8a846880b8c739507987a6d0922ec172adf6638ceed826a: Status 404 returned error can't find the container with id 88dd922e495e0e5db8a846880b8c739507987a6d0922ec172adf6638ceed826a Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.833511 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.951012 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-xx494"] Nov 26 00:44:08 crc kubenswrapper[4766]: I1126 00:44:08.959552 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-zn96g"] Nov 26 00:44:08 crc kubenswrapper[4766]: W1126 00:44:08.964311 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34ebe8c7_f89d_478a_ba71_4160910955c6.slice/crio-1156f3a65af31add44d19db5fec6ab2272862798149a1d3a304a07278334cbff WatchSource:0}: Error finding container 1156f3a65af31add44d19db5fec6ab2272862798149a1d3a304a07278334cbff: Status 404 returned error can't find the container with id 1156f3a65af31add44d19db5fec6ab2272862798149a1d3a304a07278334cbff Nov 26 00:44:08 crc kubenswrapper[4766]: W1126 00:44:08.964637 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2ec3192_d2ac_4f92_abd2_4c20548dceb3.slice/crio-7b9331ee6ab03490823a82fea7208b81b39e8c71c388c976dd52fbe0d14337c4 WatchSource:0}: Error finding container 7b9331ee6ab03490823a82fea7208b81b39e8c71c388c976dd52fbe0d14337c4: Status 404 returned error can't find the container with id 7b9331ee6ab03490823a82fea7208b81b39e8c71c388c976dd52fbe0d14337c4 Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.001987 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b064e590-4c62-4d62-841e-28a7d134632c-dns-svc\") pod \"b064e590-4c62-4d62-841e-28a7d134632c\" (UID: \"b064e590-4c62-4d62-841e-28a7d134632c\") " Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.002169 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89mf2\" (UniqueName: \"kubernetes.io/projected/b064e590-4c62-4d62-841e-28a7d134632c-kube-api-access-89mf2\") pod \"b064e590-4c62-4d62-841e-28a7d134632c\" (UID: \"b064e590-4c62-4d62-841e-28a7d134632c\") " Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.002236 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b064e590-4c62-4d62-841e-28a7d134632c-config\") pod \"b064e590-4c62-4d62-841e-28a7d134632c\" (UID: \"b064e590-4c62-4d62-841e-28a7d134632c\") " Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.011572 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b064e590-4c62-4d62-841e-28a7d134632c-kube-api-access-89mf2" (OuterVolumeSpecName: "kube-api-access-89mf2") pod "b064e590-4c62-4d62-841e-28a7d134632c" (UID: "b064e590-4c62-4d62-841e-28a7d134632c"). InnerVolumeSpecName "kube-api-access-89mf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.048296 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-xx494" event={"ID":"34ebe8c7-f89d-478a-ba71-4160910955c6","Type":"ContainerStarted","Data":"1156f3a65af31add44d19db5fec6ab2272862798149a1d3a304a07278334cbff"} Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.052031 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-zn96g" event={"ID":"c2ec3192-d2ac-4f92-abd2-4c20548dceb3","Type":"ContainerStarted","Data":"7b9331ee6ab03490823a82fea7208b81b39e8c71c388c976dd52fbe0d14337c4"} Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.054504 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" event={"ID":"b064e590-4c62-4d62-841e-28a7d134632c","Type":"ContainerDied","Data":"81199ac01bc23f7877aa38c040004ebb21d79db326a83af3d16610b4d6f447dd"} Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.054545 4766 scope.go:117] "RemoveContainer" containerID="02162a110dc761842427e2838e3c4b9002ed51c5213c5d3293a64dcf21c5fb46" Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.054554 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.056874 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b1a9-account-create-update-tzrh4" event={"ID":"bcf23dba-a1e2-41be-b115-777bda2d8dc4","Type":"ContainerStarted","Data":"ffe5fab3d7cf671c80721bf7746656bd38ec42271a90b576b96afcc749301564"} Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.056906 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b1a9-account-create-update-tzrh4" event={"ID":"bcf23dba-a1e2-41be-b115-777bda2d8dc4","Type":"ContainerStarted","Data":"88dd922e495e0e5db8a846880b8c739507987a6d0922ec172adf6638ceed826a"} Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.060769 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b064e590-4c62-4d62-841e-28a7d134632c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b064e590-4c62-4d62-841e-28a7d134632c" (UID: "b064e590-4c62-4d62-841e-28a7d134632c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.077545 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-b1a9-account-create-update-tzrh4" podStartSLOduration=2.077528715 podStartE2EDuration="2.077528715s" podCreationTimestamp="2025-11-26 00:44:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:44:09.072892879 +0000 UTC m=+1229.921663309" watchObservedRunningTime="2025-11-26 00:44:09.077528715 +0000 UTC m=+1229.926299145" Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.081969 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b064e590-4c62-4d62-841e-28a7d134632c-config" (OuterVolumeSpecName: "config") pod "b064e590-4c62-4d62-841e-28a7d134632c" (UID: "b064e590-4c62-4d62-841e-28a7d134632c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.088801 4766 scope.go:117] "RemoveContainer" containerID="be727cd0c6c7e68fa18dc05725614418c2605ff5f44a3831b5630d5171c27263" Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.103817 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b064e590-4c62-4d62-841e-28a7d134632c-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.103840 4766 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b064e590-4c62-4d62-841e-28a7d134632c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.103850 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89mf2\" (UniqueName: \"kubernetes.io/projected/b064e590-4c62-4d62-841e-28a7d134632c-kube-api-access-89mf2\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.154959 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-fbd2-account-create-update-pcmbx"] Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.504126 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-chhxg"] Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.512694 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-chhxg"] Nov 26 00:44:09 crc kubenswrapper[4766]: I1126 00:44:09.839458 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b064e590-4c62-4d62-841e-28a7d134632c" path="/var/lib/kubelet/pods/b064e590-4c62-4d62-841e-28a7d134632c/volumes" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.068176 4766 generic.go:334] "Generic (PLEG): container finished" podID="bcf23dba-a1e2-41be-b115-777bda2d8dc4" containerID="ffe5fab3d7cf671c80721bf7746656bd38ec42271a90b576b96afcc749301564" exitCode=0 Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.068239 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b1a9-account-create-update-tzrh4" event={"ID":"bcf23dba-a1e2-41be-b115-777bda2d8dc4","Type":"ContainerDied","Data":"ffe5fab3d7cf671c80721bf7746656bd38ec42271a90b576b96afcc749301564"} Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.070683 4766 generic.go:334] "Generic (PLEG): container finished" podID="095a8e06-fc50-4b1c-b27d-9efb621c523a" containerID="f57bf5ff37c0b77f28e0e5beab34da08d33c0e96573d481165a5a40dbb3ffadd" exitCode=0 Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.070754 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-fbd2-account-create-update-pcmbx" event={"ID":"095a8e06-fc50-4b1c-b27d-9efb621c523a","Type":"ContainerDied","Data":"f57bf5ff37c0b77f28e0e5beab34da08d33c0e96573d481165a5a40dbb3ffadd"} Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.070776 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-fbd2-account-create-update-pcmbx" event={"ID":"095a8e06-fc50-4b1c-b27d-9efb621c523a","Type":"ContainerStarted","Data":"457b64cb282af78bbf16a810811fd299bae09b3f156e2b0f6e502d2c7777bb97"} Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.072122 4766 generic.go:334] "Generic (PLEG): container finished" podID="34ebe8c7-f89d-478a-ba71-4160910955c6" containerID="4328dcf41f178bd6b54b42796098e40895df9e94b9bcebf8994c433d89780586" exitCode=0 Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.072152 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-xx494" event={"ID":"34ebe8c7-f89d-478a-ba71-4160910955c6","Type":"ContainerDied","Data":"4328dcf41f178bd6b54b42796098e40895df9e94b9bcebf8994c433d89780586"} Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.073722 4766 generic.go:334] "Generic (PLEG): container finished" podID="c2ec3192-d2ac-4f92-abd2-4c20548dceb3" containerID="617dc37b7d60acafb755ea09418c5c542e2891fb518eacc33195b5d71d7d306c" exitCode=0 Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.073770 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-zn96g" event={"ID":"c2ec3192-d2ac-4f92-abd2-4c20548dceb3","Type":"ContainerDied","Data":"617dc37b7d60acafb755ea09418c5c542e2891fb518eacc33195b5d71d7d306c"} Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.268925 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-92677"] Nov 26 00:44:10 crc kubenswrapper[4766]: E1126 00:44:10.269435 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b064e590-4c62-4d62-841e-28a7d134632c" containerName="dnsmasq-dns" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.269457 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="b064e590-4c62-4d62-841e-28a7d134632c" containerName="dnsmasq-dns" Nov 26 00:44:10 crc kubenswrapper[4766]: E1126 00:44:10.269498 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b064e590-4c62-4d62-841e-28a7d134632c" containerName="init" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.269506 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="b064e590-4c62-4d62-841e-28a7d134632c" containerName="init" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.269742 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="b064e590-4c62-4d62-841e-28a7d134632c" containerName="dnsmasq-dns" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.270590 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-92677" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.298754 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-92677"] Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.326219 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-2wx6s"] Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.331192 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.349917 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-2wx6s"] Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.419511 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-efe0-account-create-update-2kxdm"] Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.424640 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-efe0-account-create-update-2kxdm" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.429168 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-db-secret" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.435704 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-2wx6s\" (UID: \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\") " pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.435874 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68b0745b-3ae1-420b-ba0f-3cdbec6c0878-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-92677\" (UID: \"68b0745b-3ae1-420b-ba0f-3cdbec6c0878\") " pod="openstack/mysqld-exporter-openstack-db-create-92677" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.435919 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-2wx6s\" (UID: \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\") " pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.435962 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9rwb\" (UniqueName: \"kubernetes.io/projected/68b0745b-3ae1-420b-ba0f-3cdbec6c0878-kube-api-access-q9rwb\") pod \"mysqld-exporter-openstack-db-create-92677\" (UID: \"68b0745b-3ae1-420b-ba0f-3cdbec6c0878\") " pod="openstack/mysqld-exporter-openstack-db-create-92677" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.435996 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxwmw\" (UniqueName: \"kubernetes.io/projected/c4c2347f-5488-4f85-ab8d-c0a9324c1179-kube-api-access-fxwmw\") pod \"dnsmasq-dns-b8fbc5445-2wx6s\" (UID: \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\") " pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.436022 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-2wx6s\" (UID: \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\") " pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.436259 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-config\") pod \"dnsmasq-dns-b8fbc5445-2wx6s\" (UID: \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\") " pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.440683 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-efe0-account-create-update-2kxdm"] Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.449112 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.540727 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e-operator-scripts\") pod \"mysqld-exporter-efe0-account-create-update-2kxdm\" (UID: \"feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e\") " pod="openstack/mysqld-exporter-efe0-account-create-update-2kxdm" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.540802 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-config\") pod \"dnsmasq-dns-b8fbc5445-2wx6s\" (UID: \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\") " pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.540917 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-2wx6s\" (UID: \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\") " pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.541018 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68b0745b-3ae1-420b-ba0f-3cdbec6c0878-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-92677\" (UID: \"68b0745b-3ae1-420b-ba0f-3cdbec6c0878\") " pod="openstack/mysqld-exporter-openstack-db-create-92677" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.541122 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-2wx6s\" (UID: \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\") " pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.541267 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9rwb\" (UniqueName: \"kubernetes.io/projected/68b0745b-3ae1-420b-ba0f-3cdbec6c0878-kube-api-access-q9rwb\") pod \"mysqld-exporter-openstack-db-create-92677\" (UID: \"68b0745b-3ae1-420b-ba0f-3cdbec6c0878\") " pod="openstack/mysqld-exporter-openstack-db-create-92677" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.541308 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlb9j\" (UniqueName: \"kubernetes.io/projected/feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e-kube-api-access-vlb9j\") pod \"mysqld-exporter-efe0-account-create-update-2kxdm\" (UID: \"feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e\") " pod="openstack/mysqld-exporter-efe0-account-create-update-2kxdm" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.541402 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxwmw\" (UniqueName: \"kubernetes.io/projected/c4c2347f-5488-4f85-ab8d-c0a9324c1179-kube-api-access-fxwmw\") pod \"dnsmasq-dns-b8fbc5445-2wx6s\" (UID: \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\") " pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.541469 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-2wx6s\" (UID: \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\") " pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.541911 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-config\") pod \"dnsmasq-dns-b8fbc5445-2wx6s\" (UID: \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\") " pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.542034 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68b0745b-3ae1-420b-ba0f-3cdbec6c0878-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-92677\" (UID: \"68b0745b-3ae1-420b-ba0f-3cdbec6c0878\") " pod="openstack/mysqld-exporter-openstack-db-create-92677" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.542453 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-2wx6s\" (UID: \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\") " pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.542801 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-2wx6s\" (UID: \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\") " pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.543324 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-2wx6s\" (UID: \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\") " pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.569531 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxwmw\" (UniqueName: \"kubernetes.io/projected/c4c2347f-5488-4f85-ab8d-c0a9324c1179-kube-api-access-fxwmw\") pod \"dnsmasq-dns-b8fbc5445-2wx6s\" (UID: \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\") " pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.573190 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9rwb\" (UniqueName: \"kubernetes.io/projected/68b0745b-3ae1-420b-ba0f-3cdbec6c0878-kube-api-access-q9rwb\") pod \"mysqld-exporter-openstack-db-create-92677\" (UID: \"68b0745b-3ae1-420b-ba0f-3cdbec6c0878\") " pod="openstack/mysqld-exporter-openstack-db-create-92677" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.618444 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-92677" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.647511 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e-operator-scripts\") pod \"mysqld-exporter-efe0-account-create-update-2kxdm\" (UID: \"feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e\") " pod="openstack/mysqld-exporter-efe0-account-create-update-2kxdm" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.647773 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlb9j\" (UniqueName: \"kubernetes.io/projected/feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e-kube-api-access-vlb9j\") pod \"mysqld-exporter-efe0-account-create-update-2kxdm\" (UID: \"feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e\") " pod="openstack/mysqld-exporter-efe0-account-create-update-2kxdm" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.648601 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e-operator-scripts\") pod \"mysqld-exporter-efe0-account-create-update-2kxdm\" (UID: \"feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e\") " pod="openstack/mysqld-exporter-efe0-account-create-update-2kxdm" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.676543 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.677953 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlb9j\" (UniqueName: \"kubernetes.io/projected/feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e-kube-api-access-vlb9j\") pod \"mysqld-exporter-efe0-account-create-update-2kxdm\" (UID: \"feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e\") " pod="openstack/mysqld-exporter-efe0-account-create-update-2kxdm" Nov 26 00:44:10 crc kubenswrapper[4766]: I1126 00:44:10.752489 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-efe0-account-create-update-2kxdm" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.494925 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.501580 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.503469 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-5mxk5" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.503789 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.503905 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.504014 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.548594 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.669847 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") " pod="openstack/swift-storage-0" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.669893 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a1caa1fb-917a-4267-a947-1194557347eb-lock\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") " pod="openstack/swift-storage-0" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.669940 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh2db\" (UniqueName: \"kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-kube-api-access-dh2db\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") " pod="openstack/swift-storage-0" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.669968 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-etc-swift\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") " pod="openstack/swift-storage-0" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.669988 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a1caa1fb-917a-4267-a947-1194557347eb-cache\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") " pod="openstack/swift-storage-0" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.709322 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-tmfzn"] Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.711286 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.717062 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.717231 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.720183 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.720493 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-tmfzn"] Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.772209 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") " pod="openstack/swift-storage-0" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.772257 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a1caa1fb-917a-4267-a947-1194557347eb-lock\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") " pod="openstack/swift-storage-0" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.772304 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh2db\" (UniqueName: \"kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-kube-api-access-dh2db\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") " pod="openstack/swift-storage-0" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.772328 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-etc-swift\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") " pod="openstack/swift-storage-0" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.772353 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a1caa1fb-917a-4267-a947-1194557347eb-cache\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") " pod="openstack/swift-storage-0" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.772780 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/swift-storage-0" Nov 26 00:44:14 crc kubenswrapper[4766]: E1126 00:44:11.772842 4766 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 26 00:44:14 crc kubenswrapper[4766]: E1126 00:44:11.772858 4766 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 26 00:44:14 crc kubenswrapper[4766]: E1126 00:44:11.772903 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-etc-swift podName:a1caa1fb-917a-4267-a947-1194557347eb nodeName:}" failed. No retries permitted until 2025-11-26 00:44:12.272886331 +0000 UTC m=+1233.121656761 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-etc-swift") pod "swift-storage-0" (UID: "a1caa1fb-917a-4267-a947-1194557347eb") : configmap "swift-ring-files" not found Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.773490 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a1caa1fb-917a-4267-a947-1194557347eb-cache\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") " pod="openstack/swift-storage-0" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.773947 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a1caa1fb-917a-4267-a947-1194557347eb-lock\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") " pod="openstack/swift-storage-0" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.790884 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh2db\" (UniqueName: \"kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-kube-api-access-dh2db\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") " pod="openstack/swift-storage-0" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.798338 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") " pod="openstack/swift-storage-0" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.874638 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df26511-75fd-4bcb-ab3e-65cde68c9f46-combined-ca-bundle\") pod \"swift-ring-rebalance-tmfzn\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.874712 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6df26511-75fd-4bcb-ab3e-65cde68c9f46-dispersionconf\") pod \"swift-ring-rebalance-tmfzn\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.874785 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6df26511-75fd-4bcb-ab3e-65cde68c9f46-scripts\") pod \"swift-ring-rebalance-tmfzn\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.874856 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4p6r\" (UniqueName: \"kubernetes.io/projected/6df26511-75fd-4bcb-ab3e-65cde68c9f46-kube-api-access-c4p6r\") pod \"swift-ring-rebalance-tmfzn\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.874895 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6df26511-75fd-4bcb-ab3e-65cde68c9f46-swiftconf\") pod \"swift-ring-rebalance-tmfzn\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.874913 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6df26511-75fd-4bcb-ab3e-65cde68c9f46-etc-swift\") pod \"swift-ring-rebalance-tmfzn\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.874990 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6df26511-75fd-4bcb-ab3e-65cde68c9f46-ring-data-devices\") pod \"swift-ring-rebalance-tmfzn\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.976850 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6df26511-75fd-4bcb-ab3e-65cde68c9f46-swiftconf\") pod \"swift-ring-rebalance-tmfzn\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.976884 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6df26511-75fd-4bcb-ab3e-65cde68c9f46-etc-swift\") pod \"swift-ring-rebalance-tmfzn\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.976935 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6df26511-75fd-4bcb-ab3e-65cde68c9f46-ring-data-devices\") pod \"swift-ring-rebalance-tmfzn\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.977138 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df26511-75fd-4bcb-ab3e-65cde68c9f46-combined-ca-bundle\") pod \"swift-ring-rebalance-tmfzn\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.977168 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6df26511-75fd-4bcb-ab3e-65cde68c9f46-dispersionconf\") pod \"swift-ring-rebalance-tmfzn\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.978037 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6df26511-75fd-4bcb-ab3e-65cde68c9f46-etc-swift\") pod \"swift-ring-rebalance-tmfzn\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.978047 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6df26511-75fd-4bcb-ab3e-65cde68c9f46-scripts\") pod \"swift-ring-rebalance-tmfzn\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.978173 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4p6r\" (UniqueName: \"kubernetes.io/projected/6df26511-75fd-4bcb-ab3e-65cde68c9f46-kube-api-access-c4p6r\") pod \"swift-ring-rebalance-tmfzn\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.978664 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6df26511-75fd-4bcb-ab3e-65cde68c9f46-ring-data-devices\") pod \"swift-ring-rebalance-tmfzn\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.978673 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6df26511-75fd-4bcb-ab3e-65cde68c9f46-scripts\") pod \"swift-ring-rebalance-tmfzn\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.980632 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6df26511-75fd-4bcb-ab3e-65cde68c9f46-swiftconf\") pod \"swift-ring-rebalance-tmfzn\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.980788 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df26511-75fd-4bcb-ab3e-65cde68c9f46-combined-ca-bundle\") pod \"swift-ring-rebalance-tmfzn\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.995064 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6df26511-75fd-4bcb-ab3e-65cde68c9f46-dispersionconf\") pod \"swift-ring-rebalance-tmfzn\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:11.996700 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4p6r\" (UniqueName: \"kubernetes.io/projected/6df26511-75fd-4bcb-ab3e-65cde68c9f46-kube-api-access-c4p6r\") pod \"swift-ring-rebalance-tmfzn\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:12.028836 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:12.287241 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-etc-swift\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") " pod="openstack/swift-storage-0" Nov 26 00:44:14 crc kubenswrapper[4766]: E1126 00:44:12.287563 4766 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 26 00:44:14 crc kubenswrapper[4766]: E1126 00:44:12.287598 4766 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 26 00:44:14 crc kubenswrapper[4766]: E1126 00:44:12.287749 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-etc-swift podName:a1caa1fb-917a-4267-a947-1194557347eb nodeName:}" failed. No retries permitted until 2025-11-26 00:44:13.287722371 +0000 UTC m=+1234.136492801 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-etc-swift") pod "swift-storage-0" (UID: "a1caa1fb-917a-4267-a947-1194557347eb") : configmap "swift-ring-files" not found Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.332258 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-etc-swift\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") " pod="openstack/swift-storage-0" Nov 26 00:44:14 crc kubenswrapper[4766]: E1126 00:44:13.332530 4766 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 26 00:44:14 crc kubenswrapper[4766]: E1126 00:44:13.332546 4766 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 26 00:44:14 crc kubenswrapper[4766]: E1126 00:44:13.332612 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-etc-swift podName:a1caa1fb-917a-4267-a947-1194557347eb nodeName:}" failed. No retries permitted until 2025-11-26 00:44:15.332596345 +0000 UTC m=+1236.181366775 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-etc-swift") pod "swift-storage-0" (UID: "a1caa1fb-917a-4267-a947-1194557347eb") : configmap "swift-ring-files" not found Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.348586 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-29559"] Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.349969 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-29559" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.357449 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-29559"] Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.461339 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-3f15-account-create-update-5l657"] Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.463253 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3f15-account-create-update-5l657" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.465719 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.476643 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-3f15-account-create-update-5l657"] Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.540018 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78944618-3a02-482a-9965-ef46c0ece134-operator-scripts\") pod \"glance-db-create-29559\" (UID: \"78944618-3a02-482a-9965-ef46c0ece134\") " pod="openstack/glance-db-create-29559" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.540090 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czb69\" (UniqueName: \"kubernetes.io/projected/78944618-3a02-482a-9965-ef46c0ece134-kube-api-access-czb69\") pod \"glance-db-create-29559\" (UID: \"78944618-3a02-482a-9965-ef46c0ece134\") " pod="openstack/glance-db-create-29559" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.642055 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfjbh\" (UniqueName: \"kubernetes.io/projected/f7a0a3f7-ac14-4270-a51e-35738403f130-kube-api-access-zfjbh\") pod \"glance-3f15-account-create-update-5l657\" (UID: \"f7a0a3f7-ac14-4270-a51e-35738403f130\") " pod="openstack/glance-3f15-account-create-update-5l657" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.642223 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7a0a3f7-ac14-4270-a51e-35738403f130-operator-scripts\") pod \"glance-3f15-account-create-update-5l657\" (UID: \"f7a0a3f7-ac14-4270-a51e-35738403f130\") " pod="openstack/glance-3f15-account-create-update-5l657" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.642351 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78944618-3a02-482a-9965-ef46c0ece134-operator-scripts\") pod \"glance-db-create-29559\" (UID: \"78944618-3a02-482a-9965-ef46c0ece134\") " pod="openstack/glance-db-create-29559" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.642409 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czb69\" (UniqueName: \"kubernetes.io/projected/78944618-3a02-482a-9965-ef46c0ece134-kube-api-access-czb69\") pod \"glance-db-create-29559\" (UID: \"78944618-3a02-482a-9965-ef46c0ece134\") " pod="openstack/glance-db-create-29559" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.643543 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78944618-3a02-482a-9965-ef46c0ece134-operator-scripts\") pod \"glance-db-create-29559\" (UID: \"78944618-3a02-482a-9965-ef46c0ece134\") " pod="openstack/glance-db-create-29559" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.673173 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czb69\" (UniqueName: \"kubernetes.io/projected/78944618-3a02-482a-9965-ef46c0ece134-kube-api-access-czb69\") pod \"glance-db-create-29559\" (UID: \"78944618-3a02-482a-9965-ef46c0ece134\") " pod="openstack/glance-db-create-29559" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.745177 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7a0a3f7-ac14-4270-a51e-35738403f130-operator-scripts\") pod \"glance-3f15-account-create-update-5l657\" (UID: \"f7a0a3f7-ac14-4270-a51e-35738403f130\") " pod="openstack/glance-3f15-account-create-update-5l657" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.745324 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfjbh\" (UniqueName: \"kubernetes.io/projected/f7a0a3f7-ac14-4270-a51e-35738403f130-kube-api-access-zfjbh\") pod \"glance-3f15-account-create-update-5l657\" (UID: \"f7a0a3f7-ac14-4270-a51e-35738403f130\") " pod="openstack/glance-3f15-account-create-update-5l657" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.746045 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7a0a3f7-ac14-4270-a51e-35738403f130-operator-scripts\") pod \"glance-3f15-account-create-update-5l657\" (UID: \"f7a0a3f7-ac14-4270-a51e-35738403f130\") " pod="openstack/glance-3f15-account-create-update-5l657" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.760503 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfjbh\" (UniqueName: \"kubernetes.io/projected/f7a0a3f7-ac14-4270-a51e-35738403f130-kube-api-access-zfjbh\") pod \"glance-3f15-account-create-update-5l657\" (UID: \"f7a0a3f7-ac14-4270-a51e-35738403f130\") " pod="openstack/glance-3f15-account-create-update-5l657" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.767839 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57d769cc4f-chhxg" podUID="b064e590-4c62-4d62-841e-28a7d134632c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.119:5353: i/o timeout" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.799786 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3f15-account-create-update-5l657" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:13.970552 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-29559" Nov 26 00:44:14 crc kubenswrapper[4766]: I1126 00:44:14.980098 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-fbd2-account-create-update-pcmbx" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.007287 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b1a9-account-create-update-tzrh4" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.022813 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-92677"] Nov 26 00:44:15 crc kubenswrapper[4766]: W1126 00:44:15.023436 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68b0745b_3ae1_420b_ba0f_3cdbec6c0878.slice/crio-022f3422fe0a31cdd08004c9ee390ab31b733b6983ed637b06380a55cbcb5e29 WatchSource:0}: Error finding container 022f3422fe0a31cdd08004c9ee390ab31b733b6983ed637b06380a55cbcb5e29: Status 404 returned error can't find the container with id 022f3422fe0a31cdd08004c9ee390ab31b733b6983ed637b06380a55cbcb5e29 Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.023758 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-xx494" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.045005 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-zn96g" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.090009 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2kk2\" (UniqueName: \"kubernetes.io/projected/095a8e06-fc50-4b1c-b27d-9efb621c523a-kube-api-access-w2kk2\") pod \"095a8e06-fc50-4b1c-b27d-9efb621c523a\" (UID: \"095a8e06-fc50-4b1c-b27d-9efb621c523a\") " Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.090052 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/095a8e06-fc50-4b1c-b27d-9efb621c523a-operator-scripts\") pod \"095a8e06-fc50-4b1c-b27d-9efb621c523a\" (UID: \"095a8e06-fc50-4b1c-b27d-9efb621c523a\") " Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.091222 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/095a8e06-fc50-4b1c-b27d-9efb621c523a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "095a8e06-fc50-4b1c-b27d-9efb621c523a" (UID: "095a8e06-fc50-4b1c-b27d-9efb621c523a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.098921 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/095a8e06-fc50-4b1c-b27d-9efb621c523a-kube-api-access-w2kk2" (OuterVolumeSpecName: "kube-api-access-w2kk2") pod "095a8e06-fc50-4b1c-b27d-9efb621c523a" (UID: "095a8e06-fc50-4b1c-b27d-9efb621c523a"). InnerVolumeSpecName "kube-api-access-w2kk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.131704 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-fbd2-account-create-update-pcmbx" event={"ID":"095a8e06-fc50-4b1c-b27d-9efb621c523a","Type":"ContainerDied","Data":"457b64cb282af78bbf16a810811fd299bae09b3f156e2b0f6e502d2c7777bb97"} Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.131753 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="457b64cb282af78bbf16a810811fd299bae09b3f156e2b0f6e502d2c7777bb97" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.131826 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-fbd2-account-create-update-pcmbx" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.137567 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-xx494" event={"ID":"34ebe8c7-f89d-478a-ba71-4160910955c6","Type":"ContainerDied","Data":"1156f3a65af31add44d19db5fec6ab2272862798149a1d3a304a07278334cbff"} Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.137621 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1156f3a65af31add44d19db5fec6ab2272862798149a1d3a304a07278334cbff" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.137735 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-xx494" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.146081 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-zn96g" event={"ID":"c2ec3192-d2ac-4f92-abd2-4c20548dceb3","Type":"ContainerDied","Data":"7b9331ee6ab03490823a82fea7208b81b39e8c71c388c976dd52fbe0d14337c4"} Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.146132 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b9331ee6ab03490823a82fea7208b81b39e8c71c388c976dd52fbe0d14337c4" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.146215 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-zn96g" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.150716 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6e23ce44-c896-41b1-b427-1bc2db9955ca","Type":"ContainerStarted","Data":"b3476a207d9b96473d9efbe51664aa5329ef1bee2f07196acdf8785cbf6848bd"} Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.152934 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b1a9-account-create-update-tzrh4" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.152923 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b1a9-account-create-update-tzrh4" event={"ID":"bcf23dba-a1e2-41be-b115-777bda2d8dc4","Type":"ContainerDied","Data":"88dd922e495e0e5db8a846880b8c739507987a6d0922ec172adf6638ceed826a"} Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.153123 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88dd922e495e0e5db8a846880b8c739507987a6d0922ec172adf6638ceed826a" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.154454 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-92677" event={"ID":"68b0745b-3ae1-420b-ba0f-3cdbec6c0878","Type":"ContainerStarted","Data":"022f3422fe0a31cdd08004c9ee390ab31b733b6983ed637b06380a55cbcb5e29"} Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.181810 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-efe0-account-create-update-2kxdm"] Nov 26 00:44:15 crc kubenswrapper[4766]: W1126 00:44:15.185119 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfeb7ecda_6e5c_4a6e_baf8_fe4c87f50d4e.slice/crio-fc1eefbcc53e7ed906ad89c9f81f3bb2f3db4b95193339dd3e317ee7e004c2e2 WatchSource:0}: Error finding container fc1eefbcc53e7ed906ad89c9f81f3bb2f3db4b95193339dd3e317ee7e004c2e2: Status 404 returned error can't find the container with id fc1eefbcc53e7ed906ad89c9f81f3bb2f3db4b95193339dd3e317ee7e004c2e2 Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.191087 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrrdn\" (UniqueName: \"kubernetes.io/projected/34ebe8c7-f89d-478a-ba71-4160910955c6-kube-api-access-lrrdn\") pod \"34ebe8c7-f89d-478a-ba71-4160910955c6\" (UID: \"34ebe8c7-f89d-478a-ba71-4160910955c6\") " Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.191332 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7qsg\" (UniqueName: \"kubernetes.io/projected/bcf23dba-a1e2-41be-b115-777bda2d8dc4-kube-api-access-l7qsg\") pod \"bcf23dba-a1e2-41be-b115-777bda2d8dc4\" (UID: \"bcf23dba-a1e2-41be-b115-777bda2d8dc4\") " Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.191365 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2ec3192-d2ac-4f92-abd2-4c20548dceb3-operator-scripts\") pod \"c2ec3192-d2ac-4f92-abd2-4c20548dceb3\" (UID: \"c2ec3192-d2ac-4f92-abd2-4c20548dceb3\") " Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.191400 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34ebe8c7-f89d-478a-ba71-4160910955c6-operator-scripts\") pod \"34ebe8c7-f89d-478a-ba71-4160910955c6\" (UID: \"34ebe8c7-f89d-478a-ba71-4160910955c6\") " Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.191431 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcf23dba-a1e2-41be-b115-777bda2d8dc4-operator-scripts\") pod \"bcf23dba-a1e2-41be-b115-777bda2d8dc4\" (UID: \"bcf23dba-a1e2-41be-b115-777bda2d8dc4\") " Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.191463 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xw9qj\" (UniqueName: \"kubernetes.io/projected/c2ec3192-d2ac-4f92-abd2-4c20548dceb3-kube-api-access-xw9qj\") pod \"c2ec3192-d2ac-4f92-abd2-4c20548dceb3\" (UID: \"c2ec3192-d2ac-4f92-abd2-4c20548dceb3\") " Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.191877 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcf23dba-a1e2-41be-b115-777bda2d8dc4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bcf23dba-a1e2-41be-b115-777bda2d8dc4" (UID: "bcf23dba-a1e2-41be-b115-777bda2d8dc4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.191925 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34ebe8c7-f89d-478a-ba71-4160910955c6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "34ebe8c7-f89d-478a-ba71-4160910955c6" (UID: "34ebe8c7-f89d-478a-ba71-4160910955c6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.191978 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2ec3192-d2ac-4f92-abd2-4c20548dceb3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c2ec3192-d2ac-4f92-abd2-4c20548dceb3" (UID: "c2ec3192-d2ac-4f92-abd2-4c20548dceb3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.192134 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2ec3192-d2ac-4f92-abd2-4c20548dceb3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.192150 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34ebe8c7-f89d-478a-ba71-4160910955c6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.192161 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcf23dba-a1e2-41be-b115-777bda2d8dc4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.192172 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2kk2\" (UniqueName: \"kubernetes.io/projected/095a8e06-fc50-4b1c-b27d-9efb621c523a-kube-api-access-w2kk2\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.192185 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/095a8e06-fc50-4b1c-b27d-9efb621c523a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.195351 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2ec3192-d2ac-4f92-abd2-4c20548dceb3-kube-api-access-xw9qj" (OuterVolumeSpecName: "kube-api-access-xw9qj") pod "c2ec3192-d2ac-4f92-abd2-4c20548dceb3" (UID: "c2ec3192-d2ac-4f92-abd2-4c20548dceb3"). InnerVolumeSpecName "kube-api-access-xw9qj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.195892 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcf23dba-a1e2-41be-b115-777bda2d8dc4-kube-api-access-l7qsg" (OuterVolumeSpecName: "kube-api-access-l7qsg") pod "bcf23dba-a1e2-41be-b115-777bda2d8dc4" (UID: "bcf23dba-a1e2-41be-b115-777bda2d8dc4"). InnerVolumeSpecName "kube-api-access-l7qsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.196531 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34ebe8c7-f89d-478a-ba71-4160910955c6-kube-api-access-lrrdn" (OuterVolumeSpecName: "kube-api-access-lrrdn") pod "34ebe8c7-f89d-478a-ba71-4160910955c6" (UID: "34ebe8c7-f89d-478a-ba71-4160910955c6"). InnerVolumeSpecName "kube-api-access-lrrdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.293945 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7qsg\" (UniqueName: \"kubernetes.io/projected/bcf23dba-a1e2-41be-b115-777bda2d8dc4-kube-api-access-l7qsg\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.293975 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xw9qj\" (UniqueName: \"kubernetes.io/projected/c2ec3192-d2ac-4f92-abd2-4c20548dceb3-kube-api-access-xw9qj\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.293984 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrrdn\" (UniqueName: \"kubernetes.io/projected/34ebe8c7-f89d-478a-ba71-4160910955c6-kube-api-access-lrrdn\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.296587 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-2wx6s"] Nov 26 00:44:15 crc kubenswrapper[4766]: W1126 00:44:15.300318 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4c2347f_5488_4f85_ab8d_c0a9324c1179.slice/crio-c648fdbb68fe5ed5ad4a8b58910220f7c023bc3cc36cb6abe2d7a11f373afae4 WatchSource:0}: Error finding container c648fdbb68fe5ed5ad4a8b58910220f7c023bc3cc36cb6abe2d7a11f373afae4: Status 404 returned error can't find the container with id c648fdbb68fe5ed5ad4a8b58910220f7c023bc3cc36cb6abe2d7a11f373afae4 Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.389833 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-tmfzn"] Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.397221 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-etc-swift\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") " pod="openstack/swift-storage-0" Nov 26 00:44:15 crc kubenswrapper[4766]: E1126 00:44:15.397546 4766 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 26 00:44:15 crc kubenswrapper[4766]: E1126 00:44:15.397568 4766 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 26 00:44:15 crc kubenswrapper[4766]: E1126 00:44:15.397622 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-etc-swift podName:a1caa1fb-917a-4267-a947-1194557347eb nodeName:}" failed. No retries permitted until 2025-11-26 00:44:19.397601983 +0000 UTC m=+1240.246372413 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-etc-swift") pod "swift-storage-0" (UID: "a1caa1fb-917a-4267-a947-1194557347eb") : configmap "swift-ring-files" not found Nov 26 00:44:15 crc kubenswrapper[4766]: W1126 00:44:15.403632 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6df26511_75fd_4bcb_ab3e_65cde68c9f46.slice/crio-1eb5080cd70fb96e1a9e1dd44e5f78f41567d622808c0e060db6f81e1ce1db9f WatchSource:0}: Error finding container 1eb5080cd70fb96e1a9e1dd44e5f78f41567d622808c0e060db6f81e1ce1db9f: Status 404 returned error can't find the container with id 1eb5080cd70fb96e1a9e1dd44e5f78f41567d622808c0e060db6f81e1ce1db9f Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.407101 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-3f15-account-create-update-5l657"] Nov 26 00:44:15 crc kubenswrapper[4766]: I1126 00:44:15.429674 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-29559"] Nov 26 00:44:16 crc kubenswrapper[4766]: I1126 00:44:16.172100 4766 generic.go:334] "Generic (PLEG): container finished" podID="68b0745b-3ae1-420b-ba0f-3cdbec6c0878" containerID="d4c6a754f1e46d07682e51b007f6b063944a5b52ef1a6fec9d341fff511e1639" exitCode=0 Nov 26 00:44:16 crc kubenswrapper[4766]: I1126 00:44:16.172469 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-92677" event={"ID":"68b0745b-3ae1-420b-ba0f-3cdbec6c0878","Type":"ContainerDied","Data":"d4c6a754f1e46d07682e51b007f6b063944a5b52ef1a6fec9d341fff511e1639"} Nov 26 00:44:16 crc kubenswrapper[4766]: I1126 00:44:16.177124 4766 generic.go:334] "Generic (PLEG): container finished" podID="f7a0a3f7-ac14-4270-a51e-35738403f130" containerID="355ab80164cf5594f074f0ee22d5d68a441801af3e96264aca5d72a76f929c68" exitCode=0 Nov 26 00:44:16 crc kubenswrapper[4766]: I1126 00:44:16.177281 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-3f15-account-create-update-5l657" event={"ID":"f7a0a3f7-ac14-4270-a51e-35738403f130","Type":"ContainerDied","Data":"355ab80164cf5594f074f0ee22d5d68a441801af3e96264aca5d72a76f929c68"} Nov 26 00:44:16 crc kubenswrapper[4766]: I1126 00:44:16.177347 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-3f15-account-create-update-5l657" event={"ID":"f7a0a3f7-ac14-4270-a51e-35738403f130","Type":"ContainerStarted","Data":"91af94636c88f0325bf8db354af4eeeae7d3565b440ef184aa504104c3625e7e"} Nov 26 00:44:16 crc kubenswrapper[4766]: I1126 00:44:16.179173 4766 generic.go:334] "Generic (PLEG): container finished" podID="c4c2347f-5488-4f85-ab8d-c0a9324c1179" containerID="c21a4ede7b6d1408e7ae84444e13c7a43bac3b60340457338c7d5780e812e108" exitCode=0 Nov 26 00:44:16 crc kubenswrapper[4766]: I1126 00:44:16.179222 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" event={"ID":"c4c2347f-5488-4f85-ab8d-c0a9324c1179","Type":"ContainerDied","Data":"c21a4ede7b6d1408e7ae84444e13c7a43bac3b60340457338c7d5780e812e108"} Nov 26 00:44:16 crc kubenswrapper[4766]: I1126 00:44:16.179246 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" event={"ID":"c4c2347f-5488-4f85-ab8d-c0a9324c1179","Type":"ContainerStarted","Data":"c648fdbb68fe5ed5ad4a8b58910220f7c023bc3cc36cb6abe2d7a11f373afae4"} Nov 26 00:44:16 crc kubenswrapper[4766]: I1126 00:44:16.181346 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tmfzn" event={"ID":"6df26511-75fd-4bcb-ab3e-65cde68c9f46","Type":"ContainerStarted","Data":"1eb5080cd70fb96e1a9e1dd44e5f78f41567d622808c0e060db6f81e1ce1db9f"} Nov 26 00:44:16 crc kubenswrapper[4766]: I1126 00:44:16.182987 4766 generic.go:334] "Generic (PLEG): container finished" podID="78944618-3a02-482a-9965-ef46c0ece134" containerID="88bf9d7a8f06ab171a2e337b223f19130ce74feffb2baaa08750f237903784fe" exitCode=0 Nov 26 00:44:16 crc kubenswrapper[4766]: I1126 00:44:16.183088 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-29559" event={"ID":"78944618-3a02-482a-9965-ef46c0ece134","Type":"ContainerDied","Data":"88bf9d7a8f06ab171a2e337b223f19130ce74feffb2baaa08750f237903784fe"} Nov 26 00:44:16 crc kubenswrapper[4766]: I1126 00:44:16.183119 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-29559" event={"ID":"78944618-3a02-482a-9965-ef46c0ece134","Type":"ContainerStarted","Data":"501323ac945a2fd0292bb9b322d97d4bff6053f4705f98929f058a58e39af16e"} Nov 26 00:44:16 crc kubenswrapper[4766]: I1126 00:44:16.184442 4766 generic.go:334] "Generic (PLEG): container finished" podID="feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e" containerID="dcc1454763f21e13b343f565f06d9546ef5926d64aa121466f653116ecada477" exitCode=0 Nov 26 00:44:16 crc kubenswrapper[4766]: I1126 00:44:16.184469 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-efe0-account-create-update-2kxdm" event={"ID":"feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e","Type":"ContainerDied","Data":"dcc1454763f21e13b343f565f06d9546ef5926d64aa121466f653116ecada477"} Nov 26 00:44:16 crc kubenswrapper[4766]: I1126 00:44:16.184482 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-efe0-account-create-update-2kxdm" event={"ID":"feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e","Type":"ContainerStarted","Data":"fc1eefbcc53e7ed906ad89c9f81f3bb2f3db4b95193339dd3e317ee7e004c2e2"} Nov 26 00:44:16 crc kubenswrapper[4766]: I1126 00:44:16.837390 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-5bc7576dcd-sqbpf" podUID="e1ca8391-338e-410b-a7ce-83da6f733c87" containerName="console" containerID="cri-o://ece9fdfedc40fbd881d5f91cfbe1c0398a1f91bdecc08911f99a8d8e6be1e380" gracePeriod=15 Nov 26 00:44:17 crc kubenswrapper[4766]: I1126 00:44:17.197378 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6e23ce44-c896-41b1-b427-1bc2db9955ca","Type":"ContainerStarted","Data":"62285b131a0ff5824a4f3305369e761cddd7daf66d555ea5e28446a413301a1a"} Nov 26 00:44:17 crc kubenswrapper[4766]: I1126 00:44:17.199532 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5bc7576dcd-sqbpf_e1ca8391-338e-410b-a7ce-83da6f733c87/console/0.log" Nov 26 00:44:17 crc kubenswrapper[4766]: I1126 00:44:17.199577 4766 generic.go:334] "Generic (PLEG): container finished" podID="e1ca8391-338e-410b-a7ce-83da6f733c87" containerID="ece9fdfedc40fbd881d5f91cfbe1c0398a1f91bdecc08911f99a8d8e6be1e380" exitCode=2 Nov 26 00:44:17 crc kubenswrapper[4766]: I1126 00:44:17.199641 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5bc7576dcd-sqbpf" event={"ID":"e1ca8391-338e-410b-a7ce-83da6f733c87","Type":"ContainerDied","Data":"ece9fdfedc40fbd881d5f91cfbe1c0398a1f91bdecc08911f99a8d8e6be1e380"} Nov 26 00:44:17 crc kubenswrapper[4766]: I1126 00:44:17.203314 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" event={"ID":"c4c2347f-5488-4f85-ab8d-c0a9324c1179","Type":"ContainerStarted","Data":"968b54683480167549f5590c4888d75550afea27a313610b67f89220484f4e73"} Nov 26 00:44:17 crc kubenswrapper[4766]: I1126 00:44:17.203871 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" Nov 26 00:44:18 crc kubenswrapper[4766]: I1126 00:44:18.184436 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-92677" Nov 26 00:44:18 crc kubenswrapper[4766]: I1126 00:44:18.190270 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-29559" Nov 26 00:44:18 crc kubenswrapper[4766]: I1126 00:44:18.202838 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" podStartSLOduration=8.202780052 podStartE2EDuration="8.202780052s" podCreationTimestamp="2025-11-26 00:44:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:44:17.229571425 +0000 UTC m=+1238.078341875" watchObservedRunningTime="2025-11-26 00:44:18.202780052 +0000 UTC m=+1239.051550502" Nov 26 00:44:18 crc kubenswrapper[4766]: I1126 00:44:18.230180 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-29559" event={"ID":"78944618-3a02-482a-9965-ef46c0ece134","Type":"ContainerDied","Data":"501323ac945a2fd0292bb9b322d97d4bff6053f4705f98929f058a58e39af16e"} Nov 26 00:44:18 crc kubenswrapper[4766]: I1126 00:44:18.230227 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="501323ac945a2fd0292bb9b322d97d4bff6053f4705f98929f058a58e39af16e" Nov 26 00:44:18 crc kubenswrapper[4766]: I1126 00:44:18.230297 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-29559" Nov 26 00:44:18 crc kubenswrapper[4766]: I1126 00:44:18.234858 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-92677" Nov 26 00:44:18 crc kubenswrapper[4766]: I1126 00:44:18.234887 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-92677" event={"ID":"68b0745b-3ae1-420b-ba0f-3cdbec6c0878","Type":"ContainerDied","Data":"022f3422fe0a31cdd08004c9ee390ab31b733b6983ed637b06380a55cbcb5e29"} Nov 26 00:44:18 crc kubenswrapper[4766]: I1126 00:44:18.234961 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="022f3422fe0a31cdd08004c9ee390ab31b733b6983ed637b06380a55cbcb5e29" Nov 26 00:44:18 crc kubenswrapper[4766]: I1126 00:44:18.362120 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czb69\" (UniqueName: \"kubernetes.io/projected/78944618-3a02-482a-9965-ef46c0ece134-kube-api-access-czb69\") pod \"78944618-3a02-482a-9965-ef46c0ece134\" (UID: \"78944618-3a02-482a-9965-ef46c0ece134\") " Nov 26 00:44:18 crc kubenswrapper[4766]: I1126 00:44:18.362186 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78944618-3a02-482a-9965-ef46c0ece134-operator-scripts\") pod \"78944618-3a02-482a-9965-ef46c0ece134\" (UID: \"78944618-3a02-482a-9965-ef46c0ece134\") " Nov 26 00:44:18 crc kubenswrapper[4766]: I1126 00:44:18.362360 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68b0745b-3ae1-420b-ba0f-3cdbec6c0878-operator-scripts\") pod \"68b0745b-3ae1-420b-ba0f-3cdbec6c0878\" (UID: \"68b0745b-3ae1-420b-ba0f-3cdbec6c0878\") " Nov 26 00:44:18 crc kubenswrapper[4766]: I1126 00:44:18.362540 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9rwb\" (UniqueName: \"kubernetes.io/projected/68b0745b-3ae1-420b-ba0f-3cdbec6c0878-kube-api-access-q9rwb\") pod \"68b0745b-3ae1-420b-ba0f-3cdbec6c0878\" (UID: \"68b0745b-3ae1-420b-ba0f-3cdbec6c0878\") " Nov 26 00:44:18 crc kubenswrapper[4766]: I1126 00:44:18.362888 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68b0745b-3ae1-420b-ba0f-3cdbec6c0878-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "68b0745b-3ae1-420b-ba0f-3cdbec6c0878" (UID: "68b0745b-3ae1-420b-ba0f-3cdbec6c0878"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:18 crc kubenswrapper[4766]: I1126 00:44:18.362894 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78944618-3a02-482a-9965-ef46c0ece134-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "78944618-3a02-482a-9965-ef46c0ece134" (UID: "78944618-3a02-482a-9965-ef46c0ece134"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:18 crc kubenswrapper[4766]: I1126 00:44:18.364147 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78944618-3a02-482a-9965-ef46c0ece134-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:18 crc kubenswrapper[4766]: I1126 00:44:18.364176 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68b0745b-3ae1-420b-ba0f-3cdbec6c0878-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:18 crc kubenswrapper[4766]: I1126 00:44:18.372998 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78944618-3a02-482a-9965-ef46c0ece134-kube-api-access-czb69" (OuterVolumeSpecName: "kube-api-access-czb69") pod "78944618-3a02-482a-9965-ef46c0ece134" (UID: "78944618-3a02-482a-9965-ef46c0ece134"). InnerVolumeSpecName "kube-api-access-czb69". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:18 crc kubenswrapper[4766]: I1126 00:44:18.374062 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68b0745b-3ae1-420b-ba0f-3cdbec6c0878-kube-api-access-q9rwb" (OuterVolumeSpecName: "kube-api-access-q9rwb") pod "68b0745b-3ae1-420b-ba0f-3cdbec6c0878" (UID: "68b0745b-3ae1-420b-ba0f-3cdbec6c0878"). InnerVolumeSpecName "kube-api-access-q9rwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:18 crc kubenswrapper[4766]: I1126 00:44:18.465878 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9rwb\" (UniqueName: \"kubernetes.io/projected/68b0745b-3ae1-420b-ba0f-3cdbec6c0878-kube-api-access-q9rwb\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:18 crc kubenswrapper[4766]: I1126 00:44:18.465917 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czb69\" (UniqueName: \"kubernetes.io/projected/78944618-3a02-482a-9965-ef46c0ece134-kube-api-access-czb69\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.482826 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-etc-swift\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") " pod="openstack/swift-storage-0" Nov 26 00:44:19 crc kubenswrapper[4766]: E1126 00:44:19.483083 4766 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 26 00:44:19 crc kubenswrapper[4766]: E1126 00:44:19.483911 4766 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 26 00:44:19 crc kubenswrapper[4766]: E1126 00:44:19.484010 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-etc-swift podName:a1caa1fb-917a-4267-a947-1194557347eb nodeName:}" failed. No retries permitted until 2025-11-26 00:44:27.483991122 +0000 UTC m=+1248.332761552 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-etc-swift") pod "swift-storage-0" (UID: "a1caa1fb-917a-4267-a947-1194557347eb") : configmap "swift-ring-files" not found Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.813946 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5bc7576dcd-sqbpf_e1ca8391-338e-410b-a7ce-83da6f733c87/console/0.log" Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.814000 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.822184 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-efe0-account-create-update-2kxdm" Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.832587 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3f15-account-create-update-5l657" Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.994600 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-oauth-serving-cert\") pod \"e1ca8391-338e-410b-a7ce-83da6f733c87\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.995052 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e1ca8391-338e-410b-a7ce-83da6f733c87-console-oauth-config\") pod \"e1ca8391-338e-410b-a7ce-83da6f733c87\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.995123 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-trusted-ca-bundle\") pod \"e1ca8391-338e-410b-a7ce-83da6f733c87\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.995155 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e1ca8391-338e-410b-a7ce-83da6f733c87-console-serving-cert\") pod \"e1ca8391-338e-410b-a7ce-83da6f733c87\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.995236 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e-operator-scripts\") pod \"feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e\" (UID: \"feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e\") " Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.995288 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlb9j\" (UniqueName: \"kubernetes.io/projected/feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e-kube-api-access-vlb9j\") pod \"feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e\" (UID: \"feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e\") " Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.995346 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7a0a3f7-ac14-4270-a51e-35738403f130-operator-scripts\") pod \"f7a0a3f7-ac14-4270-a51e-35738403f130\" (UID: \"f7a0a3f7-ac14-4270-a51e-35738403f130\") " Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.995400 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfjbh\" (UniqueName: \"kubernetes.io/projected/f7a0a3f7-ac14-4270-a51e-35738403f130-kube-api-access-zfjbh\") pod \"f7a0a3f7-ac14-4270-a51e-35738403f130\" (UID: \"f7a0a3f7-ac14-4270-a51e-35738403f130\") " Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.995413 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "e1ca8391-338e-410b-a7ce-83da6f733c87" (UID: "e1ca8391-338e-410b-a7ce-83da6f733c87"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.995489 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-console-config\") pod \"e1ca8391-338e-410b-a7ce-83da6f733c87\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.995562 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbh7t\" (UniqueName: \"kubernetes.io/projected/e1ca8391-338e-410b-a7ce-83da6f733c87-kube-api-access-sbh7t\") pod \"e1ca8391-338e-410b-a7ce-83da6f733c87\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.995858 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-service-ca\") pod \"e1ca8391-338e-410b-a7ce-83da6f733c87\" (UID: \"e1ca8391-338e-410b-a7ce-83da6f733c87\") " Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.996792 4766 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.997692 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e" (UID: "feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.998377 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-service-ca" (OuterVolumeSpecName: "service-ca") pod "e1ca8391-338e-410b-a7ce-83da6f733c87" (UID: "e1ca8391-338e-410b-a7ce-83da6f733c87"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.998442 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7a0a3f7-ac14-4270-a51e-35738403f130-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f7a0a3f7-ac14-4270-a51e-35738403f130" (UID: "f7a0a3f7-ac14-4270-a51e-35738403f130"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.998449 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-console-config" (OuterVolumeSpecName: "console-config") pod "e1ca8391-338e-410b-a7ce-83da6f733c87" (UID: "e1ca8391-338e-410b-a7ce-83da6f733c87"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:19 crc kubenswrapper[4766]: I1126 00:44:19.998934 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "e1ca8391-338e-410b-a7ce-83da6f733c87" (UID: "e1ca8391-338e-410b-a7ce-83da6f733c87"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.001447 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1ca8391-338e-410b-a7ce-83da6f733c87-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "e1ca8391-338e-410b-a7ce-83da6f733c87" (UID: "e1ca8391-338e-410b-a7ce-83da6f733c87"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.001910 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e-kube-api-access-vlb9j" (OuterVolumeSpecName: "kube-api-access-vlb9j") pod "feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e" (UID: "feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e"). InnerVolumeSpecName "kube-api-access-vlb9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.002068 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1ca8391-338e-410b-a7ce-83da6f733c87-kube-api-access-sbh7t" (OuterVolumeSpecName: "kube-api-access-sbh7t") pod "e1ca8391-338e-410b-a7ce-83da6f733c87" (UID: "e1ca8391-338e-410b-a7ce-83da6f733c87"). InnerVolumeSpecName "kube-api-access-sbh7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.002973 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1ca8391-338e-410b-a7ce-83da6f733c87-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "e1ca8391-338e-410b-a7ce-83da6f733c87" (UID: "e1ca8391-338e-410b-a7ce-83da6f733c87"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.003145 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7a0a3f7-ac14-4270-a51e-35738403f130-kube-api-access-zfjbh" (OuterVolumeSpecName: "kube-api-access-zfjbh") pod "f7a0a3f7-ac14-4270-a51e-35738403f130" (UID: "f7a0a3f7-ac14-4270-a51e-35738403f130"). InnerVolumeSpecName "kube-api-access-zfjbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.099127 4766 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.099172 4766 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e1ca8391-338e-410b-a7ce-83da6f733c87-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.099182 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.099193 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlb9j\" (UniqueName: \"kubernetes.io/projected/feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e-kube-api-access-vlb9j\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.099203 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7a0a3f7-ac14-4270-a51e-35738403f130-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.099213 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfjbh\" (UniqueName: \"kubernetes.io/projected/f7a0a3f7-ac14-4270-a51e-35738403f130-kube-api-access-zfjbh\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.099221 4766 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-console-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.099229 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbh7t\" (UniqueName: \"kubernetes.io/projected/e1ca8391-338e-410b-a7ce-83da6f733c87-kube-api-access-sbh7t\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.099237 4766 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e1ca8391-338e-410b-a7ce-83da6f733c87-service-ca\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.099244 4766 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e1ca8391-338e-410b-a7ce-83da6f733c87-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.270900 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-3f15-account-create-update-5l657" event={"ID":"f7a0a3f7-ac14-4270-a51e-35738403f130","Type":"ContainerDied","Data":"91af94636c88f0325bf8db354af4eeeae7d3565b440ef184aa504104c3625e7e"} Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.271002 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91af94636c88f0325bf8db354af4eeeae7d3565b440ef184aa504104c3625e7e" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.271072 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3f15-account-create-update-5l657" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.285001 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tmfzn" event={"ID":"6df26511-75fd-4bcb-ab3e-65cde68c9f46","Type":"ContainerStarted","Data":"0ebe2f5c3b38d2b4eec49b248117db093222176db9107fee4e0eb22e1ffb11a5"} Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.288417 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-efe0-account-create-update-2kxdm" event={"ID":"feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e","Type":"ContainerDied","Data":"fc1eefbcc53e7ed906ad89c9f81f3bb2f3db4b95193339dd3e317ee7e004c2e2"} Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.288459 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc1eefbcc53e7ed906ad89c9f81f3bb2f3db4b95193339dd3e317ee7e004c2e2" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.288517 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-efe0-account-create-update-2kxdm" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.296272 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5bc7576dcd-sqbpf_e1ca8391-338e-410b-a7ce-83da6f733c87/console/0.log" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.296346 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5bc7576dcd-sqbpf" event={"ID":"e1ca8391-338e-410b-a7ce-83da6f733c87","Type":"ContainerDied","Data":"b0ed3c9a2e9e49fb3d62aaee91a75aabfe7d60df162036f05febc5d2766942f4"} Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.296379 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5bc7576dcd-sqbpf" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.296393 4766 scope.go:117] "RemoveContainer" containerID="ece9fdfedc40fbd881d5f91cfbe1c0398a1f91bdecc08911f99a8d8e6be1e380" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.323704 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-tmfzn" podStartSLOduration=5.1415925 podStartE2EDuration="9.32368168s" podCreationTimestamp="2025-11-26 00:44:11 +0000 UTC" firstStartedPulling="2025-11-26 00:44:15.416322302 +0000 UTC m=+1236.265092732" lastFinishedPulling="2025-11-26 00:44:19.598411482 +0000 UTC m=+1240.447181912" observedRunningTime="2025-11-26 00:44:20.31447728 +0000 UTC m=+1241.163247730" watchObservedRunningTime="2025-11-26 00:44:20.32368168 +0000 UTC m=+1241.172452110" Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.342805 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-5bc7576dcd-sqbpf"] Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.351545 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-5bc7576dcd-sqbpf"] Nov 26 00:44:20 crc kubenswrapper[4766]: I1126 00:44:20.661536 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 26 00:44:21 crc kubenswrapper[4766]: I1126 00:44:21.836597 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1ca8391-338e-410b-a7ce-83da6f733c87" path="/var/lib/kubelet/pods/e1ca8391-338e-410b-a7ce-83da6f733c87/volumes" Nov 26 00:44:22 crc kubenswrapper[4766]: I1126 00:44:22.318243 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6e23ce44-c896-41b1-b427-1bc2db9955ca","Type":"ContainerStarted","Data":"72396edd9978790d62df23bcdc03d5954680ea48b0bba190e1175e08db198b14"} Nov 26 00:44:22 crc kubenswrapper[4766]: I1126 00:44:22.349676 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=12.805938535 podStartE2EDuration="52.349641909s" podCreationTimestamp="2025-11-26 00:43:30 +0000 UTC" firstStartedPulling="2025-11-26 00:43:42.362499854 +0000 UTC m=+1203.211270284" lastFinishedPulling="2025-11-26 00:44:21.906203228 +0000 UTC m=+1242.754973658" observedRunningTime="2025-11-26 00:44:22.344341046 +0000 UTC m=+1243.193111496" watchObservedRunningTime="2025-11-26 00:44:22.349641909 +0000 UTC m=+1243.198412339" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.766418 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-vh2cq"] Nov 26 00:44:23 crc kubenswrapper[4766]: E1126 00:44:23.766865 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2ec3192-d2ac-4f92-abd2-4c20548dceb3" containerName="mariadb-database-create" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.766881 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2ec3192-d2ac-4f92-abd2-4c20548dceb3" containerName="mariadb-database-create" Nov 26 00:44:23 crc kubenswrapper[4766]: E1126 00:44:23.766893 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b0745b-3ae1-420b-ba0f-3cdbec6c0878" containerName="mariadb-database-create" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.766901 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b0745b-3ae1-420b-ba0f-3cdbec6c0878" containerName="mariadb-database-create" Nov 26 00:44:23 crc kubenswrapper[4766]: E1126 00:44:23.766914 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34ebe8c7-f89d-478a-ba71-4160910955c6" containerName="mariadb-database-create" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.766925 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="34ebe8c7-f89d-478a-ba71-4160910955c6" containerName="mariadb-database-create" Nov 26 00:44:23 crc kubenswrapper[4766]: E1126 00:44:23.766945 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1ca8391-338e-410b-a7ce-83da6f733c87" containerName="console" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.766952 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1ca8391-338e-410b-a7ce-83da6f733c87" containerName="console" Nov 26 00:44:23 crc kubenswrapper[4766]: E1126 00:44:23.766968 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcf23dba-a1e2-41be-b115-777bda2d8dc4" containerName="mariadb-account-create-update" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.766976 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcf23dba-a1e2-41be-b115-777bda2d8dc4" containerName="mariadb-account-create-update" Nov 26 00:44:23 crc kubenswrapper[4766]: E1126 00:44:23.766988 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="095a8e06-fc50-4b1c-b27d-9efb621c523a" containerName="mariadb-account-create-update" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.766996 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="095a8e06-fc50-4b1c-b27d-9efb621c523a" containerName="mariadb-account-create-update" Nov 26 00:44:23 crc kubenswrapper[4766]: E1126 00:44:23.767012 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e" containerName="mariadb-account-create-update" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.767021 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e" containerName="mariadb-account-create-update" Nov 26 00:44:23 crc kubenswrapper[4766]: E1126 00:44:23.767034 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7a0a3f7-ac14-4270-a51e-35738403f130" containerName="mariadb-account-create-update" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.767042 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7a0a3f7-ac14-4270-a51e-35738403f130" containerName="mariadb-account-create-update" Nov 26 00:44:23 crc kubenswrapper[4766]: E1126 00:44:23.767062 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78944618-3a02-482a-9965-ef46c0ece134" containerName="mariadb-database-create" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.767070 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="78944618-3a02-482a-9965-ef46c0ece134" containerName="mariadb-database-create" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.767260 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2ec3192-d2ac-4f92-abd2-4c20548dceb3" containerName="mariadb-database-create" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.767595 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="78944618-3a02-482a-9965-ef46c0ece134" containerName="mariadb-database-create" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.767607 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e" containerName="mariadb-account-create-update" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.767619 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="34ebe8c7-f89d-478a-ba71-4160910955c6" containerName="mariadb-database-create" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.767628 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="095a8e06-fc50-4b1c-b27d-9efb621c523a" containerName="mariadb-account-create-update" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.767637 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1ca8391-338e-410b-a7ce-83da6f733c87" containerName="console" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.767673 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcf23dba-a1e2-41be-b115-777bda2d8dc4" containerName="mariadb-account-create-update" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.767695 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7a0a3f7-ac14-4270-a51e-35738403f130" containerName="mariadb-account-create-update" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.767705 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b0745b-3ae1-420b-ba0f-3cdbec6c0878" containerName="mariadb-database-create" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.768377 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-vh2cq" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.770380 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-7t54l" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.770726 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.802772 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-vh2cq"] Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.939007 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04d8e581-a471-4aea-8c66-016ae86cd9e8-config-data\") pod \"glance-db-sync-vh2cq\" (UID: \"04d8e581-a471-4aea-8c66-016ae86cd9e8\") " pod="openstack/glance-db-sync-vh2cq" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.939076 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d8e581-a471-4aea-8c66-016ae86cd9e8-combined-ca-bundle\") pod \"glance-db-sync-vh2cq\" (UID: \"04d8e581-a471-4aea-8c66-016ae86cd9e8\") " pod="openstack/glance-db-sync-vh2cq" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.939129 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm9fx\" (UniqueName: \"kubernetes.io/projected/04d8e581-a471-4aea-8c66-016ae86cd9e8-kube-api-access-lm9fx\") pod \"glance-db-sync-vh2cq\" (UID: \"04d8e581-a471-4aea-8c66-016ae86cd9e8\") " pod="openstack/glance-db-sync-vh2cq" Nov 26 00:44:23 crc kubenswrapper[4766]: I1126 00:44:23.939160 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04d8e581-a471-4aea-8c66-016ae86cd9e8-db-sync-config-data\") pod \"glance-db-sync-vh2cq\" (UID: \"04d8e581-a471-4aea-8c66-016ae86cd9e8\") " pod="openstack/glance-db-sync-vh2cq" Nov 26 00:44:24 crc kubenswrapper[4766]: I1126 00:44:24.040602 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm9fx\" (UniqueName: \"kubernetes.io/projected/04d8e581-a471-4aea-8c66-016ae86cd9e8-kube-api-access-lm9fx\") pod \"glance-db-sync-vh2cq\" (UID: \"04d8e581-a471-4aea-8c66-016ae86cd9e8\") " pod="openstack/glance-db-sync-vh2cq" Nov 26 00:44:24 crc kubenswrapper[4766]: I1126 00:44:24.040902 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04d8e581-a471-4aea-8c66-016ae86cd9e8-db-sync-config-data\") pod \"glance-db-sync-vh2cq\" (UID: \"04d8e581-a471-4aea-8c66-016ae86cd9e8\") " pod="openstack/glance-db-sync-vh2cq" Nov 26 00:44:24 crc kubenswrapper[4766]: I1126 00:44:24.041474 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04d8e581-a471-4aea-8c66-016ae86cd9e8-config-data\") pod \"glance-db-sync-vh2cq\" (UID: \"04d8e581-a471-4aea-8c66-016ae86cd9e8\") " pod="openstack/glance-db-sync-vh2cq" Nov 26 00:44:24 crc kubenswrapper[4766]: I1126 00:44:24.041629 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d8e581-a471-4aea-8c66-016ae86cd9e8-combined-ca-bundle\") pod \"glance-db-sync-vh2cq\" (UID: \"04d8e581-a471-4aea-8c66-016ae86cd9e8\") " pod="openstack/glance-db-sync-vh2cq" Nov 26 00:44:24 crc kubenswrapper[4766]: I1126 00:44:24.046720 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04d8e581-a471-4aea-8c66-016ae86cd9e8-db-sync-config-data\") pod \"glance-db-sync-vh2cq\" (UID: \"04d8e581-a471-4aea-8c66-016ae86cd9e8\") " pod="openstack/glance-db-sync-vh2cq" Nov 26 00:44:24 crc kubenswrapper[4766]: I1126 00:44:24.046878 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04d8e581-a471-4aea-8c66-016ae86cd9e8-config-data\") pod \"glance-db-sync-vh2cq\" (UID: \"04d8e581-a471-4aea-8c66-016ae86cd9e8\") " pod="openstack/glance-db-sync-vh2cq" Nov 26 00:44:24 crc kubenswrapper[4766]: I1126 00:44:24.053520 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d8e581-a471-4aea-8c66-016ae86cd9e8-combined-ca-bundle\") pod \"glance-db-sync-vh2cq\" (UID: \"04d8e581-a471-4aea-8c66-016ae86cd9e8\") " pod="openstack/glance-db-sync-vh2cq" Nov 26 00:44:24 crc kubenswrapper[4766]: I1126 00:44:24.059568 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm9fx\" (UniqueName: \"kubernetes.io/projected/04d8e581-a471-4aea-8c66-016ae86cd9e8-kube-api-access-lm9fx\") pod \"glance-db-sync-vh2cq\" (UID: \"04d8e581-a471-4aea-8c66-016ae86cd9e8\") " pod="openstack/glance-db-sync-vh2cq" Nov 26 00:44:24 crc kubenswrapper[4766]: I1126 00:44:24.088854 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-vh2cq" Nov 26 00:44:24 crc kubenswrapper[4766]: I1126 00:44:24.926360 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-vh2cq"] Nov 26 00:44:24 crc kubenswrapper[4766]: W1126 00:44:24.933484 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04d8e581_a471_4aea_8c66_016ae86cd9e8.slice/crio-78ac1a2f5bf8b3fe68101bad7155b91bbdf992785d8bea5f3c01e94f27f15b24 WatchSource:0}: Error finding container 78ac1a2f5bf8b3fe68101bad7155b91bbdf992785d8bea5f3c01e94f27f15b24: Status 404 returned error can't find the container with id 78ac1a2f5bf8b3fe68101bad7155b91bbdf992785d8bea5f3c01e94f27f15b24 Nov 26 00:44:25 crc kubenswrapper[4766]: I1126 00:44:25.348159 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-vh2cq" event={"ID":"04d8e581-a471-4aea-8c66-016ae86cd9e8","Type":"ContainerStarted","Data":"78ac1a2f5bf8b3fe68101bad7155b91bbdf992785d8bea5f3c01e94f27f15b24"} Nov 26 00:44:25 crc kubenswrapper[4766]: I1126 00:44:25.678747 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" Nov 26 00:44:25 crc kubenswrapper[4766]: I1126 00:44:25.725079 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-7mdpn"] Nov 26 00:44:25 crc kubenswrapper[4766]: I1126 00:44:25.725337 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-7mdpn" podUID="de34e733-406f-4d3c-8c30-e36c41fdf3a1" containerName="dnsmasq-dns" containerID="cri-o://a7d6d18d696bd29e80eb1e79e59a4744e8f12f5145baf077cd2db32620687e1a" gracePeriod=10 Nov 26 00:44:25 crc kubenswrapper[4766]: I1126 00:44:25.759669 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-v2bnt"] Nov 26 00:44:25 crc kubenswrapper[4766]: I1126 00:44:25.766826 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-v2bnt" Nov 26 00:44:25 crc kubenswrapper[4766]: I1126 00:44:25.774049 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-v2bnt"] Nov 26 00:44:25 crc kubenswrapper[4766]: I1126 00:44:25.875762 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-686kt\" (UniqueName: \"kubernetes.io/projected/c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0-kube-api-access-686kt\") pod \"mysqld-exporter-openstack-cell1-db-create-v2bnt\" (UID: \"c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-v2bnt" Nov 26 00:44:25 crc kubenswrapper[4766]: I1126 00:44:25.876012 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-v2bnt\" (UID: \"c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-v2bnt" Nov 26 00:44:25 crc kubenswrapper[4766]: I1126 00:44:25.955132 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-def0-account-create-update-rmc7d"] Nov 26 00:44:25 crc kubenswrapper[4766]: I1126 00:44:25.958454 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-def0-account-create-update-rmc7d" Nov 26 00:44:25 crc kubenswrapper[4766]: I1126 00:44:25.961265 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-cell1-db-secret" Nov 26 00:44:25 crc kubenswrapper[4766]: I1126 00:44:25.965361 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-def0-account-create-update-rmc7d"] Nov 26 00:44:25 crc kubenswrapper[4766]: I1126 00:44:25.979575 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-686kt\" (UniqueName: \"kubernetes.io/projected/c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0-kube-api-access-686kt\") pod \"mysqld-exporter-openstack-cell1-db-create-v2bnt\" (UID: \"c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-v2bnt" Nov 26 00:44:25 crc kubenswrapper[4766]: I1126 00:44:25.979617 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-v2bnt\" (UID: \"c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-v2bnt" Nov 26 00:44:25 crc kubenswrapper[4766]: I1126 00:44:25.979714 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/618ca9d6-4ffb-4c75-9b16-b471471f91c3-operator-scripts\") pod \"mysqld-exporter-def0-account-create-update-rmc7d\" (UID: \"618ca9d6-4ffb-4c75-9b16-b471471f91c3\") " pod="openstack/mysqld-exporter-def0-account-create-update-rmc7d" Nov 26 00:44:25 crc kubenswrapper[4766]: I1126 00:44:25.979740 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v878l\" (UniqueName: \"kubernetes.io/projected/618ca9d6-4ffb-4c75-9b16-b471471f91c3-kube-api-access-v878l\") pod \"mysqld-exporter-def0-account-create-update-rmc7d\" (UID: \"618ca9d6-4ffb-4c75-9b16-b471471f91c3\") " pod="openstack/mysqld-exporter-def0-account-create-update-rmc7d" Nov 26 00:44:25 crc kubenswrapper[4766]: I1126 00:44:25.980805 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-v2bnt\" (UID: \"c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-v2bnt" Nov 26 00:44:26 crc kubenswrapper[4766]: I1126 00:44:26.008332 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-686kt\" (UniqueName: \"kubernetes.io/projected/c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0-kube-api-access-686kt\") pod \"mysqld-exporter-openstack-cell1-db-create-v2bnt\" (UID: \"c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-v2bnt" Nov 26 00:44:26 crc kubenswrapper[4766]: I1126 00:44:26.081353 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/618ca9d6-4ffb-4c75-9b16-b471471f91c3-operator-scripts\") pod \"mysqld-exporter-def0-account-create-update-rmc7d\" (UID: \"618ca9d6-4ffb-4c75-9b16-b471471f91c3\") " pod="openstack/mysqld-exporter-def0-account-create-update-rmc7d" Nov 26 00:44:26 crc kubenswrapper[4766]: I1126 00:44:26.081412 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v878l\" (UniqueName: \"kubernetes.io/projected/618ca9d6-4ffb-4c75-9b16-b471471f91c3-kube-api-access-v878l\") pod \"mysqld-exporter-def0-account-create-update-rmc7d\" (UID: \"618ca9d6-4ffb-4c75-9b16-b471471f91c3\") " pod="openstack/mysqld-exporter-def0-account-create-update-rmc7d" Nov 26 00:44:26 crc kubenswrapper[4766]: I1126 00:44:26.082013 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/618ca9d6-4ffb-4c75-9b16-b471471f91c3-operator-scripts\") pod \"mysqld-exporter-def0-account-create-update-rmc7d\" (UID: \"618ca9d6-4ffb-4c75-9b16-b471471f91c3\") " pod="openstack/mysqld-exporter-def0-account-create-update-rmc7d" Nov 26 00:44:26 crc kubenswrapper[4766]: I1126 00:44:26.089085 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-v2bnt" Nov 26 00:44:26 crc kubenswrapper[4766]: I1126 00:44:26.099009 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v878l\" (UniqueName: \"kubernetes.io/projected/618ca9d6-4ffb-4c75-9b16-b471471f91c3-kube-api-access-v878l\") pod \"mysqld-exporter-def0-account-create-update-rmc7d\" (UID: \"618ca9d6-4ffb-4c75-9b16-b471471f91c3\") " pod="openstack/mysqld-exporter-def0-account-create-update-rmc7d" Nov 26 00:44:26 crc kubenswrapper[4766]: I1126 00:44:26.291441 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-def0-account-create-update-rmc7d" Nov 26 00:44:26 crc kubenswrapper[4766]: I1126 00:44:26.384140 4766 generic.go:334] "Generic (PLEG): container finished" podID="de34e733-406f-4d3c-8c30-e36c41fdf3a1" containerID="a7d6d18d696bd29e80eb1e79e59a4744e8f12f5145baf077cd2db32620687e1a" exitCode=0 Nov 26 00:44:26 crc kubenswrapper[4766]: I1126 00:44:26.384195 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-7mdpn" event={"ID":"de34e733-406f-4d3c-8c30-e36c41fdf3a1","Type":"ContainerDied","Data":"a7d6d18d696bd29e80eb1e79e59a4744e8f12f5145baf077cd2db32620687e1a"} Nov 26 00:44:26 crc kubenswrapper[4766]: I1126 00:44:26.683799 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:26 crc kubenswrapper[4766]: I1126 00:44:26.749420 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-v2bnt"] Nov 26 00:44:26 crc kubenswrapper[4766]: W1126 00:44:26.782897 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9bc49f8_8ab0_4f1a_b32c_677ab40cc3f0.slice/crio-a554a9ebe17e1e7b8ee2a9d314608204bfc979fa69aa9de8e3e1b50e41059b2b WatchSource:0}: Error finding container a554a9ebe17e1e7b8ee2a9d314608204bfc979fa69aa9de8e3e1b50e41059b2b: Status 404 returned error can't find the container with id a554a9ebe17e1e7b8ee2a9d314608204bfc979fa69aa9de8e3e1b50e41059b2b Nov 26 00:44:26 crc kubenswrapper[4766]: I1126 00:44:26.949756 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-def0-account-create-update-rmc7d"] Nov 26 00:44:27 crc kubenswrapper[4766]: W1126 00:44:27.025086 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod618ca9d6_4ffb_4c75_9b16_b471471f91c3.slice/crio-1ee9d9c18ba9372c76822c85bfa692ff689cdba9c625d72c1981414146b94e1b WatchSource:0}: Error finding container 1ee9d9c18ba9372c76822c85bfa692ff689cdba9c625d72c1981414146b94e1b: Status 404 returned error can't find the container with id 1ee9d9c18ba9372c76822c85bfa692ff689cdba9c625d72c1981414146b94e1b Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.065392 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-7mdpn" Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.204010 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-config\") pod \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\" (UID: \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\") " Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.204093 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9twth\" (UniqueName: \"kubernetes.io/projected/de34e733-406f-4d3c-8c30-e36c41fdf3a1-kube-api-access-9twth\") pod \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\" (UID: \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\") " Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.204139 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-dns-svc\") pod \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\" (UID: \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\") " Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.204238 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-ovsdbserver-nb\") pod \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\" (UID: \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\") " Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.204318 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-ovsdbserver-sb\") pod \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\" (UID: \"de34e733-406f-4d3c-8c30-e36c41fdf3a1\") " Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.211046 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de34e733-406f-4d3c-8c30-e36c41fdf3a1-kube-api-access-9twth" (OuterVolumeSpecName: "kube-api-access-9twth") pod "de34e733-406f-4d3c-8c30-e36c41fdf3a1" (UID: "de34e733-406f-4d3c-8c30-e36c41fdf3a1"). InnerVolumeSpecName "kube-api-access-9twth". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.280465 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "de34e733-406f-4d3c-8c30-e36c41fdf3a1" (UID: "de34e733-406f-4d3c-8c30-e36c41fdf3a1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.283803 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-config" (OuterVolumeSpecName: "config") pod "de34e733-406f-4d3c-8c30-e36c41fdf3a1" (UID: "de34e733-406f-4d3c-8c30-e36c41fdf3a1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.289915 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "de34e733-406f-4d3c-8c30-e36c41fdf3a1" (UID: "de34e733-406f-4d3c-8c30-e36c41fdf3a1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.306962 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.307011 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.307027 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9twth\" (UniqueName: \"kubernetes.io/projected/de34e733-406f-4d3c-8c30-e36c41fdf3a1-kube-api-access-9twth\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.307042 4766 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.310488 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "de34e733-406f-4d3c-8c30-e36c41fdf3a1" (UID: "de34e733-406f-4d3c-8c30-e36c41fdf3a1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.397536 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-7mdpn" event={"ID":"de34e733-406f-4d3c-8c30-e36c41fdf3a1","Type":"ContainerDied","Data":"67fc6f75f029111fb0ed46f71b4741e95e39b3b9e798d51e9e4c8224b2a2c4af"} Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.397587 4766 scope.go:117] "RemoveContainer" containerID="a7d6d18d696bd29e80eb1e79e59a4744e8f12f5145baf077cd2db32620687e1a" Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.397719 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-7mdpn" Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.402392 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-v2bnt" event={"ID":"c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0","Type":"ContainerStarted","Data":"192c1c9406b9ae618d7e353c7cf5e6b63badfda270f1bdfb72db674c960f3664"} Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.402514 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-v2bnt" event={"ID":"c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0","Type":"ContainerStarted","Data":"a554a9ebe17e1e7b8ee2a9d314608204bfc979fa69aa9de8e3e1b50e41059b2b"} Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.404518 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-def0-account-create-update-rmc7d" event={"ID":"618ca9d6-4ffb-4c75-9b16-b471471f91c3","Type":"ContainerStarted","Data":"2582ba9afa41abc0ae5d3e47291ea56d3d29f9cc2862d627fa98cbf0f68f93b9"} Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.404644 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-def0-account-create-update-rmc7d" event={"ID":"618ca9d6-4ffb-4c75-9b16-b471471f91c3","Type":"ContainerStarted","Data":"1ee9d9c18ba9372c76822c85bfa692ff689cdba9c625d72c1981414146b94e1b"} Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.408384 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de34e733-406f-4d3c-8c30-e36c41fdf3a1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.422392 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-openstack-cell1-db-create-v2bnt" podStartSLOduration=2.422341153 podStartE2EDuration="2.422341153s" podCreationTimestamp="2025-11-26 00:44:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:44:27.415863491 +0000 UTC m=+1248.264633921" watchObservedRunningTime="2025-11-26 00:44:27.422341153 +0000 UTC m=+1248.271111583" Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.437430 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-def0-account-create-update-rmc7d" podStartSLOduration=2.437403941 podStartE2EDuration="2.437403941s" podCreationTimestamp="2025-11-26 00:44:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:44:27.433462952 +0000 UTC m=+1248.282233382" watchObservedRunningTime="2025-11-26 00:44:27.437403941 +0000 UTC m=+1248.286174361" Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.467356 4766 scope.go:117] "RemoveContainer" containerID="88feb284c7e95b21c8d7ffe541f9c8bc6c46f6e259cbbe36ad1dbc891ae8723e" Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.487521 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-7mdpn"] Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.497772 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-7mdpn"] Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.510263 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-etc-swift\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") " pod="openstack/swift-storage-0" Nov 26 00:44:27 crc kubenswrapper[4766]: E1126 00:44:27.510874 4766 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 26 00:44:27 crc kubenswrapper[4766]: E1126 00:44:27.510909 4766 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 26 00:44:27 crc kubenswrapper[4766]: E1126 00:44:27.510965 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-etc-swift podName:a1caa1fb-917a-4267-a947-1194557347eb nodeName:}" failed. No retries permitted until 2025-11-26 00:44:43.510948055 +0000 UTC m=+1264.359718485 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-etc-swift") pod "swift-storage-0" (UID: "a1caa1fb-917a-4267-a947-1194557347eb") : configmap "swift-ring-files" not found Nov 26 00:44:27 crc kubenswrapper[4766]: I1126 00:44:27.841113 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de34e733-406f-4d3c-8c30-e36c41fdf3a1" path="/var/lib/kubelet/pods/de34e733-406f-4d3c-8c30-e36c41fdf3a1/volumes" Nov 26 00:44:28 crc kubenswrapper[4766]: I1126 00:44:28.418079 4766 generic.go:334] "Generic (PLEG): container finished" podID="6df26511-75fd-4bcb-ab3e-65cde68c9f46" containerID="0ebe2f5c3b38d2b4eec49b248117db093222176db9107fee4e0eb22e1ffb11a5" exitCode=0 Nov 26 00:44:28 crc kubenswrapper[4766]: I1126 00:44:28.418804 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tmfzn" event={"ID":"6df26511-75fd-4bcb-ab3e-65cde68c9f46","Type":"ContainerDied","Data":"0ebe2f5c3b38d2b4eec49b248117db093222176db9107fee4e0eb22e1ffb11a5"} Nov 26 00:44:28 crc kubenswrapper[4766]: I1126 00:44:28.422088 4766 generic.go:334] "Generic (PLEG): container finished" podID="c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0" containerID="192c1c9406b9ae618d7e353c7cf5e6b63badfda270f1bdfb72db674c960f3664" exitCode=0 Nov 26 00:44:28 crc kubenswrapper[4766]: I1126 00:44:28.422170 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-v2bnt" event={"ID":"c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0","Type":"ContainerDied","Data":"192c1c9406b9ae618d7e353c7cf5e6b63badfda270f1bdfb72db674c960f3664"} Nov 26 00:44:28 crc kubenswrapper[4766]: I1126 00:44:28.424060 4766 generic.go:334] "Generic (PLEG): container finished" podID="618ca9d6-4ffb-4c75-9b16-b471471f91c3" containerID="2582ba9afa41abc0ae5d3e47291ea56d3d29f9cc2862d627fa98cbf0f68f93b9" exitCode=0 Nov 26 00:44:28 crc kubenswrapper[4766]: I1126 00:44:28.424104 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-def0-account-create-update-rmc7d" event={"ID":"618ca9d6-4ffb-4c75-9b16-b471471f91c3","Type":"ContainerDied","Data":"2582ba9afa41abc0ae5d3e47291ea56d3d29f9cc2862d627fa98cbf0f68f93b9"} Nov 26 00:44:28 crc kubenswrapper[4766]: I1126 00:44:28.850570 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-p42tr" podUID="22168318-cef8-4363-8b3e-3042e82037ea" containerName="ovn-controller" probeResult="failure" output=< Nov 26 00:44:28 crc kubenswrapper[4766]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 26 00:44:28 crc kubenswrapper[4766]: > Nov 26 00:44:29 crc kubenswrapper[4766]: I1126 00:44:29.436063 4766 generic.go:334] "Generic (PLEG): container finished" podID="8a944def-677c-4bdc-a7df-2586d0c7937f" containerID="ce25f81214edc08620b9866c07d4fdf518d951e03f1f623a3eeb20ca690ef230" exitCode=0 Nov 26 00:44:29 crc kubenswrapper[4766]: I1126 00:44:29.436156 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8a944def-677c-4bdc-a7df-2586d0c7937f","Type":"ContainerDied","Data":"ce25f81214edc08620b9866c07d4fdf518d951e03f1f623a3eeb20ca690ef230"} Nov 26 00:44:29 crc kubenswrapper[4766]: I1126 00:44:29.912819 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-v2bnt" Nov 26 00:44:29 crc kubenswrapper[4766]: I1126 00:44:29.968705 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-686kt\" (UniqueName: \"kubernetes.io/projected/c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0-kube-api-access-686kt\") pod \"c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0\" (UID: \"c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0\") " Nov 26 00:44:29 crc kubenswrapper[4766]: I1126 00:44:29.968850 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0-operator-scripts\") pod \"c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0\" (UID: \"c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0\") " Nov 26 00:44:29 crc kubenswrapper[4766]: I1126 00:44:29.969693 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0" (UID: "c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:29 crc kubenswrapper[4766]: I1126 00:44:29.975344 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0-kube-api-access-686kt" (OuterVolumeSpecName: "kube-api-access-686kt") pod "c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0" (UID: "c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0"). InnerVolumeSpecName "kube-api-access-686kt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.013407 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.032825 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-def0-account-create-update-rmc7d" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.070336 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6df26511-75fd-4bcb-ab3e-65cde68c9f46-dispersionconf\") pod \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.070453 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4p6r\" (UniqueName: \"kubernetes.io/projected/6df26511-75fd-4bcb-ab3e-65cde68c9f46-kube-api-access-c4p6r\") pod \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.070524 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/618ca9d6-4ffb-4c75-9b16-b471471f91c3-operator-scripts\") pod \"618ca9d6-4ffb-4c75-9b16-b471471f91c3\" (UID: \"618ca9d6-4ffb-4c75-9b16-b471471f91c3\") " Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.070573 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6df26511-75fd-4bcb-ab3e-65cde68c9f46-ring-data-devices\") pod \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.070602 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6df26511-75fd-4bcb-ab3e-65cde68c9f46-scripts\") pod \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.070644 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df26511-75fd-4bcb-ab3e-65cde68c9f46-combined-ca-bundle\") pod \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.070686 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6df26511-75fd-4bcb-ab3e-65cde68c9f46-etc-swift\") pod \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.070752 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6df26511-75fd-4bcb-ab3e-65cde68c9f46-swiftconf\") pod \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\" (UID: \"6df26511-75fd-4bcb-ab3e-65cde68c9f46\") " Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.070821 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v878l\" (UniqueName: \"kubernetes.io/projected/618ca9d6-4ffb-4c75-9b16-b471471f91c3-kube-api-access-v878l\") pod \"618ca9d6-4ffb-4c75-9b16-b471471f91c3\" (UID: \"618ca9d6-4ffb-4c75-9b16-b471471f91c3\") " Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.071162 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6df26511-75fd-4bcb-ab3e-65cde68c9f46-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "6df26511-75fd-4bcb-ab3e-65cde68c9f46" (UID: "6df26511-75fd-4bcb-ab3e-65cde68c9f46"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.071345 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-686kt\" (UniqueName: \"kubernetes.io/projected/c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0-kube-api-access-686kt\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.071367 4766 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6df26511-75fd-4bcb-ab3e-65cde68c9f46-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.071380 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.071821 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/618ca9d6-4ffb-4c75-9b16-b471471f91c3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "618ca9d6-4ffb-4c75-9b16-b471471f91c3" (UID: "618ca9d6-4ffb-4c75-9b16-b471471f91c3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.074580 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/618ca9d6-4ffb-4c75-9b16-b471471f91c3-kube-api-access-v878l" (OuterVolumeSpecName: "kube-api-access-v878l") pod "618ca9d6-4ffb-4c75-9b16-b471471f91c3" (UID: "618ca9d6-4ffb-4c75-9b16-b471471f91c3"). InnerVolumeSpecName "kube-api-access-v878l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.075352 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6df26511-75fd-4bcb-ab3e-65cde68c9f46-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "6df26511-75fd-4bcb-ab3e-65cde68c9f46" (UID: "6df26511-75fd-4bcb-ab3e-65cde68c9f46"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.079762 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6df26511-75fd-4bcb-ab3e-65cde68c9f46-kube-api-access-c4p6r" (OuterVolumeSpecName: "kube-api-access-c4p6r") pod "6df26511-75fd-4bcb-ab3e-65cde68c9f46" (UID: "6df26511-75fd-4bcb-ab3e-65cde68c9f46"). InnerVolumeSpecName "kube-api-access-c4p6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.081916 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6df26511-75fd-4bcb-ab3e-65cde68c9f46-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "6df26511-75fd-4bcb-ab3e-65cde68c9f46" (UID: "6df26511-75fd-4bcb-ab3e-65cde68c9f46"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.123920 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6df26511-75fd-4bcb-ab3e-65cde68c9f46-scripts" (OuterVolumeSpecName: "scripts") pod "6df26511-75fd-4bcb-ab3e-65cde68c9f46" (UID: "6df26511-75fd-4bcb-ab3e-65cde68c9f46"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.135841 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6df26511-75fd-4bcb-ab3e-65cde68c9f46-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "6df26511-75fd-4bcb-ab3e-65cde68c9f46" (UID: "6df26511-75fd-4bcb-ab3e-65cde68c9f46"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.137269 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6df26511-75fd-4bcb-ab3e-65cde68c9f46-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6df26511-75fd-4bcb-ab3e-65cde68c9f46" (UID: "6df26511-75fd-4bcb-ab3e-65cde68c9f46"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.174343 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/618ca9d6-4ffb-4c75-9b16-b471471f91c3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.174381 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6df26511-75fd-4bcb-ab3e-65cde68c9f46-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.174392 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df26511-75fd-4bcb-ab3e-65cde68c9f46-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.174405 4766 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6df26511-75fd-4bcb-ab3e-65cde68c9f46-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.174415 4766 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6df26511-75fd-4bcb-ab3e-65cde68c9f46-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.174428 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v878l\" (UniqueName: \"kubernetes.io/projected/618ca9d6-4ffb-4c75-9b16-b471471f91c3-kube-api-access-v878l\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.174441 4766 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6df26511-75fd-4bcb-ab3e-65cde68c9f46-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.174452 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4p6r\" (UniqueName: \"kubernetes.io/projected/6df26511-75fd-4bcb-ab3e-65cde68c9f46-kube-api-access-c4p6r\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.466986 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tmfzn" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.467016 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tmfzn" event={"ID":"6df26511-75fd-4bcb-ab3e-65cde68c9f46","Type":"ContainerDied","Data":"1eb5080cd70fb96e1a9e1dd44e5f78f41567d622808c0e060db6f81e1ce1db9f"} Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.467058 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1eb5080cd70fb96e1a9e1dd44e5f78f41567d622808c0e060db6f81e1ce1db9f" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.474797 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-v2bnt" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.475814 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-v2bnt" event={"ID":"c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0","Type":"ContainerDied","Data":"a554a9ebe17e1e7b8ee2a9d314608204bfc979fa69aa9de8e3e1b50e41059b2b"} Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.475844 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a554a9ebe17e1e7b8ee2a9d314608204bfc979fa69aa9de8e3e1b50e41059b2b" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.484988 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-def0-account-create-update-rmc7d" event={"ID":"618ca9d6-4ffb-4c75-9b16-b471471f91c3","Type":"ContainerDied","Data":"1ee9d9c18ba9372c76822c85bfa692ff689cdba9c625d72c1981414146b94e1b"} Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.485043 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ee9d9c18ba9372c76822c85bfa692ff689cdba9c625d72c1981414146b94e1b" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.485143 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-def0-account-create-update-rmc7d" Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.494268 4766 generic.go:334] "Generic (PLEG): container finished" podID="0527e2fe-678e-4712-ac14-3f2e1f3f0fe0" containerID="84d5f79a02963d431e4b2b22e8c479a860d1676c7ccd73a598d1c7c0726230be" exitCode=0 Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.494369 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0","Type":"ContainerDied","Data":"84d5f79a02963d431e4b2b22e8c479a860d1676c7ccd73a598d1c7c0726230be"} Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.500494 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8a944def-677c-4bdc-a7df-2586d0c7937f","Type":"ContainerStarted","Data":"8a31ae18469d427edc200ce8bf40458cbdf1027220aa0c2d71e7e146412710b9"} Nov 26 00:44:30 crc kubenswrapper[4766]: I1126 00:44:30.500931 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.145387 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=62.465879618 podStartE2EDuration="1m8.14536636s" podCreationTimestamp="2025-11-26 00:43:23 +0000 UTC" firstStartedPulling="2025-11-26 00:43:41.659330209 +0000 UTC m=+1202.508100639" lastFinishedPulling="2025-11-26 00:43:47.338816951 +0000 UTC m=+1208.187587381" observedRunningTime="2025-11-26 00:44:30.562830571 +0000 UTC m=+1251.411601001" watchObservedRunningTime="2025-11-26 00:44:31.14536636 +0000 UTC m=+1251.994136791" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.151695 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Nov 26 00:44:31 crc kubenswrapper[4766]: E1126 00:44:31.152163 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="618ca9d6-4ffb-4c75-9b16-b471471f91c3" containerName="mariadb-account-create-update" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.152184 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="618ca9d6-4ffb-4c75-9b16-b471471f91c3" containerName="mariadb-account-create-update" Nov 26 00:44:31 crc kubenswrapper[4766]: E1126 00:44:31.152203 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6df26511-75fd-4bcb-ab3e-65cde68c9f46" containerName="swift-ring-rebalance" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.152211 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6df26511-75fd-4bcb-ab3e-65cde68c9f46" containerName="swift-ring-rebalance" Nov 26 00:44:31 crc kubenswrapper[4766]: E1126 00:44:31.152230 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0" containerName="mariadb-database-create" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.152238 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0" containerName="mariadb-database-create" Nov 26 00:44:31 crc kubenswrapper[4766]: E1126 00:44:31.152251 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de34e733-406f-4d3c-8c30-e36c41fdf3a1" containerName="init" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.152260 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="de34e733-406f-4d3c-8c30-e36c41fdf3a1" containerName="init" Nov 26 00:44:31 crc kubenswrapper[4766]: E1126 00:44:31.152274 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de34e733-406f-4d3c-8c30-e36c41fdf3a1" containerName="dnsmasq-dns" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.152283 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="de34e733-406f-4d3c-8c30-e36c41fdf3a1" containerName="dnsmasq-dns" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.152495 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0" containerName="mariadb-database-create" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.152519 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="de34e733-406f-4d3c-8c30-e36c41fdf3a1" containerName="dnsmasq-dns" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.152542 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="6df26511-75fd-4bcb-ab3e-65cde68c9f46" containerName="swift-ring-rebalance" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.152553 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="618ca9d6-4ffb-4c75-9b16-b471471f91c3" containerName="mariadb-account-create-update" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.155429 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.157472 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.170125 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.308547 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e35fc291-f4f0-41fe-93d3-dc2683811a39-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"e35fc291-f4f0-41fe-93d3-dc2683811a39\") " pod="openstack/mysqld-exporter-0" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.308688 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e35fc291-f4f0-41fe-93d3-dc2683811a39-config-data\") pod \"mysqld-exporter-0\" (UID: \"e35fc291-f4f0-41fe-93d3-dc2683811a39\") " pod="openstack/mysqld-exporter-0" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.308744 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw5zf\" (UniqueName: \"kubernetes.io/projected/e35fc291-f4f0-41fe-93d3-dc2683811a39-kube-api-access-tw5zf\") pod \"mysqld-exporter-0\" (UID: \"e35fc291-f4f0-41fe-93d3-dc2683811a39\") " pod="openstack/mysqld-exporter-0" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.410142 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e35fc291-f4f0-41fe-93d3-dc2683811a39-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"e35fc291-f4f0-41fe-93d3-dc2683811a39\") " pod="openstack/mysqld-exporter-0" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.410235 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e35fc291-f4f0-41fe-93d3-dc2683811a39-config-data\") pod \"mysqld-exporter-0\" (UID: \"e35fc291-f4f0-41fe-93d3-dc2683811a39\") " pod="openstack/mysqld-exporter-0" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.410286 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw5zf\" (UniqueName: \"kubernetes.io/projected/e35fc291-f4f0-41fe-93d3-dc2683811a39-kube-api-access-tw5zf\") pod \"mysqld-exporter-0\" (UID: \"e35fc291-f4f0-41fe-93d3-dc2683811a39\") " pod="openstack/mysqld-exporter-0" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.415421 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e35fc291-f4f0-41fe-93d3-dc2683811a39-config-data\") pod \"mysqld-exporter-0\" (UID: \"e35fc291-f4f0-41fe-93d3-dc2683811a39\") " pod="openstack/mysqld-exporter-0" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.415604 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e35fc291-f4f0-41fe-93d3-dc2683811a39-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"e35fc291-f4f0-41fe-93d3-dc2683811a39\") " pod="openstack/mysqld-exporter-0" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.430941 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw5zf\" (UniqueName: \"kubernetes.io/projected/e35fc291-f4f0-41fe-93d3-dc2683811a39-kube-api-access-tw5zf\") pod \"mysqld-exporter-0\" (UID: \"e35fc291-f4f0-41fe-93d3-dc2683811a39\") " pod="openstack/mysqld-exporter-0" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.480948 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.517744 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0","Type":"ContainerStarted","Data":"2406d6b7cfd725312646e13f515b83a477e306195d4a6f5b4240ac62783379e0"} Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.517950 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.548287 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=60.606022466 podStartE2EDuration="1m8.548267385s" podCreationTimestamp="2025-11-26 00:43:23 +0000 UTC" firstStartedPulling="2025-11-26 00:43:39.886267163 +0000 UTC m=+1200.735037593" lastFinishedPulling="2025-11-26 00:43:47.828512082 +0000 UTC m=+1208.677282512" observedRunningTime="2025-11-26 00:44:31.542477889 +0000 UTC m=+1252.391248319" watchObservedRunningTime="2025-11-26 00:44:31.548267385 +0000 UTC m=+1252.397037815" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.684094 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:31 crc kubenswrapper[4766]: I1126 00:44:31.686928 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:32 crc kubenswrapper[4766]: I1126 00:44:32.030691 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 26 00:44:32 crc kubenswrapper[4766]: I1126 00:44:32.538627 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"e35fc291-f4f0-41fe-93d3-dc2683811a39","Type":"ContainerStarted","Data":"68dcaf6bfe9d46079cdcd180975134d8c3c33d493ab36a127d3a072ba1933397"} Nov 26 00:44:32 crc kubenswrapper[4766]: I1126 00:44:32.540124 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:33 crc kubenswrapper[4766]: I1126 00:44:33.782250 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-p42tr" podUID="22168318-cef8-4363-8b3e-3042e82037ea" containerName="ovn-controller" probeResult="failure" output=< Nov 26 00:44:33 crc kubenswrapper[4766]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 26 00:44:33 crc kubenswrapper[4766]: > Nov 26 00:44:33 crc kubenswrapper[4766]: I1126 00:44:33.880745 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:44:33 crc kubenswrapper[4766]: I1126 00:44:33.893055 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-6ghtc" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.098400 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-p42tr-config-tcctj"] Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.099873 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.102031 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.114151 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-p42tr-config-tcctj"] Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.170813 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-scripts\") pod \"ovn-controller-p42tr-config-tcctj\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.170906 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-var-run\") pod \"ovn-controller-p42tr-config-tcctj\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.170930 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9wzk\" (UniqueName: \"kubernetes.io/projected/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-kube-api-access-s9wzk\") pod \"ovn-controller-p42tr-config-tcctj\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.171064 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-var-log-ovn\") pod \"ovn-controller-p42tr-config-tcctj\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.171112 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-var-run-ovn\") pod \"ovn-controller-p42tr-config-tcctj\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.171135 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-additional-scripts\") pod \"ovn-controller-p42tr-config-tcctj\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.272622 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-var-run-ovn\") pod \"ovn-controller-p42tr-config-tcctj\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.272921 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-var-run-ovn\") pod \"ovn-controller-p42tr-config-tcctj\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.272953 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-additional-scripts\") pod \"ovn-controller-p42tr-config-tcctj\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.273108 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-scripts\") pod \"ovn-controller-p42tr-config-tcctj\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.273168 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-var-run\") pod \"ovn-controller-p42tr-config-tcctj\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.273187 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9wzk\" (UniqueName: \"kubernetes.io/projected/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-kube-api-access-s9wzk\") pod \"ovn-controller-p42tr-config-tcctj\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.273284 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-var-log-ovn\") pod \"ovn-controller-p42tr-config-tcctj\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.273532 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-var-run\") pod \"ovn-controller-p42tr-config-tcctj\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.274249 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-var-log-ovn\") pod \"ovn-controller-p42tr-config-tcctj\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.274356 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-additional-scripts\") pod \"ovn-controller-p42tr-config-tcctj\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.276351 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-scripts\") pod \"ovn-controller-p42tr-config-tcctj\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.308306 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9wzk\" (UniqueName: \"kubernetes.io/projected/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-kube-api-access-s9wzk\") pod \"ovn-controller-p42tr-config-tcctj\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.472892 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.974328 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.974599 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="6e23ce44-c896-41b1-b427-1bc2db9955ca" containerName="prometheus" containerID="cri-o://b3476a207d9b96473d9efbe51664aa5329ef1bee2f07196acdf8785cbf6848bd" gracePeriod=600 Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.974979 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="6e23ce44-c896-41b1-b427-1bc2db9955ca" containerName="thanos-sidecar" containerID="cri-o://72396edd9978790d62df23bcdc03d5954680ea48b0bba190e1175e08db198b14" gracePeriod=600 Nov 26 00:44:34 crc kubenswrapper[4766]: I1126 00:44:34.975030 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="6e23ce44-c896-41b1-b427-1bc2db9955ca" containerName="config-reloader" containerID="cri-o://62285b131a0ff5824a4f3305369e761cddd7daf66d555ea5e28446a413301a1a" gracePeriod=600 Nov 26 00:44:35 crc kubenswrapper[4766]: I1126 00:44:35.569684 4766 generic.go:334] "Generic (PLEG): container finished" podID="6e23ce44-c896-41b1-b427-1bc2db9955ca" containerID="72396edd9978790d62df23bcdc03d5954680ea48b0bba190e1175e08db198b14" exitCode=0 Nov 26 00:44:35 crc kubenswrapper[4766]: I1126 00:44:35.570012 4766 generic.go:334] "Generic (PLEG): container finished" podID="6e23ce44-c896-41b1-b427-1bc2db9955ca" containerID="62285b131a0ff5824a4f3305369e761cddd7daf66d555ea5e28446a413301a1a" exitCode=0 Nov 26 00:44:35 crc kubenswrapper[4766]: I1126 00:44:35.570026 4766 generic.go:334] "Generic (PLEG): container finished" podID="6e23ce44-c896-41b1-b427-1bc2db9955ca" containerID="b3476a207d9b96473d9efbe51664aa5329ef1bee2f07196acdf8785cbf6848bd" exitCode=0 Nov 26 00:44:35 crc kubenswrapper[4766]: I1126 00:44:35.569691 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6e23ce44-c896-41b1-b427-1bc2db9955ca","Type":"ContainerDied","Data":"72396edd9978790d62df23bcdc03d5954680ea48b0bba190e1175e08db198b14"} Nov 26 00:44:35 crc kubenswrapper[4766]: I1126 00:44:35.570068 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6e23ce44-c896-41b1-b427-1bc2db9955ca","Type":"ContainerDied","Data":"62285b131a0ff5824a4f3305369e761cddd7daf66d555ea5e28446a413301a1a"} Nov 26 00:44:35 crc kubenswrapper[4766]: I1126 00:44:35.570086 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6e23ce44-c896-41b1-b427-1bc2db9955ca","Type":"ContainerDied","Data":"b3476a207d9b96473d9efbe51664aa5329ef1bee2f07196acdf8785cbf6848bd"} Nov 26 00:44:36 crc kubenswrapper[4766]: I1126 00:44:36.684083 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="6e23ce44-c896-41b1-b427-1bc2db9955ca" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.128:9090/-/ready\": dial tcp 10.217.0.128:9090: connect: connection refused" Nov 26 00:44:38 crc kubenswrapper[4766]: I1126 00:44:38.775184 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-p42tr" podUID="22168318-cef8-4363-8b3e-3042e82037ea" containerName="ovn-controller" probeResult="failure" output=< Nov 26 00:44:38 crc kubenswrapper[4766]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 26 00:44:38 crc kubenswrapper[4766]: > Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.498600 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.634766 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6e23ce44-c896-41b1-b427-1bc2db9955ca-config\") pod \"6e23ce44-c896-41b1-b427-1bc2db9955ca\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.634828 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6e23ce44-c896-41b1-b427-1bc2db9955ca-tls-assets\") pod \"6e23ce44-c896-41b1-b427-1bc2db9955ca\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.634939 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7hjr\" (UniqueName: \"kubernetes.io/projected/6e23ce44-c896-41b1-b427-1bc2db9955ca-kube-api-access-g7hjr\") pod \"6e23ce44-c896-41b1-b427-1bc2db9955ca\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.634989 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6e23ce44-c896-41b1-b427-1bc2db9955ca-config-out\") pod \"6e23ce44-c896-41b1-b427-1bc2db9955ca\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.635113 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf\") pod \"6e23ce44-c896-41b1-b427-1bc2db9955ca\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.635164 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6e23ce44-c896-41b1-b427-1bc2db9955ca-web-config\") pod \"6e23ce44-c896-41b1-b427-1bc2db9955ca\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.635232 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6e23ce44-c896-41b1-b427-1bc2db9955ca-prometheus-metric-storage-rulefiles-0\") pod \"6e23ce44-c896-41b1-b427-1bc2db9955ca\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.635269 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6e23ce44-c896-41b1-b427-1bc2db9955ca-thanos-prometheus-http-client-file\") pod \"6e23ce44-c896-41b1-b427-1bc2db9955ca\" (UID: \"6e23ce44-c896-41b1-b427-1bc2db9955ca\") " Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.662451 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e23ce44-c896-41b1-b427-1bc2db9955ca-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "6e23ce44-c896-41b1-b427-1bc2db9955ca" (UID: "6e23ce44-c896-41b1-b427-1bc2db9955ca"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.662558 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e23ce44-c896-41b1-b427-1bc2db9955ca-config" (OuterVolumeSpecName: "config") pod "6e23ce44-c896-41b1-b427-1bc2db9955ca" (UID: "6e23ce44-c896-41b1-b427-1bc2db9955ca"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.667327 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e23ce44-c896-41b1-b427-1bc2db9955ca-config-out" (OuterVolumeSpecName: "config-out") pod "6e23ce44-c896-41b1-b427-1bc2db9955ca" (UID: "6e23ce44-c896-41b1-b427-1bc2db9955ca"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.667423 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e23ce44-c896-41b1-b427-1bc2db9955ca-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "6e23ce44-c896-41b1-b427-1bc2db9955ca" (UID: "6e23ce44-c896-41b1-b427-1bc2db9955ca"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.669937 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e23ce44-c896-41b1-b427-1bc2db9955ca-kube-api-access-g7hjr" (OuterVolumeSpecName: "kube-api-access-g7hjr") pod "6e23ce44-c896-41b1-b427-1bc2db9955ca" (UID: "6e23ce44-c896-41b1-b427-1bc2db9955ca"). InnerVolumeSpecName "kube-api-access-g7hjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.670823 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e23ce44-c896-41b1-b427-1bc2db9955ca-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "6e23ce44-c896-41b1-b427-1bc2db9955ca" (UID: "6e23ce44-c896-41b1-b427-1bc2db9955ca"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.751627 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-p42tr-config-tcctj"] Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.761729 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6e23ce44-c896-41b1-b427-1bc2db9955ca-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.761782 4766 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6e23ce44-c896-41b1-b427-1bc2db9955ca-tls-assets\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.761792 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7hjr\" (UniqueName: \"kubernetes.io/projected/6e23ce44-c896-41b1-b427-1bc2db9955ca-kube-api-access-g7hjr\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.761801 4766 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6e23ce44-c896-41b1-b427-1bc2db9955ca-config-out\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.761811 4766 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6e23ce44-c896-41b1-b427-1bc2db9955ca-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.761819 4766 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6e23ce44-c896-41b1-b427-1bc2db9955ca-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.767979 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e23ce44-c896-41b1-b427-1bc2db9955ca-web-config" (OuterVolumeSpecName: "web-config") pod "6e23ce44-c896-41b1-b427-1bc2db9955ca" (UID: "6e23ce44-c896-41b1-b427-1bc2db9955ca"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.774083 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6e23ce44-c896-41b1-b427-1bc2db9955ca","Type":"ContainerDied","Data":"3fe19dd68ce379294bd9201b944d402e76e55120cf5f90ad8c7531eef13219c4"} Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.774140 4766 scope.go:117] "RemoveContainer" containerID="72396edd9978790d62df23bcdc03d5954680ea48b0bba190e1175e08db198b14" Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.774300 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.854867 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "6e23ce44-c896-41b1-b427-1bc2db9955ca" (UID: "6e23ce44-c896-41b1-b427-1bc2db9955ca"). InnerVolumeSpecName "pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.863192 4766 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf\") on node \"crc\" " Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.863226 4766 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6e23ce44-c896-41b1-b427-1bc2db9955ca-web-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.887219 4766 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.887375 4766 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf") on node "crc" Nov 26 00:44:40 crc kubenswrapper[4766]: I1126 00:44:40.965225 4766 reconciler_common.go:293] "Volume detached for volume \"pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.173196 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.200086 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.224922 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 26 00:44:41 crc kubenswrapper[4766]: E1126 00:44:41.225453 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e23ce44-c896-41b1-b427-1bc2db9955ca" containerName="config-reloader" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.225475 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e23ce44-c896-41b1-b427-1bc2db9955ca" containerName="config-reloader" Nov 26 00:44:41 crc kubenswrapper[4766]: E1126 00:44:41.225501 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e23ce44-c896-41b1-b427-1bc2db9955ca" containerName="prometheus" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.225509 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e23ce44-c896-41b1-b427-1bc2db9955ca" containerName="prometheus" Nov 26 00:44:41 crc kubenswrapper[4766]: E1126 00:44:41.225540 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e23ce44-c896-41b1-b427-1bc2db9955ca" containerName="init-config-reloader" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.225549 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e23ce44-c896-41b1-b427-1bc2db9955ca" containerName="init-config-reloader" Nov 26 00:44:41 crc kubenswrapper[4766]: E1126 00:44:41.225563 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e23ce44-c896-41b1-b427-1bc2db9955ca" containerName="thanos-sidecar" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.225570 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e23ce44-c896-41b1-b427-1bc2db9955ca" containerName="thanos-sidecar" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.225811 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e23ce44-c896-41b1-b427-1bc2db9955ca" containerName="config-reloader" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.225847 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e23ce44-c896-41b1-b427-1bc2db9955ca" containerName="thanos-sidecar" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.225871 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e23ce44-c896-41b1-b427-1bc2db9955ca" containerName="prometheus" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.228083 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.232002 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.233752 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.234452 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.235026 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.235502 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-k8x7n" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.235527 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.242947 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.250713 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.372059 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0c404609-2299-4fe5-8c43-da5614df251e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.372114 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0c404609-2299-4fe5-8c43-da5614df251e-config\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.372153 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0c404609-2299-4fe5-8c43-da5614df251e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.372209 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0c404609-2299-4fe5-8c43-da5614df251e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.372247 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c404609-2299-4fe5-8c43-da5614df251e-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.372272 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.372327 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/0c404609-2299-4fe5-8c43-da5614df251e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.372354 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgvkr\" (UniqueName: \"kubernetes.io/projected/0c404609-2299-4fe5-8c43-da5614df251e-kube-api-access-sgvkr\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.372371 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0c404609-2299-4fe5-8c43-da5614df251e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.372391 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0c404609-2299-4fe5-8c43-da5614df251e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.372433 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/0c404609-2299-4fe5-8c43-da5614df251e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.474030 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0c404609-2299-4fe5-8c43-da5614df251e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.474136 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0c404609-2299-4fe5-8c43-da5614df251e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.474189 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c404609-2299-4fe5-8c43-da5614df251e-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.474223 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.474266 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/0c404609-2299-4fe5-8c43-da5614df251e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.474309 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgvkr\" (UniqueName: \"kubernetes.io/projected/0c404609-2299-4fe5-8c43-da5614df251e-kube-api-access-sgvkr\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.474339 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0c404609-2299-4fe5-8c43-da5614df251e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.474364 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0c404609-2299-4fe5-8c43-da5614df251e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.474423 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/0c404609-2299-4fe5-8c43-da5614df251e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.474453 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0c404609-2299-4fe5-8c43-da5614df251e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.474487 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0c404609-2299-4fe5-8c43-da5614df251e-config\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.477111 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/0c404609-2299-4fe5-8c43-da5614df251e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.481639 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0c404609-2299-4fe5-8c43-da5614df251e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.481713 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c404609-2299-4fe5-8c43-da5614df251e-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.483673 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/0c404609-2299-4fe5-8c43-da5614df251e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.483994 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0c404609-2299-4fe5-8c43-da5614df251e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.485120 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0c404609-2299-4fe5-8c43-da5614df251e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.486171 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0c404609-2299-4fe5-8c43-da5614df251e-config\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.487232 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0c404609-2299-4fe5-8c43-da5614df251e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.487553 4766 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.487588 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/149ea7c544c1e9392b1f012edc83d0a3042f9a6ea5986794ddffaab68b75c51a/globalmount\"" pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.498338 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0c404609-2299-4fe5-8c43-da5614df251e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.504150 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgvkr\" (UniqueName: \"kubernetes.io/projected/0c404609-2299-4fe5-8c43-da5614df251e-kube-api-access-sgvkr\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.553870 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7e3ce641-1fbc-453d-a77c-ef6031041bdf\") pod \"prometheus-metric-storage-0\" (UID: \"0c404609-2299-4fe5-8c43-da5614df251e\") " pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.652400 4766 scope.go:117] "RemoveContainer" containerID="62285b131a0ff5824a4f3305369e761cddd7daf66d555ea5e28446a413301a1a" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.725914 4766 scope.go:117] "RemoveContainer" containerID="b3476a207d9b96473d9efbe51664aa5329ef1bee2f07196acdf8785cbf6848bd" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.782597 4766 scope.go:117] "RemoveContainer" containerID="a70fc172683be2ed9e0b00864643e55a3d8ede537b6d0359d28f439575913435" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.791677 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-vh2cq" event={"ID":"04d8e581-a471-4aea-8c66-016ae86cd9e8","Type":"ContainerStarted","Data":"28f98fa01a48d0f4bded085fc78451fad87337d0a0eecc968095e0c60f6edffe"} Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.793128 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p42tr-config-tcctj" event={"ID":"be0ded9d-d02c-4ac3-919b-d9d6e94866d9","Type":"ContainerStarted","Data":"8158aeef73b2b877e391a716178952897613f896443a870604fd1c1ff23cef1e"} Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.817232 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-vh2cq" podStartSLOduration=3.589707142 podStartE2EDuration="18.817207973s" podCreationTimestamp="2025-11-26 00:44:23 +0000 UTC" firstStartedPulling="2025-11-26 00:44:24.935851246 +0000 UTC m=+1245.784621676" lastFinishedPulling="2025-11-26 00:44:40.163352077 +0000 UTC m=+1261.012122507" observedRunningTime="2025-11-26 00:44:41.807049538 +0000 UTC m=+1262.655819968" watchObservedRunningTime="2025-11-26 00:44:41.817207973 +0000 UTC m=+1262.665978423" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.840409 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e23ce44-c896-41b1-b427-1bc2db9955ca" path="/var/lib/kubelet/pods/6e23ce44-c896-41b1-b427-1bc2db9955ca/volumes" Nov 26 00:44:41 crc kubenswrapper[4766]: I1126 00:44:41.848321 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 26 00:44:42 crc kubenswrapper[4766]: I1126 00:44:42.375022 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 26 00:44:42 crc kubenswrapper[4766]: W1126 00:44:42.395418 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c404609_2299_4fe5_8c43_da5614df251e.slice/crio-42762178ee238aeea6e42249adb313bfb5c83a40ca683cff2afd485497c57944 WatchSource:0}: Error finding container 42762178ee238aeea6e42249adb313bfb5c83a40ca683cff2afd485497c57944: Status 404 returned error can't find the container with id 42762178ee238aeea6e42249adb313bfb5c83a40ca683cff2afd485497c57944 Nov 26 00:44:42 crc kubenswrapper[4766]: I1126 00:44:42.802154 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0c404609-2299-4fe5-8c43-da5614df251e","Type":"ContainerStarted","Data":"42762178ee238aeea6e42249adb313bfb5c83a40ca683cff2afd485497c57944"} Nov 26 00:44:42 crc kubenswrapper[4766]: I1126 00:44:42.803883 4766 generic.go:334] "Generic (PLEG): container finished" podID="be0ded9d-d02c-4ac3-919b-d9d6e94866d9" containerID="3230fd677c493d3befdf9f609024a491fa15783fbb4f924c1484a0015f258849" exitCode=0 Nov 26 00:44:42 crc kubenswrapper[4766]: I1126 00:44:42.803973 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p42tr-config-tcctj" event={"ID":"be0ded9d-d02c-4ac3-919b-d9d6e94866d9","Type":"ContainerDied","Data":"3230fd677c493d3befdf9f609024a491fa15783fbb4f924c1484a0015f258849"} Nov 26 00:44:42 crc kubenswrapper[4766]: I1126 00:44:42.806643 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"e35fc291-f4f0-41fe-93d3-dc2683811a39","Type":"ContainerStarted","Data":"b52dd15229e1e324f3c932df79226b0271e73d92dd9491f1bcb45846272b430c"} Nov 26 00:44:42 crc kubenswrapper[4766]: I1126 00:44:42.838495 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=2.137667244 podStartE2EDuration="11.838472474s" podCreationTimestamp="2025-11-26 00:44:31 +0000 UTC" firstStartedPulling="2025-11-26 00:44:32.036906109 +0000 UTC m=+1252.885676539" lastFinishedPulling="2025-11-26 00:44:41.737711339 +0000 UTC m=+1262.586481769" observedRunningTime="2025-11-26 00:44:42.831486709 +0000 UTC m=+1263.680257149" watchObservedRunningTime="2025-11-26 00:44:42.838472474 +0000 UTC m=+1263.687242914" Nov 26 00:44:43 crc kubenswrapper[4766]: I1126 00:44:43.513634 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-etc-swift\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") " pod="openstack/swift-storage-0" Nov 26 00:44:43 crc kubenswrapper[4766]: I1126 00:44:43.533406 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a1caa1fb-917a-4267-a947-1194557347eb-etc-swift\") pod \"swift-storage-0\" (UID: \"a1caa1fb-917a-4267-a947-1194557347eb\") " pod="openstack/swift-storage-0" Nov 26 00:44:43 crc kubenswrapper[4766]: I1126 00:44:43.661731 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 26 00:44:43 crc kubenswrapper[4766]: I1126 00:44:43.779229 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-p42tr" Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.209437 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 26 00:44:44 crc kubenswrapper[4766]: W1126 00:44:44.409818 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1caa1fb_917a_4267_a947_1194557347eb.slice/crio-a180c1f7ccf53a923d5c14f764df3de23d7170b406a182724a1fde73971073e2 WatchSource:0}: Error finding container a180c1f7ccf53a923d5c14f764df3de23d7170b406a182724a1fde73971073e2: Status 404 returned error can't find the container with id a180c1f7ccf53a923d5c14f764df3de23d7170b406a182724a1fde73971073e2 Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.637866 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.701631 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.735022 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-scripts\") pod \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.735069 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-additional-scripts\") pod \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.735143 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9wzk\" (UniqueName: \"kubernetes.io/projected/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-kube-api-access-s9wzk\") pod \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.735193 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-var-run-ovn\") pod \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.735253 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-var-log-ovn\") pod \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.735320 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-var-run\") pod \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\" (UID: \"be0ded9d-d02c-4ac3-919b-d9d6e94866d9\") " Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.735378 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "be0ded9d-d02c-4ac3-919b-d9d6e94866d9" (UID: "be0ded9d-d02c-4ac3-919b-d9d6e94866d9"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.735438 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "be0ded9d-d02c-4ac3-919b-d9d6e94866d9" (UID: "be0ded9d-d02c-4ac3-919b-d9d6e94866d9"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.735493 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-var-run" (OuterVolumeSpecName: "var-run") pod "be0ded9d-d02c-4ac3-919b-d9d6e94866d9" (UID: "be0ded9d-d02c-4ac3-919b-d9d6e94866d9"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.735954 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "be0ded9d-d02c-4ac3-919b-d9d6e94866d9" (UID: "be0ded9d-d02c-4ac3-919b-d9d6e94866d9"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.736747 4766 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-var-run\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.736773 4766 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.736785 4766 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.736798 4766 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.738483 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-scripts" (OuterVolumeSpecName: "scripts") pod "be0ded9d-d02c-4ac3-919b-d9d6e94866d9" (UID: "be0ded9d-d02c-4ac3-919b-d9d6e94866d9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.760275 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-kube-api-access-s9wzk" (OuterVolumeSpecName: "kube-api-access-s9wzk") pod "be0ded9d-d02c-4ac3-919b-d9d6e94866d9" (UID: "be0ded9d-d02c-4ac3-919b-d9d6e94866d9"). InnerVolumeSpecName "kube-api-access-s9wzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.840943 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.840986 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9wzk\" (UniqueName: \"kubernetes.io/projected/be0ded9d-d02c-4ac3-919b-d9d6e94866d9-kube-api-access-s9wzk\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.848856 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a1caa1fb-917a-4267-a947-1194557347eb","Type":"ContainerStarted","Data":"a180c1f7ccf53a923d5c14f764df3de23d7170b406a182724a1fde73971073e2"} Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.868665 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p42tr-config-tcctj" event={"ID":"be0ded9d-d02c-4ac3-919b-d9d6e94866d9","Type":"ContainerDied","Data":"8158aeef73b2b877e391a716178952897613f896443a870604fd1c1ff23cef1e"} Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.868707 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8158aeef73b2b877e391a716178952897613f896443a870604fd1c1ff23cef1e" Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.868894 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p42tr-config-tcctj" Nov 26 00:44:44 crc kubenswrapper[4766]: I1126 00:44:44.954852 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.064691 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-7hkqm"] Nov 26 00:44:45 crc kubenswrapper[4766]: E1126 00:44:45.065146 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be0ded9d-d02c-4ac3-919b-d9d6e94866d9" containerName="ovn-config" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.065162 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="be0ded9d-d02c-4ac3-919b-d9d6e94866d9" containerName="ovn-config" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.065355 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="be0ded9d-d02c-4ac3-919b-d9d6e94866d9" containerName="ovn-config" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.066024 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7hkqm" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.075559 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-7hkqm"] Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.085104 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-6008-account-create-update-dbw86"] Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.086643 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6008-account-create-update-dbw86" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.091961 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.099586 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6008-account-create-update-dbw86"] Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.149742 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec-operator-scripts\") pod \"barbican-6008-account-create-update-dbw86\" (UID: \"ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec\") " pod="openstack/barbican-6008-account-create-update-dbw86" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.149853 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47lrj\" (UniqueName: \"kubernetes.io/projected/ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec-kube-api-access-47lrj\") pod \"barbican-6008-account-create-update-dbw86\" (UID: \"ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec\") " pod="openstack/barbican-6008-account-create-update-dbw86" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.149884 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b80a95b-04b4-4b49-a799-cb6469c7b198-operator-scripts\") pod \"barbican-db-create-7hkqm\" (UID: \"0b80a95b-04b4-4b49-a799-cb6469c7b198\") " pod="openstack/barbican-db-create-7hkqm" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.149946 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql6t5\" (UniqueName: \"kubernetes.io/projected/0b80a95b-04b4-4b49-a799-cb6469c7b198-kube-api-access-ql6t5\") pod \"barbican-db-create-7hkqm\" (UID: \"0b80a95b-04b4-4b49-a799-cb6469c7b198\") " pod="openstack/barbican-db-create-7hkqm" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.200826 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-9e39-account-create-update-9s4pv"] Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.202409 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-9e39-account-create-update-9s4pv" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.213101 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.253568 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-6wtrt"] Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.255797 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6wtrt" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.258066 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92fd9e2d-9152-46a4-8686-ef9fe6d0b790-operator-scripts\") pod \"heat-9e39-account-create-update-9s4pv\" (UID: \"92fd9e2d-9152-46a4-8686-ef9fe6d0b790\") " pod="openstack/heat-9e39-account-create-update-9s4pv" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.258427 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th79p\" (UniqueName: \"kubernetes.io/projected/92fd9e2d-9152-46a4-8686-ef9fe6d0b790-kube-api-access-th79p\") pod \"heat-9e39-account-create-update-9s4pv\" (UID: \"92fd9e2d-9152-46a4-8686-ef9fe6d0b790\") " pod="openstack/heat-9e39-account-create-update-9s4pv" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.258914 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec-operator-scripts\") pod \"barbican-6008-account-create-update-dbw86\" (UID: \"ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec\") " pod="openstack/barbican-6008-account-create-update-dbw86" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.259189 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47lrj\" (UniqueName: \"kubernetes.io/projected/ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec-kube-api-access-47lrj\") pod \"barbican-6008-account-create-update-dbw86\" (UID: \"ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec\") " pod="openstack/barbican-6008-account-create-update-dbw86" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.259397 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b80a95b-04b4-4b49-a799-cb6469c7b198-operator-scripts\") pod \"barbican-db-create-7hkqm\" (UID: \"0b80a95b-04b4-4b49-a799-cb6469c7b198\") " pod="openstack/barbican-db-create-7hkqm" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.261491 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec-operator-scripts\") pod \"barbican-6008-account-create-update-dbw86\" (UID: \"ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec\") " pod="openstack/barbican-6008-account-create-update-dbw86" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.262494 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql6t5\" (UniqueName: \"kubernetes.io/projected/0b80a95b-04b4-4b49-a799-cb6469c7b198-kube-api-access-ql6t5\") pod \"barbican-db-create-7hkqm\" (UID: \"0b80a95b-04b4-4b49-a799-cb6469c7b198\") " pod="openstack/barbican-db-create-7hkqm" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.263325 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b80a95b-04b4-4b49-a799-cb6469c7b198-operator-scripts\") pod \"barbican-db-create-7hkqm\" (UID: \"0b80a95b-04b4-4b49-a799-cb6469c7b198\") " pod="openstack/barbican-db-create-7hkqm" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.307405 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-9e39-account-create-update-9s4pv"] Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.315361 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-6wtrt"] Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.318803 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47lrj\" (UniqueName: \"kubernetes.io/projected/ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec-kube-api-access-47lrj\") pod \"barbican-6008-account-create-update-dbw86\" (UID: \"ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec\") " pod="openstack/barbican-6008-account-create-update-dbw86" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.318866 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql6t5\" (UniqueName: \"kubernetes.io/projected/0b80a95b-04b4-4b49-a799-cb6469c7b198-kube-api-access-ql6t5\") pod \"barbican-db-create-7hkqm\" (UID: \"0b80a95b-04b4-4b49-a799-cb6469c7b198\") " pod="openstack/barbican-db-create-7hkqm" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.364555 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-79smp"] Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.365849 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-79smp" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.367427 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjtft\" (UniqueName: \"kubernetes.io/projected/4873ca17-8a36-4c66-b86b-5a6784c7e4f2-kube-api-access-gjtft\") pod \"cinder-db-create-6wtrt\" (UID: \"4873ca17-8a36-4c66-b86b-5a6784c7e4f2\") " pod="openstack/cinder-db-create-6wtrt" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.367462 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4873ca17-8a36-4c66-b86b-5a6784c7e4f2-operator-scripts\") pod \"cinder-db-create-6wtrt\" (UID: \"4873ca17-8a36-4c66-b86b-5a6784c7e4f2\") " pod="openstack/cinder-db-create-6wtrt" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.367494 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92fd9e2d-9152-46a4-8686-ef9fe6d0b790-operator-scripts\") pod \"heat-9e39-account-create-update-9s4pv\" (UID: \"92fd9e2d-9152-46a4-8686-ef9fe6d0b790\") " pod="openstack/heat-9e39-account-create-update-9s4pv" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.367539 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th79p\" (UniqueName: \"kubernetes.io/projected/92fd9e2d-9152-46a4-8686-ef9fe6d0b790-kube-api-access-th79p\") pod \"heat-9e39-account-create-update-9s4pv\" (UID: \"92fd9e2d-9152-46a4-8686-ef9fe6d0b790\") " pod="openstack/heat-9e39-account-create-update-9s4pv" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.370190 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92fd9e2d-9152-46a4-8686-ef9fe6d0b790-operator-scripts\") pod \"heat-9e39-account-create-update-9s4pv\" (UID: \"92fd9e2d-9152-46a4-8686-ef9fe6d0b790\") " pod="openstack/heat-9e39-account-create-update-9s4pv" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.380259 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-5dd7-account-create-update-q2fxk"] Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.382039 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5dd7-account-create-update-q2fxk" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.383527 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7hkqm" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.387053 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.397098 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th79p\" (UniqueName: \"kubernetes.io/projected/92fd9e2d-9152-46a4-8686-ef9fe6d0b790-kube-api-access-th79p\") pod \"heat-9e39-account-create-update-9s4pv\" (UID: \"92fd9e2d-9152-46a4-8686-ef9fe6d0b790\") " pod="openstack/heat-9e39-account-create-update-9s4pv" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.398457 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-79smp"] Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.418929 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6008-account-create-update-dbw86" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.437315 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5dd7-account-create-update-q2fxk"] Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.469578 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ced743c6-6d0e-439e-9d16-46ed28450a25-operator-scripts\") pod \"heat-db-create-79smp\" (UID: \"ced743c6-6d0e-439e-9d16-46ed28450a25\") " pod="openstack/heat-db-create-79smp" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.469774 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5szfg\" (UniqueName: \"kubernetes.io/projected/ced743c6-6d0e-439e-9d16-46ed28450a25-kube-api-access-5szfg\") pod \"heat-db-create-79smp\" (UID: \"ced743c6-6d0e-439e-9d16-46ed28450a25\") " pod="openstack/heat-db-create-79smp" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.469798 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjtft\" (UniqueName: \"kubernetes.io/projected/4873ca17-8a36-4c66-b86b-5a6784c7e4f2-kube-api-access-gjtft\") pod \"cinder-db-create-6wtrt\" (UID: \"4873ca17-8a36-4c66-b86b-5a6784c7e4f2\") " pod="openstack/cinder-db-create-6wtrt" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.469822 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4873ca17-8a36-4c66-b86b-5a6784c7e4f2-operator-scripts\") pod \"cinder-db-create-6wtrt\" (UID: \"4873ca17-8a36-4c66-b86b-5a6784c7e4f2\") " pod="openstack/cinder-db-create-6wtrt" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.469866 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27rdp\" (UniqueName: \"kubernetes.io/projected/adde2fcf-e073-4936-a856-16bd660658a5-kube-api-access-27rdp\") pod \"cinder-5dd7-account-create-update-q2fxk\" (UID: \"adde2fcf-e073-4936-a856-16bd660658a5\") " pod="openstack/cinder-5dd7-account-create-update-q2fxk" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.469887 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/adde2fcf-e073-4936-a856-16bd660658a5-operator-scripts\") pod \"cinder-5dd7-account-create-update-q2fxk\" (UID: \"adde2fcf-e073-4936-a856-16bd660658a5\") " pod="openstack/cinder-5dd7-account-create-update-q2fxk" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.470767 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4873ca17-8a36-4c66-b86b-5a6784c7e4f2-operator-scripts\") pod \"cinder-db-create-6wtrt\" (UID: \"4873ca17-8a36-4c66-b86b-5a6784c7e4f2\") " pod="openstack/cinder-db-create-6wtrt" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.500138 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjtft\" (UniqueName: \"kubernetes.io/projected/4873ca17-8a36-4c66-b86b-5a6784c7e4f2-kube-api-access-gjtft\") pod \"cinder-db-create-6wtrt\" (UID: \"4873ca17-8a36-4c66-b86b-5a6784c7e4f2\") " pod="openstack/cinder-db-create-6wtrt" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.540996 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-g5j89"] Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.542330 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-g5j89" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.543544 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-9e39-account-create-update-9s4pv" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.554154 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.554498 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.554772 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ksmzm" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.555088 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.556738 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-g5j89"] Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.572874 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5szfg\" (UniqueName: \"kubernetes.io/projected/ced743c6-6d0e-439e-9d16-46ed28450a25-kube-api-access-5szfg\") pod \"heat-db-create-79smp\" (UID: \"ced743c6-6d0e-439e-9d16-46ed28450a25\") " pod="openstack/heat-db-create-79smp" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.572978 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp7vx\" (UniqueName: \"kubernetes.io/projected/7e0f8d57-5be8-4851-b8cf-e4ebdf753048-kube-api-access-hp7vx\") pod \"keystone-db-sync-g5j89\" (UID: \"7e0f8d57-5be8-4851-b8cf-e4ebdf753048\") " pod="openstack/keystone-db-sync-g5j89" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.573011 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27rdp\" (UniqueName: \"kubernetes.io/projected/adde2fcf-e073-4936-a856-16bd660658a5-kube-api-access-27rdp\") pod \"cinder-5dd7-account-create-update-q2fxk\" (UID: \"adde2fcf-e073-4936-a856-16bd660658a5\") " pod="openstack/cinder-5dd7-account-create-update-q2fxk" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.573040 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/adde2fcf-e073-4936-a856-16bd660658a5-operator-scripts\") pod \"cinder-5dd7-account-create-update-q2fxk\" (UID: \"adde2fcf-e073-4936-a856-16bd660658a5\") " pod="openstack/cinder-5dd7-account-create-update-q2fxk" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.573195 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e0f8d57-5be8-4851-b8cf-e4ebdf753048-config-data\") pod \"keystone-db-sync-g5j89\" (UID: \"7e0f8d57-5be8-4851-b8cf-e4ebdf753048\") " pod="openstack/keystone-db-sync-g5j89" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.573242 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e0f8d57-5be8-4851-b8cf-e4ebdf753048-combined-ca-bundle\") pod \"keystone-db-sync-g5j89\" (UID: \"7e0f8d57-5be8-4851-b8cf-e4ebdf753048\") " pod="openstack/keystone-db-sync-g5j89" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.573282 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ced743c6-6d0e-439e-9d16-46ed28450a25-operator-scripts\") pod \"heat-db-create-79smp\" (UID: \"ced743c6-6d0e-439e-9d16-46ed28450a25\") " pod="openstack/heat-db-create-79smp" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.574590 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/adde2fcf-e073-4936-a856-16bd660658a5-operator-scripts\") pod \"cinder-5dd7-account-create-update-q2fxk\" (UID: \"adde2fcf-e073-4936-a856-16bd660658a5\") " pod="openstack/cinder-5dd7-account-create-update-q2fxk" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.574893 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ced743c6-6d0e-439e-9d16-46ed28450a25-operator-scripts\") pod \"heat-db-create-79smp\" (UID: \"ced743c6-6d0e-439e-9d16-46ed28450a25\") " pod="openstack/heat-db-create-79smp" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.582914 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6wtrt" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.616271 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5szfg\" (UniqueName: \"kubernetes.io/projected/ced743c6-6d0e-439e-9d16-46ed28450a25-kube-api-access-5szfg\") pod \"heat-db-create-79smp\" (UID: \"ced743c6-6d0e-439e-9d16-46ed28450a25\") " pod="openstack/heat-db-create-79smp" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.621925 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27rdp\" (UniqueName: \"kubernetes.io/projected/adde2fcf-e073-4936-a856-16bd660658a5-kube-api-access-27rdp\") pod \"cinder-5dd7-account-create-update-q2fxk\" (UID: \"adde2fcf-e073-4936-a856-16bd660658a5\") " pod="openstack/cinder-5dd7-account-create-update-q2fxk" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.681297 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-zfpqn"] Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.683293 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp7vx\" (UniqueName: \"kubernetes.io/projected/7e0f8d57-5be8-4851-b8cf-e4ebdf753048-kube-api-access-hp7vx\") pod \"keystone-db-sync-g5j89\" (UID: \"7e0f8d57-5be8-4851-b8cf-e4ebdf753048\") " pod="openstack/keystone-db-sync-g5j89" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.683390 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e0f8d57-5be8-4851-b8cf-e4ebdf753048-config-data\") pod \"keystone-db-sync-g5j89\" (UID: \"7e0f8d57-5be8-4851-b8cf-e4ebdf753048\") " pod="openstack/keystone-db-sync-g5j89" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.683419 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e0f8d57-5be8-4851-b8cf-e4ebdf753048-combined-ca-bundle\") pod \"keystone-db-sync-g5j89\" (UID: \"7e0f8d57-5be8-4851-b8cf-e4ebdf753048\") " pod="openstack/keystone-db-sync-g5j89" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.686310 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zfpqn" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.694076 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e0f8d57-5be8-4851-b8cf-e4ebdf753048-combined-ca-bundle\") pod \"keystone-db-sync-g5j89\" (UID: \"7e0f8d57-5be8-4851-b8cf-e4ebdf753048\") " pod="openstack/keystone-db-sync-g5j89" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.695446 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e0f8d57-5be8-4851-b8cf-e4ebdf753048-config-data\") pod \"keystone-db-sync-g5j89\" (UID: \"7e0f8d57-5be8-4851-b8cf-e4ebdf753048\") " pod="openstack/keystone-db-sync-g5j89" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.697716 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-zfpqn"] Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.709198 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-0831-account-create-update-l6rfv"] Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.710777 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-0831-account-create-update-l6rfv" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.715736 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.720203 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp7vx\" (UniqueName: \"kubernetes.io/projected/7e0f8d57-5be8-4851-b8cf-e4ebdf753048-kube-api-access-hp7vx\") pod \"keystone-db-sync-g5j89\" (UID: \"7e0f8d57-5be8-4851-b8cf-e4ebdf753048\") " pod="openstack/keystone-db-sync-g5j89" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.723512 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-0831-account-create-update-l6rfv"] Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.774631 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-79smp" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.784129 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5dd7-account-create-update-q2fxk" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.785474 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d44n5\" (UniqueName: \"kubernetes.io/projected/e25f6c19-7b36-4144-8416-7df74580e906-kube-api-access-d44n5\") pod \"neutron-0831-account-create-update-l6rfv\" (UID: \"e25f6c19-7b36-4144-8416-7df74580e906\") " pod="openstack/neutron-0831-account-create-update-l6rfv" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.785522 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5abdded9-8645-49e8-910e-3ce04650f19e-operator-scripts\") pod \"neutron-db-create-zfpqn\" (UID: \"5abdded9-8645-49e8-910e-3ce04650f19e\") " pod="openstack/neutron-db-create-zfpqn" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.785574 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxr9q\" (UniqueName: \"kubernetes.io/projected/5abdded9-8645-49e8-910e-3ce04650f19e-kube-api-access-jxr9q\") pod \"neutron-db-create-zfpqn\" (UID: \"5abdded9-8645-49e8-910e-3ce04650f19e\") " pod="openstack/neutron-db-create-zfpqn" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.785622 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e25f6c19-7b36-4144-8416-7df74580e906-operator-scripts\") pod \"neutron-0831-account-create-update-l6rfv\" (UID: \"e25f6c19-7b36-4144-8416-7df74580e906\") " pod="openstack/neutron-0831-account-create-update-l6rfv" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.888033 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-g5j89" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.891142 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d44n5\" (UniqueName: \"kubernetes.io/projected/e25f6c19-7b36-4144-8416-7df74580e906-kube-api-access-d44n5\") pod \"neutron-0831-account-create-update-l6rfv\" (UID: \"e25f6c19-7b36-4144-8416-7df74580e906\") " pod="openstack/neutron-0831-account-create-update-l6rfv" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.891176 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5abdded9-8645-49e8-910e-3ce04650f19e-operator-scripts\") pod \"neutron-db-create-zfpqn\" (UID: \"5abdded9-8645-49e8-910e-3ce04650f19e\") " pod="openstack/neutron-db-create-zfpqn" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.891214 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxr9q\" (UniqueName: \"kubernetes.io/projected/5abdded9-8645-49e8-910e-3ce04650f19e-kube-api-access-jxr9q\") pod \"neutron-db-create-zfpqn\" (UID: \"5abdded9-8645-49e8-910e-3ce04650f19e\") " pod="openstack/neutron-db-create-zfpqn" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.891244 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e25f6c19-7b36-4144-8416-7df74580e906-operator-scripts\") pod \"neutron-0831-account-create-update-l6rfv\" (UID: \"e25f6c19-7b36-4144-8416-7df74580e906\") " pod="openstack/neutron-0831-account-create-update-l6rfv" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.891874 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e25f6c19-7b36-4144-8416-7df74580e906-operator-scripts\") pod \"neutron-0831-account-create-update-l6rfv\" (UID: \"e25f6c19-7b36-4144-8416-7df74580e906\") " pod="openstack/neutron-0831-account-create-update-l6rfv" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.893699 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5abdded9-8645-49e8-910e-3ce04650f19e-operator-scripts\") pod \"neutron-db-create-zfpqn\" (UID: \"5abdded9-8645-49e8-910e-3ce04650f19e\") " pod="openstack/neutron-db-create-zfpqn" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.900319 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0c404609-2299-4fe5-8c43-da5614df251e","Type":"ContainerStarted","Data":"adeae26ac312023c36ab25e47c8fa15f872a653fdd0446b054ff3362eb059b38"} Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.927038 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d44n5\" (UniqueName: \"kubernetes.io/projected/e25f6c19-7b36-4144-8416-7df74580e906-kube-api-access-d44n5\") pod \"neutron-0831-account-create-update-l6rfv\" (UID: \"e25f6c19-7b36-4144-8416-7df74580e906\") " pod="openstack/neutron-0831-account-create-update-l6rfv" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.932898 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxr9q\" (UniqueName: \"kubernetes.io/projected/5abdded9-8645-49e8-910e-3ce04650f19e-kube-api-access-jxr9q\") pod \"neutron-db-create-zfpqn\" (UID: \"5abdded9-8645-49e8-910e-3ce04650f19e\") " pod="openstack/neutron-db-create-zfpqn" Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.951713 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-p42tr-config-tcctj"] Nov 26 00:44:45 crc kubenswrapper[4766]: I1126 00:44:45.977430 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-p42tr-config-tcctj"] Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.106410 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zfpqn" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.108177 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-7hkqm"] Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.135917 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6008-account-create-update-dbw86"] Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.142623 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-0831-account-create-update-l6rfv" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.200602 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-p42tr-config-5zvhd"] Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.201797 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.204455 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.236100 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-p42tr-config-5zvhd"] Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.303647 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e801320b-14cb-42ce-9f2d-e8d93164ee5f-additional-scripts\") pod \"ovn-controller-p42tr-config-5zvhd\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.303725 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e801320b-14cb-42ce-9f2d-e8d93164ee5f-var-log-ovn\") pod \"ovn-controller-p42tr-config-5zvhd\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.303787 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t46sc\" (UniqueName: \"kubernetes.io/projected/e801320b-14cb-42ce-9f2d-e8d93164ee5f-kube-api-access-t46sc\") pod \"ovn-controller-p42tr-config-5zvhd\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.303812 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e801320b-14cb-42ce-9f2d-e8d93164ee5f-var-run\") pod \"ovn-controller-p42tr-config-5zvhd\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.303891 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e801320b-14cb-42ce-9f2d-e8d93164ee5f-scripts\") pod \"ovn-controller-p42tr-config-5zvhd\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.303919 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e801320b-14cb-42ce-9f2d-e8d93164ee5f-var-run-ovn\") pod \"ovn-controller-p42tr-config-5zvhd\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.405035 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e801320b-14cb-42ce-9f2d-e8d93164ee5f-scripts\") pod \"ovn-controller-p42tr-config-5zvhd\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.405092 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e801320b-14cb-42ce-9f2d-e8d93164ee5f-var-run-ovn\") pod \"ovn-controller-p42tr-config-5zvhd\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.405145 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e801320b-14cb-42ce-9f2d-e8d93164ee5f-additional-scripts\") pod \"ovn-controller-p42tr-config-5zvhd\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.405173 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e801320b-14cb-42ce-9f2d-e8d93164ee5f-var-log-ovn\") pod \"ovn-controller-p42tr-config-5zvhd\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.405229 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e801320b-14cb-42ce-9f2d-e8d93164ee5f-var-run\") pod \"ovn-controller-p42tr-config-5zvhd\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.405244 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t46sc\" (UniqueName: \"kubernetes.io/projected/e801320b-14cb-42ce-9f2d-e8d93164ee5f-kube-api-access-t46sc\") pod \"ovn-controller-p42tr-config-5zvhd\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.405636 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e801320b-14cb-42ce-9f2d-e8d93164ee5f-var-log-ovn\") pod \"ovn-controller-p42tr-config-5zvhd\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.405693 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e801320b-14cb-42ce-9f2d-e8d93164ee5f-var-run\") pod \"ovn-controller-p42tr-config-5zvhd\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.405673 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e801320b-14cb-42ce-9f2d-e8d93164ee5f-var-run-ovn\") pod \"ovn-controller-p42tr-config-5zvhd\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.406373 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e801320b-14cb-42ce-9f2d-e8d93164ee5f-additional-scripts\") pod \"ovn-controller-p42tr-config-5zvhd\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.407273 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e801320b-14cb-42ce-9f2d-e8d93164ee5f-scripts\") pod \"ovn-controller-p42tr-config-5zvhd\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.424489 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t46sc\" (UniqueName: \"kubernetes.io/projected/e801320b-14cb-42ce-9f2d-e8d93164ee5f-kube-api-access-t46sc\") pod \"ovn-controller-p42tr-config-5zvhd\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.523367 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:46 crc kubenswrapper[4766]: W1126 00:44:46.561392 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b80a95b_04b4_4b49_a799_cb6469c7b198.slice/crio-2da026e8c58f85e67191d1a77b0d5292a432b147d0fed63dbedb3d31336d792f WatchSource:0}: Error finding container 2da026e8c58f85e67191d1a77b0d5292a432b147d0fed63dbedb3d31336d792f: Status 404 returned error can't find the container with id 2da026e8c58f85e67191d1a77b0d5292a432b147d0fed63dbedb3d31336d792f Nov 26 00:44:46 crc kubenswrapper[4766]: W1126 00:44:46.563539 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef5c1bcc_7d69_4ec1_848b_244c7bc4b3ec.slice/crio-be761d86000f1c330fbba5c464d7e437ee294a596921f25f39c818832f639d5f WatchSource:0}: Error finding container be761d86000f1c330fbba5c464d7e437ee294a596921f25f39c818832f639d5f: Status 404 returned error can't find the container with id be761d86000f1c330fbba5c464d7e437ee294a596921f25f39c818832f639d5f Nov 26 00:44:46 crc kubenswrapper[4766]: I1126 00:44:46.974247 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6008-account-create-update-dbw86" event={"ID":"ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec","Type":"ContainerStarted","Data":"be761d86000f1c330fbba5c464d7e437ee294a596921f25f39c818832f639d5f"} Nov 26 00:44:47 crc kubenswrapper[4766]: I1126 00:44:47.019584 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-7hkqm" event={"ID":"0b80a95b-04b4-4b49-a799-cb6469c7b198","Type":"ContainerStarted","Data":"2da026e8c58f85e67191d1a77b0d5292a432b147d0fed63dbedb3d31336d792f"} Nov 26 00:44:47 crc kubenswrapper[4766]: I1126 00:44:47.219170 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-g5j89"] Nov 26 00:44:47 crc kubenswrapper[4766]: W1126 00:44:47.267392 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e0f8d57_5be8_4851_b8cf_e4ebdf753048.slice/crio-da0139c7b26f4d4f56704812db5bf52162d08491221570b93645951721c771f7 WatchSource:0}: Error finding container da0139c7b26f4d4f56704812db5bf52162d08491221570b93645951721c771f7: Status 404 returned error can't find the container with id da0139c7b26f4d4f56704812db5bf52162d08491221570b93645951721c771f7 Nov 26 00:44:47 crc kubenswrapper[4766]: I1126 00:44:47.600905 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-0831-account-create-update-l6rfv"] Nov 26 00:44:47 crc kubenswrapper[4766]: W1126 00:44:47.608166 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode25f6c19_7b36_4144_8416_7df74580e906.slice/crio-4a9781dff2cd9d052dc20dab61858a30722ba9607d000b472f434629431c80f8 WatchSource:0}: Error finding container 4a9781dff2cd9d052dc20dab61858a30722ba9607d000b472f434629431c80f8: Status 404 returned error can't find the container with id 4a9781dff2cd9d052dc20dab61858a30722ba9607d000b472f434629431c80f8 Nov 26 00:44:47 crc kubenswrapper[4766]: I1126 00:44:47.793408 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-79smp"] Nov 26 00:44:47 crc kubenswrapper[4766]: W1126 00:44:47.850704 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podced743c6_6d0e_439e_9d16_46ed28450a25.slice/crio-ed62cedf17f35663d1b0f19d10d44d20e64d4a0fa32285e68671972d4b60fbaf WatchSource:0}: Error finding container ed62cedf17f35663d1b0f19d10d44d20e64d4a0fa32285e68671972d4b60fbaf: Status 404 returned error can't find the container with id ed62cedf17f35663d1b0f19d10d44d20e64d4a0fa32285e68671972d4b60fbaf Nov 26 00:44:47 crc kubenswrapper[4766]: I1126 00:44:47.859988 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be0ded9d-d02c-4ac3-919b-d9d6e94866d9" path="/var/lib/kubelet/pods/be0ded9d-d02c-4ac3-919b-d9d6e94866d9/volumes" Nov 26 00:44:47 crc kubenswrapper[4766]: I1126 00:44:47.997836 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-zfpqn"] Nov 26 00:44:48 crc kubenswrapper[4766]: I1126 00:44:48.030819 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-9e39-account-create-update-9s4pv"] Nov 26 00:44:48 crc kubenswrapper[4766]: I1126 00:44:48.057740 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5dd7-account-create-update-q2fxk"] Nov 26 00:44:48 crc kubenswrapper[4766]: I1126 00:44:48.063396 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-g5j89" event={"ID":"7e0f8d57-5be8-4851-b8cf-e4ebdf753048","Type":"ContainerStarted","Data":"da0139c7b26f4d4f56704812db5bf52162d08491221570b93645951721c771f7"} Nov 26 00:44:48 crc kubenswrapper[4766]: I1126 00:44:48.082496 4766 generic.go:334] "Generic (PLEG): container finished" podID="ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec" containerID="3dec080f0849364d716930c5377aabf98641dbe93cb0acf2b18ebfbac25f7daf" exitCode=0 Nov 26 00:44:48 crc kubenswrapper[4766]: I1126 00:44:48.083002 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6008-account-create-update-dbw86" event={"ID":"ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec","Type":"ContainerDied","Data":"3dec080f0849364d716930c5377aabf98641dbe93cb0acf2b18ebfbac25f7daf"} Nov 26 00:44:48 crc kubenswrapper[4766]: I1126 00:44:48.139806 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a1caa1fb-917a-4267-a947-1194557347eb","Type":"ContainerStarted","Data":"3d87208c786584cf7ceb7e7cd935fce372dace88b9cd2ac155459dacbfbcdde0"} Nov 26 00:44:48 crc kubenswrapper[4766]: I1126 00:44:48.139851 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a1caa1fb-917a-4267-a947-1194557347eb","Type":"ContainerStarted","Data":"138eb81e47ddaf6c85cce0d9054d1b124045c4d3c7f2e39aa50cbbeee788d012"} Nov 26 00:44:48 crc kubenswrapper[4766]: I1126 00:44:48.152388 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-0831-account-create-update-l6rfv" event={"ID":"e25f6c19-7b36-4144-8416-7df74580e906","Type":"ContainerStarted","Data":"4a9781dff2cd9d052dc20dab61858a30722ba9607d000b472f434629431c80f8"} Nov 26 00:44:48 crc kubenswrapper[4766]: I1126 00:44:48.166242 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-6wtrt"] Nov 26 00:44:48 crc kubenswrapper[4766]: I1126 00:44:48.168034 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-79smp" event={"ID":"ced743c6-6d0e-439e-9d16-46ed28450a25","Type":"ContainerStarted","Data":"ed62cedf17f35663d1b0f19d10d44d20e64d4a0fa32285e68671972d4b60fbaf"} Nov 26 00:44:48 crc kubenswrapper[4766]: I1126 00:44:48.179074 4766 generic.go:334] "Generic (PLEG): container finished" podID="0b80a95b-04b4-4b49-a799-cb6469c7b198" containerID="6a94d4eb134b3e7e31ef1a063827a330051743ba95d2b1a225613f8053ac406d" exitCode=0 Nov 26 00:44:48 crc kubenswrapper[4766]: I1126 00:44:48.179338 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-7hkqm" event={"ID":"0b80a95b-04b4-4b49-a799-cb6469c7b198","Type":"ContainerDied","Data":"6a94d4eb134b3e7e31ef1a063827a330051743ba95d2b1a225613f8053ac406d"} Nov 26 00:44:48 crc kubenswrapper[4766]: I1126 00:44:48.201568 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-p42tr-config-5zvhd"] Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.229544 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5dd7-account-create-update-q2fxk" event={"ID":"adde2fcf-e073-4936-a856-16bd660658a5","Type":"ContainerStarted","Data":"5985e4cf0d953149e690837b7699efb7eff149dc2bcbe52131652bccb2d58583"} Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.231228 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a1caa1fb-917a-4267-a947-1194557347eb","Type":"ContainerStarted","Data":"ed28f9caba2d31315dc18a1ef0a978092995ee61479323b40af6145a3fc3b220"} Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.232234 4766 generic.go:334] "Generic (PLEG): container finished" podID="e25f6c19-7b36-4144-8416-7df74580e906" containerID="0fc5bc8ceed46933f386b560ac88b6d46950acf7e3fca8e34371c897c1589ba6" exitCode=0 Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.232266 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-0831-account-create-update-l6rfv" event={"ID":"e25f6c19-7b36-4144-8416-7df74580e906","Type":"ContainerDied","Data":"0fc5bc8ceed46933f386b560ac88b6d46950acf7e3fca8e34371c897c1589ba6"} Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.234929 4766 generic.go:334] "Generic (PLEG): container finished" podID="92fd9e2d-9152-46a4-8686-ef9fe6d0b790" containerID="d379cb5e72f4f66844ab6f351a9c5cdec0bbee9e536777d4b418b62a5c1a2b1b" exitCode=0 Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.234986 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-9e39-account-create-update-9s4pv" event={"ID":"92fd9e2d-9152-46a4-8686-ef9fe6d0b790","Type":"ContainerDied","Data":"d379cb5e72f4f66844ab6f351a9c5cdec0bbee9e536777d4b418b62a5c1a2b1b"} Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.235012 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-9e39-account-create-update-9s4pv" event={"ID":"92fd9e2d-9152-46a4-8686-ef9fe6d0b790","Type":"ContainerStarted","Data":"810f304ef09c6793016086c9143d94e9c9333ecd2ea1ce67259ec0d5b446d3a2"} Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.236152 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p42tr-config-5zvhd" event={"ID":"e801320b-14cb-42ce-9f2d-e8d93164ee5f","Type":"ContainerStarted","Data":"175c2261ba6217f19dd9623af693000aea7a2c77414d56a67f9d7bf3d38ce14a"} Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.237345 4766 generic.go:334] "Generic (PLEG): container finished" podID="ced743c6-6d0e-439e-9d16-46ed28450a25" containerID="e2cd4853efb6adc3594e8c0b2ed32609691e4a0a453aeb7a20751172fce520dc" exitCode=0 Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.237379 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-79smp" event={"ID":"ced743c6-6d0e-439e-9d16-46ed28450a25","Type":"ContainerDied","Data":"e2cd4853efb6adc3594e8c0b2ed32609691e4a0a453aeb7a20751172fce520dc"} Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.238389 4766 generic.go:334] "Generic (PLEG): container finished" podID="5abdded9-8645-49e8-910e-3ce04650f19e" containerID="4e715311f76e353b420abed0419f3ac40cd086666dfd92cac3e058990bcc330b" exitCode=0 Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.238428 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zfpqn" event={"ID":"5abdded9-8645-49e8-910e-3ce04650f19e","Type":"ContainerDied","Data":"4e715311f76e353b420abed0419f3ac40cd086666dfd92cac3e058990bcc330b"} Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.238443 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zfpqn" event={"ID":"5abdded9-8645-49e8-910e-3ce04650f19e","Type":"ContainerStarted","Data":"272b5e07175934b4aaf337eb6aa9b29ac4c38b246717e5afd6df709deb9931bd"} Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.239320 4766 generic.go:334] "Generic (PLEG): container finished" podID="4873ca17-8a36-4c66-b86b-5a6784c7e4f2" containerID="8cdd22a5a27b876b5463eb0cadb66912b4a961a13e341b9c2f852800c9abf62c" exitCode=0 Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.239463 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6wtrt" event={"ID":"4873ca17-8a36-4c66-b86b-5a6784c7e4f2","Type":"ContainerDied","Data":"8cdd22a5a27b876b5463eb0cadb66912b4a961a13e341b9c2f852800c9abf62c"} Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.239476 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6wtrt" event={"ID":"4873ca17-8a36-4c66-b86b-5a6784c7e4f2","Type":"ContainerStarted","Data":"e6cf2b2836022a1b369455cc865aa9fade50d9832afd60008335bd0180287751"} Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.784282 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6008-account-create-update-dbw86" Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.793512 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7hkqm" Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.930665 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec-operator-scripts\") pod \"ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec\" (UID: \"ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec\") " Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.931065 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ql6t5\" (UniqueName: \"kubernetes.io/projected/0b80a95b-04b4-4b49-a799-cb6469c7b198-kube-api-access-ql6t5\") pod \"0b80a95b-04b4-4b49-a799-cb6469c7b198\" (UID: \"0b80a95b-04b4-4b49-a799-cb6469c7b198\") " Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.931118 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47lrj\" (UniqueName: \"kubernetes.io/projected/ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec-kube-api-access-47lrj\") pod \"ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec\" (UID: \"ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec\") " Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.931226 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b80a95b-04b4-4b49-a799-cb6469c7b198-operator-scripts\") pod \"0b80a95b-04b4-4b49-a799-cb6469c7b198\" (UID: \"0b80a95b-04b4-4b49-a799-cb6469c7b198\") " Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.932144 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b80a95b-04b4-4b49-a799-cb6469c7b198-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0b80a95b-04b4-4b49-a799-cb6469c7b198" (UID: "0b80a95b-04b4-4b49-a799-cb6469c7b198"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.934686 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec" (UID: "ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.944065 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec-kube-api-access-47lrj" (OuterVolumeSpecName: "kube-api-access-47lrj") pod "ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec" (UID: "ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec"). InnerVolumeSpecName "kube-api-access-47lrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:49 crc kubenswrapper[4766]: I1126 00:44:49.958416 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b80a95b-04b4-4b49-a799-cb6469c7b198-kube-api-access-ql6t5" (OuterVolumeSpecName: "kube-api-access-ql6t5") pod "0b80a95b-04b4-4b49-a799-cb6469c7b198" (UID: "0b80a95b-04b4-4b49-a799-cb6469c7b198"). InnerVolumeSpecName "kube-api-access-ql6t5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:50 crc kubenswrapper[4766]: I1126 00:44:50.033174 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b80a95b-04b4-4b49-a799-cb6469c7b198-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:50 crc kubenswrapper[4766]: I1126 00:44:50.033207 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:50 crc kubenswrapper[4766]: I1126 00:44:50.033217 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ql6t5\" (UniqueName: \"kubernetes.io/projected/0b80a95b-04b4-4b49-a799-cb6469c7b198-kube-api-access-ql6t5\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:50 crc kubenswrapper[4766]: I1126 00:44:50.033228 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47lrj\" (UniqueName: \"kubernetes.io/projected/ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec-kube-api-access-47lrj\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:50 crc kubenswrapper[4766]: I1126 00:44:50.251927 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a1caa1fb-917a-4267-a947-1194557347eb","Type":"ContainerStarted","Data":"52351320e843e6043116741905bf67a8b69ba86cccbf3f3a2160372b2de7f802"} Nov 26 00:44:50 crc kubenswrapper[4766]: I1126 00:44:50.254277 4766 generic.go:334] "Generic (PLEG): container finished" podID="e801320b-14cb-42ce-9f2d-e8d93164ee5f" containerID="6d403e118929942f889cbb9cc1f0a5ff12be8b1d9c96aadbde1c60c3b7d47810" exitCode=0 Nov 26 00:44:50 crc kubenswrapper[4766]: I1126 00:44:50.254746 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p42tr-config-5zvhd" event={"ID":"e801320b-14cb-42ce-9f2d-e8d93164ee5f","Type":"ContainerDied","Data":"6d403e118929942f889cbb9cc1f0a5ff12be8b1d9c96aadbde1c60c3b7d47810"} Nov 26 00:44:50 crc kubenswrapper[4766]: I1126 00:44:50.256408 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-7hkqm" event={"ID":"0b80a95b-04b4-4b49-a799-cb6469c7b198","Type":"ContainerDied","Data":"2da026e8c58f85e67191d1a77b0d5292a432b147d0fed63dbedb3d31336d792f"} Nov 26 00:44:50 crc kubenswrapper[4766]: I1126 00:44:50.256435 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2da026e8c58f85e67191d1a77b0d5292a432b147d0fed63dbedb3d31336d792f" Nov 26 00:44:50 crc kubenswrapper[4766]: I1126 00:44:50.256410 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7hkqm" Nov 26 00:44:50 crc kubenswrapper[4766]: I1126 00:44:50.258941 4766 generic.go:334] "Generic (PLEG): container finished" podID="adde2fcf-e073-4936-a856-16bd660658a5" containerID="bb7285908ffbbf45d040e880724046e64e81ab2eea095e40609884bee5494e08" exitCode=0 Nov 26 00:44:50 crc kubenswrapper[4766]: I1126 00:44:50.259007 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5dd7-account-create-update-q2fxk" event={"ID":"adde2fcf-e073-4936-a856-16bd660658a5","Type":"ContainerDied","Data":"bb7285908ffbbf45d040e880724046e64e81ab2eea095e40609884bee5494e08"} Nov 26 00:44:50 crc kubenswrapper[4766]: I1126 00:44:50.262520 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6008-account-create-update-dbw86" Nov 26 00:44:50 crc kubenswrapper[4766]: I1126 00:44:50.262589 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6008-account-create-update-dbw86" event={"ID":"ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec","Type":"ContainerDied","Data":"be761d86000f1c330fbba5c464d7e437ee294a596921f25f39c818832f639d5f"} Nov 26 00:44:50 crc kubenswrapper[4766]: I1126 00:44:50.262623 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be761d86000f1c330fbba5c464d7e437ee294a596921f25f39c818832f639d5f" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.276955 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-79smp" event={"ID":"ced743c6-6d0e-439e-9d16-46ed28450a25","Type":"ContainerDied","Data":"ed62cedf17f35663d1b0f19d10d44d20e64d4a0fa32285e68671972d4b60fbaf"} Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.277210 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed62cedf17f35663d1b0f19d10d44d20e64d4a0fa32285e68671972d4b60fbaf" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.282230 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zfpqn" event={"ID":"5abdded9-8645-49e8-910e-3ce04650f19e","Type":"ContainerDied","Data":"272b5e07175934b4aaf337eb6aa9b29ac4c38b246717e5afd6df709deb9931bd"} Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.282273 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="272b5e07175934b4aaf337eb6aa9b29ac4c38b246717e5afd6df709deb9931bd" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.285164 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6wtrt" event={"ID":"4873ca17-8a36-4c66-b86b-5a6784c7e4f2","Type":"ContainerDied","Data":"e6cf2b2836022a1b369455cc865aa9fade50d9832afd60008335bd0180287751"} Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.285198 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6cf2b2836022a1b369455cc865aa9fade50d9832afd60008335bd0180287751" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.287134 4766 generic.go:334] "Generic (PLEG): container finished" podID="0c404609-2299-4fe5-8c43-da5614df251e" containerID="adeae26ac312023c36ab25e47c8fa15f872a653fdd0446b054ff3362eb059b38" exitCode=0 Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.287203 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0c404609-2299-4fe5-8c43-da5614df251e","Type":"ContainerDied","Data":"adeae26ac312023c36ab25e47c8fa15f872a653fdd0446b054ff3362eb059b38"} Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.288756 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-0831-account-create-update-l6rfv" event={"ID":"e25f6c19-7b36-4144-8416-7df74580e906","Type":"ContainerDied","Data":"4a9781dff2cd9d052dc20dab61858a30722ba9607d000b472f434629431c80f8"} Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.288785 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a9781dff2cd9d052dc20dab61858a30722ba9607d000b472f434629431c80f8" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.292164 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-9e39-account-create-update-9s4pv" event={"ID":"92fd9e2d-9152-46a4-8686-ef9fe6d0b790","Type":"ContainerDied","Data":"810f304ef09c6793016086c9143d94e9c9333ecd2ea1ce67259ec0d5b446d3a2"} Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.292200 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="810f304ef09c6793016086c9143d94e9c9333ecd2ea1ce67259ec0d5b446d3a2" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.439606 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zfpqn" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.467903 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-9e39-account-create-update-9s4pv" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.501928 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-0831-account-create-update-l6rfv" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.554890 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-79smp" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.555082 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6wtrt" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.564920 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-th79p\" (UniqueName: \"kubernetes.io/projected/92fd9e2d-9152-46a4-8686-ef9fe6d0b790-kube-api-access-th79p\") pod \"92fd9e2d-9152-46a4-8686-ef9fe6d0b790\" (UID: \"92fd9e2d-9152-46a4-8686-ef9fe6d0b790\") " Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.565147 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxr9q\" (UniqueName: \"kubernetes.io/projected/5abdded9-8645-49e8-910e-3ce04650f19e-kube-api-access-jxr9q\") pod \"5abdded9-8645-49e8-910e-3ce04650f19e\" (UID: \"5abdded9-8645-49e8-910e-3ce04650f19e\") " Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.565194 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5abdded9-8645-49e8-910e-3ce04650f19e-operator-scripts\") pod \"5abdded9-8645-49e8-910e-3ce04650f19e\" (UID: \"5abdded9-8645-49e8-910e-3ce04650f19e\") " Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.565254 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92fd9e2d-9152-46a4-8686-ef9fe6d0b790-operator-scripts\") pod \"92fd9e2d-9152-46a4-8686-ef9fe6d0b790\" (UID: \"92fd9e2d-9152-46a4-8686-ef9fe6d0b790\") " Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.566159 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92fd9e2d-9152-46a4-8686-ef9fe6d0b790-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "92fd9e2d-9152-46a4-8686-ef9fe6d0b790" (UID: "92fd9e2d-9152-46a4-8686-ef9fe6d0b790"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.572051 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5abdded9-8645-49e8-910e-3ce04650f19e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5abdded9-8645-49e8-910e-3ce04650f19e" (UID: "5abdded9-8645-49e8-910e-3ce04650f19e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.578700 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5abdded9-8645-49e8-910e-3ce04650f19e-kube-api-access-jxr9q" (OuterVolumeSpecName: "kube-api-access-jxr9q") pod "5abdded9-8645-49e8-910e-3ce04650f19e" (UID: "5abdded9-8645-49e8-910e-3ce04650f19e"). InnerVolumeSpecName "kube-api-access-jxr9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.598034 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92fd9e2d-9152-46a4-8686-ef9fe6d0b790-kube-api-access-th79p" (OuterVolumeSpecName: "kube-api-access-th79p") pod "92fd9e2d-9152-46a4-8686-ef9fe6d0b790" (UID: "92fd9e2d-9152-46a4-8686-ef9fe6d0b790"). InnerVolumeSpecName "kube-api-access-th79p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.668014 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d44n5\" (UniqueName: \"kubernetes.io/projected/e25f6c19-7b36-4144-8416-7df74580e906-kube-api-access-d44n5\") pod \"e25f6c19-7b36-4144-8416-7df74580e906\" (UID: \"e25f6c19-7b36-4144-8416-7df74580e906\") " Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.668107 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ced743c6-6d0e-439e-9d16-46ed28450a25-operator-scripts\") pod \"ced743c6-6d0e-439e-9d16-46ed28450a25\" (UID: \"ced743c6-6d0e-439e-9d16-46ed28450a25\") " Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.668172 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4873ca17-8a36-4c66-b86b-5a6784c7e4f2-operator-scripts\") pod \"4873ca17-8a36-4c66-b86b-5a6784c7e4f2\" (UID: \"4873ca17-8a36-4c66-b86b-5a6784c7e4f2\") " Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.668339 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e25f6c19-7b36-4144-8416-7df74580e906-operator-scripts\") pod \"e25f6c19-7b36-4144-8416-7df74580e906\" (UID: \"e25f6c19-7b36-4144-8416-7df74580e906\") " Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.668466 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5szfg\" (UniqueName: \"kubernetes.io/projected/ced743c6-6d0e-439e-9d16-46ed28450a25-kube-api-access-5szfg\") pod \"ced743c6-6d0e-439e-9d16-46ed28450a25\" (UID: \"ced743c6-6d0e-439e-9d16-46ed28450a25\") " Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.668540 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjtft\" (UniqueName: \"kubernetes.io/projected/4873ca17-8a36-4c66-b86b-5a6784c7e4f2-kube-api-access-gjtft\") pod \"4873ca17-8a36-4c66-b86b-5a6784c7e4f2\" (UID: \"4873ca17-8a36-4c66-b86b-5a6784c7e4f2\") " Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.668619 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ced743c6-6d0e-439e-9d16-46ed28450a25-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ced743c6-6d0e-439e-9d16-46ed28450a25" (UID: "ced743c6-6d0e-439e-9d16-46ed28450a25"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.668910 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e25f6c19-7b36-4144-8416-7df74580e906-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e25f6c19-7b36-4144-8416-7df74580e906" (UID: "e25f6c19-7b36-4144-8416-7df74580e906"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.669198 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-th79p\" (UniqueName: \"kubernetes.io/projected/92fd9e2d-9152-46a4-8686-ef9fe6d0b790-kube-api-access-th79p\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.669235 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e25f6c19-7b36-4144-8416-7df74580e906-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.669254 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxr9q\" (UniqueName: \"kubernetes.io/projected/5abdded9-8645-49e8-910e-3ce04650f19e-kube-api-access-jxr9q\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.669271 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5abdded9-8645-49e8-910e-3ce04650f19e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.669288 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ced743c6-6d0e-439e-9d16-46ed28450a25-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.669305 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92fd9e2d-9152-46a4-8686-ef9fe6d0b790-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.669706 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4873ca17-8a36-4c66-b86b-5a6784c7e4f2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4873ca17-8a36-4c66-b86b-5a6784c7e4f2" (UID: "4873ca17-8a36-4c66-b86b-5a6784c7e4f2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.679118 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4873ca17-8a36-4c66-b86b-5a6784c7e4f2-kube-api-access-gjtft" (OuterVolumeSpecName: "kube-api-access-gjtft") pod "4873ca17-8a36-4c66-b86b-5a6784c7e4f2" (UID: "4873ca17-8a36-4c66-b86b-5a6784c7e4f2"). InnerVolumeSpecName "kube-api-access-gjtft". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.679181 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e25f6c19-7b36-4144-8416-7df74580e906-kube-api-access-d44n5" (OuterVolumeSpecName: "kube-api-access-d44n5") pod "e25f6c19-7b36-4144-8416-7df74580e906" (UID: "e25f6c19-7b36-4144-8416-7df74580e906"). InnerVolumeSpecName "kube-api-access-d44n5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.679223 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ced743c6-6d0e-439e-9d16-46ed28450a25-kube-api-access-5szfg" (OuterVolumeSpecName: "kube-api-access-5szfg") pod "ced743c6-6d0e-439e-9d16-46ed28450a25" (UID: "ced743c6-6d0e-439e-9d16-46ed28450a25"). InnerVolumeSpecName "kube-api-access-5szfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.771001 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d44n5\" (UniqueName: \"kubernetes.io/projected/e25f6c19-7b36-4144-8416-7df74580e906-kube-api-access-d44n5\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.771037 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4873ca17-8a36-4c66-b86b-5a6784c7e4f2-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.771046 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5szfg\" (UniqueName: \"kubernetes.io/projected/ced743c6-6d0e-439e-9d16-46ed28450a25-kube-api-access-5szfg\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:51 crc kubenswrapper[4766]: I1126 00:44:51.771066 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjtft\" (UniqueName: \"kubernetes.io/projected/4873ca17-8a36-4c66-b86b-5a6784c7e4f2-kube-api-access-gjtft\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:52 crc kubenswrapper[4766]: I1126 00:44:52.306037 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a1caa1fb-917a-4267-a947-1194557347eb","Type":"ContainerStarted","Data":"25566a43b493d7a94ae3202eba08fbfcda0ef20b0d5db97c1f2aeb45f8c5c120"} Nov 26 00:44:52 crc kubenswrapper[4766]: I1126 00:44:52.308372 4766 generic.go:334] "Generic (PLEG): container finished" podID="04d8e581-a471-4aea-8c66-016ae86cd9e8" containerID="28f98fa01a48d0f4bded085fc78451fad87337d0a0eecc968095e0c60f6edffe" exitCode=0 Nov 26 00:44:52 crc kubenswrapper[4766]: I1126 00:44:52.308429 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-vh2cq" event={"ID":"04d8e581-a471-4aea-8c66-016ae86cd9e8","Type":"ContainerDied","Data":"28f98fa01a48d0f4bded085fc78451fad87337d0a0eecc968095e0c60f6edffe"} Nov 26 00:44:52 crc kubenswrapper[4766]: I1126 00:44:52.310308 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-9e39-account-create-update-9s4pv" Nov 26 00:44:52 crc kubenswrapper[4766]: I1126 00:44:52.310451 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0c404609-2299-4fe5-8c43-da5614df251e","Type":"ContainerStarted","Data":"5ee3e0772dbd1eaca0545970f974db17a8ac4570540b1e7bada966d7ad6368a5"} Nov 26 00:44:52 crc kubenswrapper[4766]: I1126 00:44:52.310512 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zfpqn" Nov 26 00:44:52 crc kubenswrapper[4766]: I1126 00:44:52.310523 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-0831-account-create-update-l6rfv" Nov 26 00:44:52 crc kubenswrapper[4766]: I1126 00:44:52.310558 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6wtrt" Nov 26 00:44:52 crc kubenswrapper[4766]: I1126 00:44:52.310708 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-79smp" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.342812 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p42tr-config-5zvhd" event={"ID":"e801320b-14cb-42ce-9f2d-e8d93164ee5f","Type":"ContainerDied","Data":"175c2261ba6217f19dd9623af693000aea7a2c77414d56a67f9d7bf3d38ce14a"} Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.343334 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="175c2261ba6217f19dd9623af693000aea7a2c77414d56a67f9d7bf3d38ce14a" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.346100 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5dd7-account-create-update-q2fxk" event={"ID":"adde2fcf-e073-4936-a856-16bd660658a5","Type":"ContainerDied","Data":"5985e4cf0d953149e690837b7699efb7eff149dc2bcbe52131652bccb2d58583"} Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.346136 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5985e4cf0d953149e690837b7699efb7eff149dc2bcbe52131652bccb2d58583" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.350392 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-vh2cq" event={"ID":"04d8e581-a471-4aea-8c66-016ae86cd9e8","Type":"ContainerDied","Data":"78ac1a2f5bf8b3fe68101bad7155b91bbdf992785d8bea5f3c01e94f27f15b24"} Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.350421 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78ac1a2f5bf8b3fe68101bad7155b91bbdf992785d8bea5f3c01e94f27f15b24" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.367181 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5dd7-account-create-update-q2fxk" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.426142 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-vh2cq" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.439439 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/adde2fcf-e073-4936-a856-16bd660658a5-operator-scripts\") pod \"adde2fcf-e073-4936-a856-16bd660658a5\" (UID: \"adde2fcf-e073-4936-a856-16bd660658a5\") " Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.439746 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27rdp\" (UniqueName: \"kubernetes.io/projected/adde2fcf-e073-4936-a856-16bd660658a5-kube-api-access-27rdp\") pod \"adde2fcf-e073-4936-a856-16bd660658a5\" (UID: \"adde2fcf-e073-4936-a856-16bd660658a5\") " Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.441073 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adde2fcf-e073-4936-a856-16bd660658a5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "adde2fcf-e073-4936-a856-16bd660658a5" (UID: "adde2fcf-e073-4936-a856-16bd660658a5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.445098 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adde2fcf-e073-4936-a856-16bd660658a5-kube-api-access-27rdp" (OuterVolumeSpecName: "kube-api-access-27rdp") pod "adde2fcf-e073-4936-a856-16bd660658a5" (UID: "adde2fcf-e073-4936-a856-16bd660658a5"). InnerVolumeSpecName "kube-api-access-27rdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.450283 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.541371 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e801320b-14cb-42ce-9f2d-e8d93164ee5f-var-run\") pod \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.541441 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lm9fx\" (UniqueName: \"kubernetes.io/projected/04d8e581-a471-4aea-8c66-016ae86cd9e8-kube-api-access-lm9fx\") pod \"04d8e581-a471-4aea-8c66-016ae86cd9e8\" (UID: \"04d8e581-a471-4aea-8c66-016ae86cd9e8\") " Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.541488 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04d8e581-a471-4aea-8c66-016ae86cd9e8-config-data\") pod \"04d8e581-a471-4aea-8c66-016ae86cd9e8\" (UID: \"04d8e581-a471-4aea-8c66-016ae86cd9e8\") " Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.541497 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e801320b-14cb-42ce-9f2d-e8d93164ee5f-var-run" (OuterVolumeSpecName: "var-run") pod "e801320b-14cb-42ce-9f2d-e8d93164ee5f" (UID: "e801320b-14cb-42ce-9f2d-e8d93164ee5f"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.541579 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e801320b-14cb-42ce-9f2d-e8d93164ee5f-additional-scripts\") pod \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.541603 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t46sc\" (UniqueName: \"kubernetes.io/projected/e801320b-14cb-42ce-9f2d-e8d93164ee5f-kube-api-access-t46sc\") pod \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.541638 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d8e581-a471-4aea-8c66-016ae86cd9e8-combined-ca-bundle\") pod \"04d8e581-a471-4aea-8c66-016ae86cd9e8\" (UID: \"04d8e581-a471-4aea-8c66-016ae86cd9e8\") " Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.541690 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e801320b-14cb-42ce-9f2d-e8d93164ee5f-var-log-ovn\") pod \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.541752 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e801320b-14cb-42ce-9f2d-e8d93164ee5f-var-run-ovn\") pod \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.541829 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04d8e581-a471-4aea-8c66-016ae86cd9e8-db-sync-config-data\") pod \"04d8e581-a471-4aea-8c66-016ae86cd9e8\" (UID: \"04d8e581-a471-4aea-8c66-016ae86cd9e8\") " Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.541855 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e801320b-14cb-42ce-9f2d-e8d93164ee5f-scripts\") pod \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\" (UID: \"e801320b-14cb-42ce-9f2d-e8d93164ee5f\") " Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.542009 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e801320b-14cb-42ce-9f2d-e8d93164ee5f-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "e801320b-14cb-42ce-9f2d-e8d93164ee5f" (UID: "e801320b-14cb-42ce-9f2d-e8d93164ee5f"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.542084 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e801320b-14cb-42ce-9f2d-e8d93164ee5f-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "e801320b-14cb-42ce-9f2d-e8d93164ee5f" (UID: "e801320b-14cb-42ce-9f2d-e8d93164ee5f"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.542478 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e801320b-14cb-42ce-9f2d-e8d93164ee5f-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "e801320b-14cb-42ce-9f2d-e8d93164ee5f" (UID: "e801320b-14cb-42ce-9f2d-e8d93164ee5f"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.542799 4766 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e801320b-14cb-42ce-9f2d-e8d93164ee5f-var-run\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.542879 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e801320b-14cb-42ce-9f2d-e8d93164ee5f-scripts" (OuterVolumeSpecName: "scripts") pod "e801320b-14cb-42ce-9f2d-e8d93164ee5f" (UID: "e801320b-14cb-42ce-9f2d-e8d93164ee5f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.542890 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27rdp\" (UniqueName: \"kubernetes.io/projected/adde2fcf-e073-4936-a856-16bd660658a5-kube-api-access-27rdp\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.542926 4766 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e801320b-14cb-42ce-9f2d-e8d93164ee5f-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.542936 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/adde2fcf-e073-4936-a856-16bd660658a5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.542948 4766 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e801320b-14cb-42ce-9f2d-e8d93164ee5f-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.542958 4766 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e801320b-14cb-42ce-9f2d-e8d93164ee5f-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.606326 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04d8e581-a471-4aea-8c66-016ae86cd9e8-kube-api-access-lm9fx" (OuterVolumeSpecName: "kube-api-access-lm9fx") pod "04d8e581-a471-4aea-8c66-016ae86cd9e8" (UID: "04d8e581-a471-4aea-8c66-016ae86cd9e8"). InnerVolumeSpecName "kube-api-access-lm9fx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.611254 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e801320b-14cb-42ce-9f2d-e8d93164ee5f-kube-api-access-t46sc" (OuterVolumeSpecName: "kube-api-access-t46sc") pod "e801320b-14cb-42ce-9f2d-e8d93164ee5f" (UID: "e801320b-14cb-42ce-9f2d-e8d93164ee5f"). InnerVolumeSpecName "kube-api-access-t46sc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.613878 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04d8e581-a471-4aea-8c66-016ae86cd9e8-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "04d8e581-a471-4aea-8c66-016ae86cd9e8" (UID: "04d8e581-a471-4aea-8c66-016ae86cd9e8"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.644304 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t46sc\" (UniqueName: \"kubernetes.io/projected/e801320b-14cb-42ce-9f2d-e8d93164ee5f-kube-api-access-t46sc\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.644344 4766 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04d8e581-a471-4aea-8c66-016ae86cd9e8-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.644380 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e801320b-14cb-42ce-9f2d-e8d93164ee5f-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.644393 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lm9fx\" (UniqueName: \"kubernetes.io/projected/04d8e581-a471-4aea-8c66-016ae86cd9e8-kube-api-access-lm9fx\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.646626 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04d8e581-a471-4aea-8c66-016ae86cd9e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04d8e581-a471-4aea-8c66-016ae86cd9e8" (UID: "04d8e581-a471-4aea-8c66-016ae86cd9e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.688758 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04d8e581-a471-4aea-8c66-016ae86cd9e8-config-data" (OuterVolumeSpecName: "config-data") pod "04d8e581-a471-4aea-8c66-016ae86cd9e8" (UID: "04d8e581-a471-4aea-8c66-016ae86cd9e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.748789 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04d8e581-a471-4aea-8c66-016ae86cd9e8-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:55 crc kubenswrapper[4766]: I1126 00:44:55.748827 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d8e581-a471-4aea-8c66-016ae86cd9e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.361409 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-g5j89" event={"ID":"7e0f8d57-5be8-4851-b8cf-e4ebdf753048","Type":"ContainerStarted","Data":"a0222e88d05b9e6f27da2a12b178be333fccf25b344163e8ae9b92d4c7116e94"} Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.366145 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a1caa1fb-917a-4267-a947-1194557347eb","Type":"ContainerStarted","Data":"b1d9e648d4a9f283d14d9bb1862710d957e2ebd6df209c03c751d4946e9d68cd"} Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.366177 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p42tr-config-5zvhd" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.366200 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a1caa1fb-917a-4267-a947-1194557347eb","Type":"ContainerStarted","Data":"6d5802d66c7d10d517ef1daf3c67bda1477222da3c8d055b00611c09aa3a2d96"} Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.366218 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a1caa1fb-917a-4267-a947-1194557347eb","Type":"ContainerStarted","Data":"1ea752e2cd3f5bc3779dd6ffdd149568a17bd5c1335ba306fcda1343bf82264d"} Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.366164 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5dd7-account-create-update-q2fxk" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.366396 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-vh2cq" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.389696 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-g5j89" podStartSLOduration=3.43504143 podStartE2EDuration="11.389671099s" podCreationTimestamp="2025-11-26 00:44:45 +0000 UTC" firstStartedPulling="2025-11-26 00:44:47.281828109 +0000 UTC m=+1268.130598539" lastFinishedPulling="2025-11-26 00:44:55.236457778 +0000 UTC m=+1276.085228208" observedRunningTime="2025-11-26 00:44:56.382487809 +0000 UTC m=+1277.231258259" watchObservedRunningTime="2025-11-26 00:44:56.389671099 +0000 UTC m=+1277.238441529" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.526372 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-p42tr-config-5zvhd"] Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.540272 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-p42tr-config-5zvhd"] Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.894928 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-qfdp6"] Nov 26 00:44:56 crc kubenswrapper[4766]: E1126 00:44:56.895400 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5abdded9-8645-49e8-910e-3ce04650f19e" containerName="mariadb-database-create" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.895415 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5abdded9-8645-49e8-910e-3ce04650f19e" containerName="mariadb-database-create" Nov 26 00:44:56 crc kubenswrapper[4766]: E1126 00:44:56.895435 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e801320b-14cb-42ce-9f2d-e8d93164ee5f" containerName="ovn-config" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.895442 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e801320b-14cb-42ce-9f2d-e8d93164ee5f" containerName="ovn-config" Nov 26 00:44:56 crc kubenswrapper[4766]: E1126 00:44:56.895457 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04d8e581-a471-4aea-8c66-016ae86cd9e8" containerName="glance-db-sync" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.895469 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="04d8e581-a471-4aea-8c66-016ae86cd9e8" containerName="glance-db-sync" Nov 26 00:44:56 crc kubenswrapper[4766]: E1126 00:44:56.895480 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adde2fcf-e073-4936-a856-16bd660658a5" containerName="mariadb-account-create-update" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.895487 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="adde2fcf-e073-4936-a856-16bd660658a5" containerName="mariadb-account-create-update" Nov 26 00:44:56 crc kubenswrapper[4766]: E1126 00:44:56.895504 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92fd9e2d-9152-46a4-8686-ef9fe6d0b790" containerName="mariadb-account-create-update" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.895511 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="92fd9e2d-9152-46a4-8686-ef9fe6d0b790" containerName="mariadb-account-create-update" Nov 26 00:44:56 crc kubenswrapper[4766]: E1126 00:44:56.895523 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e25f6c19-7b36-4144-8416-7df74580e906" containerName="mariadb-account-create-update" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.895530 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e25f6c19-7b36-4144-8416-7df74580e906" containerName="mariadb-account-create-update" Nov 26 00:44:56 crc kubenswrapper[4766]: E1126 00:44:56.895542 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4873ca17-8a36-4c66-b86b-5a6784c7e4f2" containerName="mariadb-database-create" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.895550 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="4873ca17-8a36-4c66-b86b-5a6784c7e4f2" containerName="mariadb-database-create" Nov 26 00:44:56 crc kubenswrapper[4766]: E1126 00:44:56.895570 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ced743c6-6d0e-439e-9d16-46ed28450a25" containerName="mariadb-database-create" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.895577 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ced743c6-6d0e-439e-9d16-46ed28450a25" containerName="mariadb-database-create" Nov 26 00:44:56 crc kubenswrapper[4766]: E1126 00:44:56.895588 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b80a95b-04b4-4b49-a799-cb6469c7b198" containerName="mariadb-database-create" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.895594 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b80a95b-04b4-4b49-a799-cb6469c7b198" containerName="mariadb-database-create" Nov 26 00:44:56 crc kubenswrapper[4766]: E1126 00:44:56.895603 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec" containerName="mariadb-account-create-update" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.895610 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec" containerName="mariadb-account-create-update" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.895844 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="04d8e581-a471-4aea-8c66-016ae86cd9e8" containerName="glance-db-sync" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.895859 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="ced743c6-6d0e-439e-9d16-46ed28450a25" containerName="mariadb-database-create" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.895868 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b80a95b-04b4-4b49-a799-cb6469c7b198" containerName="mariadb-database-create" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.895879 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="e801320b-14cb-42ce-9f2d-e8d93164ee5f" containerName="ovn-config" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.895895 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="adde2fcf-e073-4936-a856-16bd660658a5" containerName="mariadb-account-create-update" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.895909 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec" containerName="mariadb-account-create-update" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.895921 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="92fd9e2d-9152-46a4-8686-ef9fe6d0b790" containerName="mariadb-account-create-update" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.895930 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="4873ca17-8a36-4c66-b86b-5a6784c7e4f2" containerName="mariadb-database-create" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.895944 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="5abdded9-8645-49e8-910e-3ce04650f19e" containerName="mariadb-database-create" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.895952 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="e25f6c19-7b36-4144-8416-7df74580e906" containerName="mariadb-account-create-update" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.906281 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" Nov 26 00:44:56 crc kubenswrapper[4766]: I1126 00:44:56.939506 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-qfdp6"] Nov 26 00:44:57 crc kubenswrapper[4766]: I1126 00:44:57.074568 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df6nb\" (UniqueName: \"kubernetes.io/projected/b89002b9-75d6-4868-9148-d3d9d8c71db7-kube-api-access-df6nb\") pod \"dnsmasq-dns-74dc88fc-qfdp6\" (UID: \"b89002b9-75d6-4868-9148-d3d9d8c71db7\") " pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" Nov 26 00:44:57 crc kubenswrapper[4766]: I1126 00:44:57.074670 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-dns-svc\") pod \"dnsmasq-dns-74dc88fc-qfdp6\" (UID: \"b89002b9-75d6-4868-9148-d3d9d8c71db7\") " pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" Nov 26 00:44:57 crc kubenswrapper[4766]: I1126 00:44:57.074732 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-ovsdbserver-sb\") pod \"dnsmasq-dns-74dc88fc-qfdp6\" (UID: \"b89002b9-75d6-4868-9148-d3d9d8c71db7\") " pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" Nov 26 00:44:57 crc kubenswrapper[4766]: I1126 00:44:57.074792 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-config\") pod \"dnsmasq-dns-74dc88fc-qfdp6\" (UID: \"b89002b9-75d6-4868-9148-d3d9d8c71db7\") " pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" Nov 26 00:44:57 crc kubenswrapper[4766]: I1126 00:44:57.074820 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-ovsdbserver-nb\") pod \"dnsmasq-dns-74dc88fc-qfdp6\" (UID: \"b89002b9-75d6-4868-9148-d3d9d8c71db7\") " pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" Nov 26 00:44:57 crc kubenswrapper[4766]: I1126 00:44:57.176089 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-config\") pod \"dnsmasq-dns-74dc88fc-qfdp6\" (UID: \"b89002b9-75d6-4868-9148-d3d9d8c71db7\") " pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" Nov 26 00:44:57 crc kubenswrapper[4766]: I1126 00:44:57.176136 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-ovsdbserver-nb\") pod \"dnsmasq-dns-74dc88fc-qfdp6\" (UID: \"b89002b9-75d6-4868-9148-d3d9d8c71db7\") " pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" Nov 26 00:44:57 crc kubenswrapper[4766]: I1126 00:44:57.176188 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df6nb\" (UniqueName: \"kubernetes.io/projected/b89002b9-75d6-4868-9148-d3d9d8c71db7-kube-api-access-df6nb\") pod \"dnsmasq-dns-74dc88fc-qfdp6\" (UID: \"b89002b9-75d6-4868-9148-d3d9d8c71db7\") " pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" Nov 26 00:44:57 crc kubenswrapper[4766]: I1126 00:44:57.176252 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-dns-svc\") pod \"dnsmasq-dns-74dc88fc-qfdp6\" (UID: \"b89002b9-75d6-4868-9148-d3d9d8c71db7\") " pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" Nov 26 00:44:57 crc kubenswrapper[4766]: I1126 00:44:57.176333 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-ovsdbserver-sb\") pod \"dnsmasq-dns-74dc88fc-qfdp6\" (UID: \"b89002b9-75d6-4868-9148-d3d9d8c71db7\") " pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" Nov 26 00:44:57 crc kubenswrapper[4766]: I1126 00:44:57.177253 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-ovsdbserver-sb\") pod \"dnsmasq-dns-74dc88fc-qfdp6\" (UID: \"b89002b9-75d6-4868-9148-d3d9d8c71db7\") " pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" Nov 26 00:44:57 crc kubenswrapper[4766]: I1126 00:44:57.177458 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-config\") pod \"dnsmasq-dns-74dc88fc-qfdp6\" (UID: \"b89002b9-75d6-4868-9148-d3d9d8c71db7\") " pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" Nov 26 00:44:57 crc kubenswrapper[4766]: I1126 00:44:57.177519 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-ovsdbserver-nb\") pod \"dnsmasq-dns-74dc88fc-qfdp6\" (UID: \"b89002b9-75d6-4868-9148-d3d9d8c71db7\") " pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" Nov 26 00:44:57 crc kubenswrapper[4766]: I1126 00:44:57.177594 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-dns-svc\") pod \"dnsmasq-dns-74dc88fc-qfdp6\" (UID: \"b89002b9-75d6-4868-9148-d3d9d8c71db7\") " pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" Nov 26 00:44:57 crc kubenswrapper[4766]: I1126 00:44:57.193800 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df6nb\" (UniqueName: \"kubernetes.io/projected/b89002b9-75d6-4868-9148-d3d9d8c71db7-kube-api-access-df6nb\") pod \"dnsmasq-dns-74dc88fc-qfdp6\" (UID: \"b89002b9-75d6-4868-9148-d3d9d8c71db7\") " pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" Nov 26 00:44:57 crc kubenswrapper[4766]: I1126 00:44:57.232951 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" Nov 26 00:44:57 crc kubenswrapper[4766]: I1126 00:44:57.712017 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-qfdp6"] Nov 26 00:44:57 crc kubenswrapper[4766]: I1126 00:44:57.842621 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e801320b-14cb-42ce-9f2d-e8d93164ee5f" path="/var/lib/kubelet/pods/e801320b-14cb-42ce-9f2d-e8d93164ee5f/volumes" Nov 26 00:44:58 crc kubenswrapper[4766]: I1126 00:44:58.396997 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0c404609-2299-4fe5-8c43-da5614df251e","Type":"ContainerStarted","Data":"cb239b53957503da63249e94873518b66d9aa56896132440ee3ea7bb710825d0"} Nov 26 00:44:58 crc kubenswrapper[4766]: I1126 00:44:58.400611 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a1caa1fb-917a-4267-a947-1194557347eb","Type":"ContainerStarted","Data":"3af2e6ff1283e1aebd017bdc95d3472d0f64302402744a1d151faa30cabb28ee"} Nov 26 00:44:58 crc kubenswrapper[4766]: I1126 00:44:58.402014 4766 generic.go:334] "Generic (PLEG): container finished" podID="b89002b9-75d6-4868-9148-d3d9d8c71db7" containerID="11ce1280feb760acb49d0a261bcc7651b2f6a4530abfdd01f231c6d08c32d305" exitCode=0 Nov 26 00:44:58 crc kubenswrapper[4766]: I1126 00:44:58.402078 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" event={"ID":"b89002b9-75d6-4868-9148-d3d9d8c71db7","Type":"ContainerDied","Data":"11ce1280feb760acb49d0a261bcc7651b2f6a4530abfdd01f231c6d08c32d305"} Nov 26 00:44:58 crc kubenswrapper[4766]: I1126 00:44:58.402141 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" event={"ID":"b89002b9-75d6-4868-9148-d3d9d8c71db7","Type":"ContainerStarted","Data":"3bc36ef357eea6a23eed6745b53c1d3e422695e773c6a5526cd1923a8b149a60"} Nov 26 00:44:59 crc kubenswrapper[4766]: I1126 00:44:59.415453 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a1caa1fb-917a-4267-a947-1194557347eb","Type":"ContainerStarted","Data":"7cb428c03bccc31cf9b33ea29c6146e21b03989d57d8577386fe5a8b21c68900"} Nov 26 00:44:59 crc kubenswrapper[4766]: I1126 00:44:59.415790 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a1caa1fb-917a-4267-a947-1194557347eb","Type":"ContainerStarted","Data":"13b955ebec3d7c1f3ccfe766fafead4caaacc5f94d8cb573894026f527279037"} Nov 26 00:44:59 crc kubenswrapper[4766]: I1126 00:44:59.415802 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a1caa1fb-917a-4267-a947-1194557347eb","Type":"ContainerStarted","Data":"b28c8d647bb7b3c31cf07125ba83828f18572305437c80be974e204c8a7d7e72"} Nov 26 00:44:59 crc kubenswrapper[4766]: I1126 00:44:59.415810 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a1caa1fb-917a-4267-a947-1194557347eb","Type":"ContainerStarted","Data":"9c346aa6926e84edce8de6c6aa3d7c27924fa59e577f3d116fad01c38dd2dab2"} Nov 26 00:44:59 crc kubenswrapper[4766]: I1126 00:44:59.415819 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a1caa1fb-917a-4267-a947-1194557347eb","Type":"ContainerStarted","Data":"025a635e7975f2863d0090c2b6de37d00dfbf2e3de572921c915a2d54bbfe909"} Nov 26 00:44:59 crc kubenswrapper[4766]: I1126 00:44:59.420173 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" event={"ID":"b89002b9-75d6-4868-9148-d3d9d8c71db7","Type":"ContainerStarted","Data":"21b4df373ed9a5c49f23e4412a8bc927260849cc2f0d141c898fc136b51988a1"} Nov 26 00:44:59 crc kubenswrapper[4766]: I1126 00:44:59.420338 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" Nov 26 00:44:59 crc kubenswrapper[4766]: I1126 00:44:59.430786 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0c404609-2299-4fe5-8c43-da5614df251e","Type":"ContainerStarted","Data":"c90fee74ff4f413822dcda81f803850feeeca0c5ce08fcede1b30b161a633664"} Nov 26 00:44:59 crc kubenswrapper[4766]: I1126 00:44:59.441016 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" podStartSLOduration=3.440995671 podStartE2EDuration="3.440995671s" podCreationTimestamp="2025-11-26 00:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:44:59.436846867 +0000 UTC m=+1280.285617297" watchObservedRunningTime="2025-11-26 00:44:59.440995671 +0000 UTC m=+1280.289766101" Nov 26 00:44:59 crc kubenswrapper[4766]: I1126 00:44:59.458213 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=18.458190192 podStartE2EDuration="18.458190192s" podCreationTimestamp="2025-11-26 00:44:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:44:59.455961676 +0000 UTC m=+1280.304732116" watchObservedRunningTime="2025-11-26 00:44:59.458190192 +0000 UTC m=+1280.306960622" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.132579 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg"] Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.134625 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.137181 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.137782 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.152232 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg"] Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.245062 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/32dc2556-2ee6-4b11-90c0-c074b1c2e774-secret-volume\") pod \"collect-profiles-29401965-qggxg\" (UID: \"32dc2556-2ee6-4b11-90c0-c074b1c2e774\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.245127 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/32dc2556-2ee6-4b11-90c0-c074b1c2e774-config-volume\") pod \"collect-profiles-29401965-qggxg\" (UID: \"32dc2556-2ee6-4b11-90c0-c074b1c2e774\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.245184 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz7nl\" (UniqueName: \"kubernetes.io/projected/32dc2556-2ee6-4b11-90c0-c074b1c2e774-kube-api-access-nz7nl\") pod \"collect-profiles-29401965-qggxg\" (UID: \"32dc2556-2ee6-4b11-90c0-c074b1c2e774\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.346598 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/32dc2556-2ee6-4b11-90c0-c074b1c2e774-secret-volume\") pod \"collect-profiles-29401965-qggxg\" (UID: \"32dc2556-2ee6-4b11-90c0-c074b1c2e774\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.346639 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/32dc2556-2ee6-4b11-90c0-c074b1c2e774-config-volume\") pod \"collect-profiles-29401965-qggxg\" (UID: \"32dc2556-2ee6-4b11-90c0-c074b1c2e774\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.346721 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz7nl\" (UniqueName: \"kubernetes.io/projected/32dc2556-2ee6-4b11-90c0-c074b1c2e774-kube-api-access-nz7nl\") pod \"collect-profiles-29401965-qggxg\" (UID: \"32dc2556-2ee6-4b11-90c0-c074b1c2e774\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.347766 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/32dc2556-2ee6-4b11-90c0-c074b1c2e774-config-volume\") pod \"collect-profiles-29401965-qggxg\" (UID: \"32dc2556-2ee6-4b11-90c0-c074b1c2e774\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.354520 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/32dc2556-2ee6-4b11-90c0-c074b1c2e774-secret-volume\") pod \"collect-profiles-29401965-qggxg\" (UID: \"32dc2556-2ee6-4b11-90c0-c074b1c2e774\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.368641 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz7nl\" (UniqueName: \"kubernetes.io/projected/32dc2556-2ee6-4b11-90c0-c074b1c2e774-kube-api-access-nz7nl\") pod \"collect-profiles-29401965-qggxg\" (UID: \"32dc2556-2ee6-4b11-90c0-c074b1c2e774\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.452491 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a1caa1fb-917a-4267-a947-1194557347eb","Type":"ContainerStarted","Data":"df2586dd586273f1ad67e32894a58a3c6f3949a74de61959b737b0dbba3830a0"} Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.459952 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.478311 4766 generic.go:334] "Generic (PLEG): container finished" podID="7e0f8d57-5be8-4851-b8cf-e4ebdf753048" containerID="a0222e88d05b9e6f27da2a12b178be333fccf25b344163e8ae9b92d4c7116e94" exitCode=0 Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.478364 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-g5j89" event={"ID":"7e0f8d57-5be8-4851-b8cf-e4ebdf753048","Type":"ContainerDied","Data":"a0222e88d05b9e6f27da2a12b178be333fccf25b344163e8ae9b92d4c7116e94"} Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.517131 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=37.003994179 podStartE2EDuration="50.517109069s" podCreationTimestamp="2025-11-26 00:44:10 +0000 UTC" firstStartedPulling="2025-11-26 00:44:44.412708024 +0000 UTC m=+1265.261478454" lastFinishedPulling="2025-11-26 00:44:57.925822914 +0000 UTC m=+1278.774593344" observedRunningTime="2025-11-26 00:45:00.514107463 +0000 UTC m=+1281.362877913" watchObservedRunningTime="2025-11-26 00:45:00.517109069 +0000 UTC m=+1281.365879499" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.807562 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-qfdp6"] Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.831740 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-pj25t"] Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.835296 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.839053 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.865300 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-pj25t"] Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.964038 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-pj25t\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.964107 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-config\") pod \"dnsmasq-dns-5f59b8f679-pj25t\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.964136 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26x8f\" (UniqueName: \"kubernetes.io/projected/eb6091b5-cc82-44cf-93cd-825dd952dbb3-kube-api-access-26x8f\") pod \"dnsmasq-dns-5f59b8f679-pj25t\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.964180 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-pj25t\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.964363 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-pj25t\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:00 crc kubenswrapper[4766]: I1126 00:45:00.964388 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-pj25t\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.004609 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg"] Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.066137 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-pj25t\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.066184 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-pj25t\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.066237 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-pj25t\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.066271 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-config\") pod \"dnsmasq-dns-5f59b8f679-pj25t\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.066287 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26x8f\" (UniqueName: \"kubernetes.io/projected/eb6091b5-cc82-44cf-93cd-825dd952dbb3-kube-api-access-26x8f\") pod \"dnsmasq-dns-5f59b8f679-pj25t\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.066314 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-pj25t\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.067268 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-pj25t\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.068499 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-pj25t\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.068883 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-config\") pod \"dnsmasq-dns-5f59b8f679-pj25t\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.069046 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-pj25t\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.070009 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-pj25t\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.093837 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26x8f\" (UniqueName: \"kubernetes.io/projected/eb6091b5-cc82-44cf-93cd-825dd952dbb3-kube-api-access-26x8f\") pod \"dnsmasq-dns-5f59b8f679-pj25t\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.167245 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.510300 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg" event={"ID":"32dc2556-2ee6-4b11-90c0-c074b1c2e774","Type":"ContainerStarted","Data":"0e267bf2eb7b3baa5687c89d55b47b09cf78dd1b9bed01fd6d72dce260665cba"} Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.510625 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg" event={"ID":"32dc2556-2ee6-4b11-90c0-c074b1c2e774","Type":"ContainerStarted","Data":"dba76ac70a983dbece745e7f42480484eeb256530355906665b95def6c9d9e49"} Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.512296 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" podUID="b89002b9-75d6-4868-9148-d3d9d8c71db7" containerName="dnsmasq-dns" containerID="cri-o://21b4df373ed9a5c49f23e4412a8bc927260849cc2f0d141c898fc136b51988a1" gracePeriod=10 Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.669778 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-pj25t"] Nov 26 00:45:01 crc kubenswrapper[4766]: W1126 00:45:01.714052 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb6091b5_cc82_44cf_93cd_825dd952dbb3.slice/crio-0a315a25146024c38c57e119df498cbf09855ad8c814e61d8d24fec0d578e933 WatchSource:0}: Error finding container 0a315a25146024c38c57e119df498cbf09855ad8c814e61d8d24fec0d578e933: Status 404 returned error can't find the container with id 0a315a25146024c38c57e119df498cbf09855ad8c814e61d8d24fec0d578e933 Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.849762 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.963362 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-g5j89" Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.993273 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e0f8d57-5be8-4851-b8cf-e4ebdf753048-combined-ca-bundle\") pod \"7e0f8d57-5be8-4851-b8cf-e4ebdf753048\" (UID: \"7e0f8d57-5be8-4851-b8cf-e4ebdf753048\") " Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.993485 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e0f8d57-5be8-4851-b8cf-e4ebdf753048-config-data\") pod \"7e0f8d57-5be8-4851-b8cf-e4ebdf753048\" (UID: \"7e0f8d57-5be8-4851-b8cf-e4ebdf753048\") " Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.993568 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp7vx\" (UniqueName: \"kubernetes.io/projected/7e0f8d57-5be8-4851-b8cf-e4ebdf753048-kube-api-access-hp7vx\") pod \"7e0f8d57-5be8-4851-b8cf-e4ebdf753048\" (UID: \"7e0f8d57-5be8-4851-b8cf-e4ebdf753048\") " Nov 26 00:45:01 crc kubenswrapper[4766]: I1126 00:45:01.998426 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e0f8d57-5be8-4851-b8cf-e4ebdf753048-kube-api-access-hp7vx" (OuterVolumeSpecName: "kube-api-access-hp7vx") pod "7e0f8d57-5be8-4851-b8cf-e4ebdf753048" (UID: "7e0f8d57-5be8-4851-b8cf-e4ebdf753048"). InnerVolumeSpecName "kube-api-access-hp7vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.025803 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e0f8d57-5be8-4851-b8cf-e4ebdf753048-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7e0f8d57-5be8-4851-b8cf-e4ebdf753048" (UID: "7e0f8d57-5be8-4851-b8cf-e4ebdf753048"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.031873 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.046474 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e0f8d57-5be8-4851-b8cf-e4ebdf753048-config-data" (OuterVolumeSpecName: "config-data") pod "7e0f8d57-5be8-4851-b8cf-e4ebdf753048" (UID: "7e0f8d57-5be8-4851-b8cf-e4ebdf753048"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.095611 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-config\") pod \"b89002b9-75d6-4868-9148-d3d9d8c71db7\" (UID: \"b89002b9-75d6-4868-9148-d3d9d8c71db7\") " Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.095709 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-dns-svc\") pod \"b89002b9-75d6-4868-9148-d3d9d8c71db7\" (UID: \"b89002b9-75d6-4868-9148-d3d9d8c71db7\") " Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.095744 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-ovsdbserver-sb\") pod \"b89002b9-75d6-4868-9148-d3d9d8c71db7\" (UID: \"b89002b9-75d6-4868-9148-d3d9d8c71db7\") " Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.095781 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df6nb\" (UniqueName: \"kubernetes.io/projected/b89002b9-75d6-4868-9148-d3d9d8c71db7-kube-api-access-df6nb\") pod \"b89002b9-75d6-4868-9148-d3d9d8c71db7\" (UID: \"b89002b9-75d6-4868-9148-d3d9d8c71db7\") " Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.095961 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-ovsdbserver-nb\") pod \"b89002b9-75d6-4868-9148-d3d9d8c71db7\" (UID: \"b89002b9-75d6-4868-9148-d3d9d8c71db7\") " Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.096312 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e0f8d57-5be8-4851-b8cf-e4ebdf753048-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.096330 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp7vx\" (UniqueName: \"kubernetes.io/projected/7e0f8d57-5be8-4851-b8cf-e4ebdf753048-kube-api-access-hp7vx\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.096339 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e0f8d57-5be8-4851-b8cf-e4ebdf753048-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.099318 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b89002b9-75d6-4868-9148-d3d9d8c71db7-kube-api-access-df6nb" (OuterVolumeSpecName: "kube-api-access-df6nb") pod "b89002b9-75d6-4868-9148-d3d9d8c71db7" (UID: "b89002b9-75d6-4868-9148-d3d9d8c71db7"). InnerVolumeSpecName "kube-api-access-df6nb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.146262 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b89002b9-75d6-4868-9148-d3d9d8c71db7" (UID: "b89002b9-75d6-4868-9148-d3d9d8c71db7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.151834 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b89002b9-75d6-4868-9148-d3d9d8c71db7" (UID: "b89002b9-75d6-4868-9148-d3d9d8c71db7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.159670 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-config" (OuterVolumeSpecName: "config") pod "b89002b9-75d6-4868-9148-d3d9d8c71db7" (UID: "b89002b9-75d6-4868-9148-d3d9d8c71db7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.191117 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b89002b9-75d6-4868-9148-d3d9d8c71db7" (UID: "b89002b9-75d6-4868-9148-d3d9d8c71db7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.198067 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.198117 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.198134 4766 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.198149 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b89002b9-75d6-4868-9148-d3d9d8c71db7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.198162 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df6nb\" (UniqueName: \"kubernetes.io/projected/b89002b9-75d6-4868-9148-d3d9d8c71db7-kube-api-access-df6nb\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.520106 4766 generic.go:334] "Generic (PLEG): container finished" podID="32dc2556-2ee6-4b11-90c0-c074b1c2e774" containerID="0e267bf2eb7b3baa5687c89d55b47b09cf78dd1b9bed01fd6d72dce260665cba" exitCode=0 Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.520169 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg" event={"ID":"32dc2556-2ee6-4b11-90c0-c074b1c2e774","Type":"ContainerDied","Data":"0e267bf2eb7b3baa5687c89d55b47b09cf78dd1b9bed01fd6d72dce260665cba"} Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.521504 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-g5j89" event={"ID":"7e0f8d57-5be8-4851-b8cf-e4ebdf753048","Type":"ContainerDied","Data":"da0139c7b26f4d4f56704812db5bf52162d08491221570b93645951721c771f7"} Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.521532 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da0139c7b26f4d4f56704812db5bf52162d08491221570b93645951721c771f7" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.521576 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-g5j89" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.523228 4766 generic.go:334] "Generic (PLEG): container finished" podID="eb6091b5-cc82-44cf-93cd-825dd952dbb3" containerID="3d37dbe375032d57ea58c3c0cb53002b299b4aa1db9eaf7cb6ad313046dc6c6e" exitCode=0 Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.523315 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" event={"ID":"eb6091b5-cc82-44cf-93cd-825dd952dbb3","Type":"ContainerDied","Data":"3d37dbe375032d57ea58c3c0cb53002b299b4aa1db9eaf7cb6ad313046dc6c6e"} Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.523353 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" event={"ID":"eb6091b5-cc82-44cf-93cd-825dd952dbb3","Type":"ContainerStarted","Data":"0a315a25146024c38c57e119df498cbf09855ad8c814e61d8d24fec0d578e933"} Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.525713 4766 generic.go:334] "Generic (PLEG): container finished" podID="b89002b9-75d6-4868-9148-d3d9d8c71db7" containerID="21b4df373ed9a5c49f23e4412a8bc927260849cc2f0d141c898fc136b51988a1" exitCode=0 Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.525757 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.525774 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" event={"ID":"b89002b9-75d6-4868-9148-d3d9d8c71db7","Type":"ContainerDied","Data":"21b4df373ed9a5c49f23e4412a8bc927260849cc2f0d141c898fc136b51988a1"} Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.525813 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-qfdp6" event={"ID":"b89002b9-75d6-4868-9148-d3d9d8c71db7","Type":"ContainerDied","Data":"3bc36ef357eea6a23eed6745b53c1d3e422695e773c6a5526cd1923a8b149a60"} Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.525880 4766 scope.go:117] "RemoveContainer" containerID="21b4df373ed9a5c49f23e4412a8bc927260849cc2f0d141c898fc136b51988a1" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.550817 4766 scope.go:117] "RemoveContainer" containerID="11ce1280feb760acb49d0a261bcc7651b2f6a4530abfdd01f231c6d08c32d305" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.741078 4766 scope.go:117] "RemoveContainer" containerID="21b4df373ed9a5c49f23e4412a8bc927260849cc2f0d141c898fc136b51988a1" Nov 26 00:45:02 crc kubenswrapper[4766]: E1126 00:45:02.741757 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21b4df373ed9a5c49f23e4412a8bc927260849cc2f0d141c898fc136b51988a1\": container with ID starting with 21b4df373ed9a5c49f23e4412a8bc927260849cc2f0d141c898fc136b51988a1 not found: ID does not exist" containerID="21b4df373ed9a5c49f23e4412a8bc927260849cc2f0d141c898fc136b51988a1" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.741805 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21b4df373ed9a5c49f23e4412a8bc927260849cc2f0d141c898fc136b51988a1"} err="failed to get container status \"21b4df373ed9a5c49f23e4412a8bc927260849cc2f0d141c898fc136b51988a1\": rpc error: code = NotFound desc = could not find container \"21b4df373ed9a5c49f23e4412a8bc927260849cc2f0d141c898fc136b51988a1\": container with ID starting with 21b4df373ed9a5c49f23e4412a8bc927260849cc2f0d141c898fc136b51988a1 not found: ID does not exist" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.741835 4766 scope.go:117] "RemoveContainer" containerID="11ce1280feb760acb49d0a261bcc7651b2f6a4530abfdd01f231c6d08c32d305" Nov 26 00:45:02 crc kubenswrapper[4766]: E1126 00:45:02.747193 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11ce1280feb760acb49d0a261bcc7651b2f6a4530abfdd01f231c6d08c32d305\": container with ID starting with 11ce1280feb760acb49d0a261bcc7651b2f6a4530abfdd01f231c6d08c32d305 not found: ID does not exist" containerID="11ce1280feb760acb49d0a261bcc7651b2f6a4530abfdd01f231c6d08c32d305" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.747250 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11ce1280feb760acb49d0a261bcc7651b2f6a4530abfdd01f231c6d08c32d305"} err="failed to get container status \"11ce1280feb760acb49d0a261bcc7651b2f6a4530abfdd01f231c6d08c32d305\": rpc error: code = NotFound desc = could not find container \"11ce1280feb760acb49d0a261bcc7651b2f6a4530abfdd01f231c6d08c32d305\": container with ID starting with 11ce1280feb760acb49d0a261bcc7651b2f6a4530abfdd01f231c6d08c32d305 not found: ID does not exist" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.855753 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-pj25t"] Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.879372 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-qfdp6"] Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.912948 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-sbhxb"] Nov 26 00:45:02 crc kubenswrapper[4766]: E1126 00:45:02.914324 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b89002b9-75d6-4868-9148-d3d9d8c71db7" containerName="dnsmasq-dns" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.914351 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="b89002b9-75d6-4868-9148-d3d9d8c71db7" containerName="dnsmasq-dns" Nov 26 00:45:02 crc kubenswrapper[4766]: E1126 00:45:02.914374 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e0f8d57-5be8-4851-b8cf-e4ebdf753048" containerName="keystone-db-sync" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.914382 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e0f8d57-5be8-4851-b8cf-e4ebdf753048" containerName="keystone-db-sync" Nov 26 00:45:02 crc kubenswrapper[4766]: E1126 00:45:02.914427 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b89002b9-75d6-4868-9148-d3d9d8c71db7" containerName="init" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.914434 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="b89002b9-75d6-4868-9148-d3d9d8c71db7" containerName="init" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.915164 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="b89002b9-75d6-4868-9148-d3d9d8c71db7" containerName="dnsmasq-dns" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.915174 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e0f8d57-5be8-4851-b8cf-e4ebdf753048" containerName="keystone-db-sync" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.917770 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.960134 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-rg6jq"] Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.969081 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.976217 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-qfdp6"] Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.976983 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.977163 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.977188 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ksmzm" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.977533 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.977631 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 26 00:45:02 crc kubenswrapper[4766]: I1126 00:45:02.991727 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rg6jq"] Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.011461 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-sbhxb"] Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.027980 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-9259k"] Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.030845 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-9259k" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.035939 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.035966 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-8fcdz" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.078013 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-9259k"] Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.127825 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqcmd\" (UniqueName: \"kubernetes.io/projected/edf21578-364f-472d-99bd-ec283f2fb585-kube-api-access-wqcmd\") pod \"dnsmasq-dns-bbf5cc879-sbhxb\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.127887 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-fernet-keys\") pod \"keystone-bootstrap-rg6jq\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.127962 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-config\") pod \"dnsmasq-dns-bbf5cc879-sbhxb\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.128038 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-sbhxb\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.128096 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-credential-keys\") pod \"keystone-bootstrap-rg6jq\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.128122 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-sbhxb\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.128153 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-config-data\") pod \"keystone-bootstrap-rg6jq\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.128234 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-scripts\") pod \"keystone-bootstrap-rg6jq\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.128261 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbg98\" (UniqueName: \"kubernetes.io/projected/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-kube-api-access-wbg98\") pod \"keystone-bootstrap-rg6jq\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.128293 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-sbhxb\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.128318 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-combined-ca-bundle\") pod \"keystone-bootstrap-rg6jq\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.128371 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-sbhxb\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.145369 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.181743 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-8sbgx"] Nov 26 00:45:03 crc kubenswrapper[4766]: E1126 00:45:03.182417 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32dc2556-2ee6-4b11-90c0-c074b1c2e774" containerName="collect-profiles" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.182432 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="32dc2556-2ee6-4b11-90c0-c074b1c2e774" containerName="collect-profiles" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.182663 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="32dc2556-2ee6-4b11-90c0-c074b1c2e774" containerName="collect-profiles" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.183317 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.191414 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-wcjrw" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.191629 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.192032 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.210332 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-8sbgx"] Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.235345 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-scripts\") pod \"keystone-bootstrap-rg6jq\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.235394 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbg98\" (UniqueName: \"kubernetes.io/projected/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-kube-api-access-wbg98\") pod \"keystone-bootstrap-rg6jq\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.235440 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-sbhxb\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.235477 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-combined-ca-bundle\") pod \"keystone-bootstrap-rg6jq\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.235510 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fd45623-8db1-440a-960a-a848abbb0c90-config-data\") pod \"heat-db-sync-9259k\" (UID: \"2fd45623-8db1-440a-960a-a848abbb0c90\") " pod="openstack/heat-db-sync-9259k" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.235536 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-db-sync-config-data\") pod \"cinder-db-sync-8sbgx\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.235618 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-sbhxb\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.235665 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ffaf6d04-ed5d-4438-a285-bc4981448be5-etc-machine-id\") pod \"cinder-db-sync-8sbgx\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.235709 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqcmd\" (UniqueName: \"kubernetes.io/projected/edf21578-364f-472d-99bd-ec283f2fb585-kube-api-access-wqcmd\") pod \"dnsmasq-dns-bbf5cc879-sbhxb\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.235739 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-fernet-keys\") pod \"keystone-bootstrap-rg6jq\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.235789 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-config\") pod \"dnsmasq-dns-bbf5cc879-sbhxb\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.235805 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-combined-ca-bundle\") pod \"cinder-db-sync-8sbgx\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.235832 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fd45623-8db1-440a-960a-a848abbb0c90-combined-ca-bundle\") pod \"heat-db-sync-9259k\" (UID: \"2fd45623-8db1-440a-960a-a848abbb0c90\") " pod="openstack/heat-db-sync-9259k" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.235919 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-sbhxb\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.235970 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-scripts\") pod \"cinder-db-sync-8sbgx\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.235998 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf9q8\" (UniqueName: \"kubernetes.io/projected/2fd45623-8db1-440a-960a-a848abbb0c90-kube-api-access-wf9q8\") pod \"heat-db-sync-9259k\" (UID: \"2fd45623-8db1-440a-960a-a848abbb0c90\") " pod="openstack/heat-db-sync-9259k" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.236016 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-credential-keys\") pod \"keystone-bootstrap-rg6jq\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.236064 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-sbhxb\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.236080 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgxb5\" (UniqueName: \"kubernetes.io/projected/ffaf6d04-ed5d-4438-a285-bc4981448be5-kube-api-access-jgxb5\") pod \"cinder-db-sync-8sbgx\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.236094 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-config-data\") pod \"cinder-db-sync-8sbgx\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.236123 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-config-data\") pod \"keystone-bootstrap-rg6jq\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.243759 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-config\") pod \"dnsmasq-dns-bbf5cc879-sbhxb\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.244334 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-fernet-keys\") pod \"keystone-bootstrap-rg6jq\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.244553 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-sbhxb\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.244990 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-sbhxb\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.245931 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-sbhxb\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.246823 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-sbhxb\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.249347 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-scripts\") pod \"keystone-bootstrap-rg6jq\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.249748 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-config-data\") pod \"keystone-bootstrap-rg6jq\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.256570 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-combined-ca-bundle\") pod \"keystone-bootstrap-rg6jq\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.258635 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-credential-keys\") pod \"keystone-bootstrap-rg6jq\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.275065 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbg98\" (UniqueName: \"kubernetes.io/projected/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-kube-api-access-wbg98\") pod \"keystone-bootstrap-rg6jq\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.278966 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-8bfjz"] Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.280449 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-8bfjz" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.281840 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqcmd\" (UniqueName: \"kubernetes.io/projected/edf21578-364f-472d-99bd-ec283f2fb585-kube-api-access-wqcmd\") pod \"dnsmasq-dns-bbf5cc879-sbhxb\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.291470 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.291750 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-8w9h8" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.291887 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.308765 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-8bfjz"] Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.323796 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-pd2t2"] Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.325409 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pd2t2" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.328813 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-87dbv" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.329046 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.334399 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-pd2t2"] Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.337406 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/32dc2556-2ee6-4b11-90c0-c074b1c2e774-secret-volume\") pod \"32dc2556-2ee6-4b11-90c0-c074b1c2e774\" (UID: \"32dc2556-2ee6-4b11-90c0-c074b1c2e774\") " Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.337457 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nz7nl\" (UniqueName: \"kubernetes.io/projected/32dc2556-2ee6-4b11-90c0-c074b1c2e774-kube-api-access-nz7nl\") pod \"32dc2556-2ee6-4b11-90c0-c074b1c2e774\" (UID: \"32dc2556-2ee6-4b11-90c0-c074b1c2e774\") " Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.337523 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/32dc2556-2ee6-4b11-90c0-c074b1c2e774-config-volume\") pod \"32dc2556-2ee6-4b11-90c0-c074b1c2e774\" (UID: \"32dc2556-2ee6-4b11-90c0-c074b1c2e774\") " Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.337877 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ffaf6d04-ed5d-4438-a285-bc4981448be5-etc-machine-id\") pod \"cinder-db-sync-8sbgx\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.337936 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b156a1fc-37a3-45d3-999d-55852f10bcc3-combined-ca-bundle\") pod \"neutron-db-sync-8bfjz\" (UID: \"b156a1fc-37a3-45d3-999d-55852f10bcc3\") " pod="openstack/neutron-db-sync-8bfjz" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.337957 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r568f\" (UniqueName: \"kubernetes.io/projected/66feb239-5790-4cfe-96db-e9dd6413bb09-kube-api-access-r568f\") pod \"barbican-db-sync-pd2t2\" (UID: \"66feb239-5790-4cfe-96db-e9dd6413bb09\") " pod="openstack/barbican-db-sync-pd2t2" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.337979 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-combined-ca-bundle\") pod \"cinder-db-sync-8sbgx\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.338000 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fd45623-8db1-440a-960a-a848abbb0c90-combined-ca-bundle\") pod \"heat-db-sync-9259k\" (UID: \"2fd45623-8db1-440a-960a-a848abbb0c90\") " pod="openstack/heat-db-sync-9259k" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.338021 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8g58\" (UniqueName: \"kubernetes.io/projected/b156a1fc-37a3-45d3-999d-55852f10bcc3-kube-api-access-t8g58\") pod \"neutron-db-sync-8bfjz\" (UID: \"b156a1fc-37a3-45d3-999d-55852f10bcc3\") " pod="openstack/neutron-db-sync-8bfjz" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.338044 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b156a1fc-37a3-45d3-999d-55852f10bcc3-config\") pod \"neutron-db-sync-8bfjz\" (UID: \"b156a1fc-37a3-45d3-999d-55852f10bcc3\") " pod="openstack/neutron-db-sync-8bfjz" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.338061 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66feb239-5790-4cfe-96db-e9dd6413bb09-db-sync-config-data\") pod \"barbican-db-sync-pd2t2\" (UID: \"66feb239-5790-4cfe-96db-e9dd6413bb09\") " pod="openstack/barbican-db-sync-pd2t2" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.338108 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-scripts\") pod \"cinder-db-sync-8sbgx\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.338129 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf9q8\" (UniqueName: \"kubernetes.io/projected/2fd45623-8db1-440a-960a-a848abbb0c90-kube-api-access-wf9q8\") pod \"heat-db-sync-9259k\" (UID: \"2fd45623-8db1-440a-960a-a848abbb0c90\") " pod="openstack/heat-db-sync-9259k" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.338152 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgxb5\" (UniqueName: \"kubernetes.io/projected/ffaf6d04-ed5d-4438-a285-bc4981448be5-kube-api-access-jgxb5\") pod \"cinder-db-sync-8sbgx\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.338183 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-config-data\") pod \"cinder-db-sync-8sbgx\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.338210 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66feb239-5790-4cfe-96db-e9dd6413bb09-combined-ca-bundle\") pod \"barbican-db-sync-pd2t2\" (UID: \"66feb239-5790-4cfe-96db-e9dd6413bb09\") " pod="openstack/barbican-db-sync-pd2t2" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.338325 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fd45623-8db1-440a-960a-a848abbb0c90-config-data\") pod \"heat-db-sync-9259k\" (UID: \"2fd45623-8db1-440a-960a-a848abbb0c90\") " pod="openstack/heat-db-sync-9259k" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.338346 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-db-sync-config-data\") pod \"cinder-db-sync-8sbgx\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.342841 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-db-sync-config-data\") pod \"cinder-db-sync-8sbgx\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.346201 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32dc2556-2ee6-4b11-90c0-c074b1c2e774-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "32dc2556-2ee6-4b11-90c0-c074b1c2e774" (UID: "32dc2556-2ee6-4b11-90c0-c074b1c2e774"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.347509 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ffaf6d04-ed5d-4438-a285-bc4981448be5-etc-machine-id\") pod \"cinder-db-sync-8sbgx\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.352016 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32dc2556-2ee6-4b11-90c0-c074b1c2e774-kube-api-access-nz7nl" (OuterVolumeSpecName: "kube-api-access-nz7nl") pod "32dc2556-2ee6-4b11-90c0-c074b1c2e774" (UID: "32dc2556-2ee6-4b11-90c0-c074b1c2e774"). InnerVolumeSpecName "kube-api-access-nz7nl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.355705 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-config-data\") pod \"cinder-db-sync-8sbgx\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.356287 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32dc2556-2ee6-4b11-90c0-c074b1c2e774-config-volume" (OuterVolumeSpecName: "config-volume") pod "32dc2556-2ee6-4b11-90c0-c074b1c2e774" (UID: "32dc2556-2ee6-4b11-90c0-c074b1c2e774"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.358078 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fd45623-8db1-440a-960a-a848abbb0c90-config-data\") pod \"heat-db-sync-9259k\" (UID: \"2fd45623-8db1-440a-960a-a848abbb0c90\") " pod="openstack/heat-db-sync-9259k" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.362759 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-scripts\") pod \"cinder-db-sync-8sbgx\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.366985 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fd45623-8db1-440a-960a-a848abbb0c90-combined-ca-bundle\") pod \"heat-db-sync-9259k\" (UID: \"2fd45623-8db1-440a-960a-a848abbb0c90\") " pod="openstack/heat-db-sync-9259k" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.387257 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf9q8\" (UniqueName: \"kubernetes.io/projected/2fd45623-8db1-440a-960a-a848abbb0c90-kube-api-access-wf9q8\") pod \"heat-db-sync-9259k\" (UID: \"2fd45623-8db1-440a-960a-a848abbb0c90\") " pod="openstack/heat-db-sync-9259k" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.387850 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-combined-ca-bundle\") pod \"cinder-db-sync-8sbgx\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.406870 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgxb5\" (UniqueName: \"kubernetes.io/projected/ffaf6d04-ed5d-4438-a285-bc4981448be5-kube-api-access-jgxb5\") pod \"cinder-db-sync-8sbgx\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.421945 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-7rfnj"] Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.426392 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-7rfnj" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.428568 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-7rfnj"] Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.429005 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.429136 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-44jtp" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.430162 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.439837 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-sbhxb"] Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.440677 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.440744 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66feb239-5790-4cfe-96db-e9dd6413bb09-combined-ca-bundle\") pod \"barbican-db-sync-pd2t2\" (UID: \"66feb239-5790-4cfe-96db-e9dd6413bb09\") " pod="openstack/barbican-db-sync-pd2t2" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.440842 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b156a1fc-37a3-45d3-999d-55852f10bcc3-combined-ca-bundle\") pod \"neutron-db-sync-8bfjz\" (UID: \"b156a1fc-37a3-45d3-999d-55852f10bcc3\") " pod="openstack/neutron-db-sync-8bfjz" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.440863 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r568f\" (UniqueName: \"kubernetes.io/projected/66feb239-5790-4cfe-96db-e9dd6413bb09-kube-api-access-r568f\") pod \"barbican-db-sync-pd2t2\" (UID: \"66feb239-5790-4cfe-96db-e9dd6413bb09\") " pod="openstack/barbican-db-sync-pd2t2" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.440891 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8g58\" (UniqueName: \"kubernetes.io/projected/b156a1fc-37a3-45d3-999d-55852f10bcc3-kube-api-access-t8g58\") pod \"neutron-db-sync-8bfjz\" (UID: \"b156a1fc-37a3-45d3-999d-55852f10bcc3\") " pod="openstack/neutron-db-sync-8bfjz" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.440913 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b156a1fc-37a3-45d3-999d-55852f10bcc3-config\") pod \"neutron-db-sync-8bfjz\" (UID: \"b156a1fc-37a3-45d3-999d-55852f10bcc3\") " pod="openstack/neutron-db-sync-8bfjz" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.440928 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66feb239-5790-4cfe-96db-e9dd6413bb09-db-sync-config-data\") pod \"barbican-db-sync-pd2t2\" (UID: \"66feb239-5790-4cfe-96db-e9dd6413bb09\") " pod="openstack/barbican-db-sync-pd2t2" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.440994 4766 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/32dc2556-2ee6-4b11-90c0-c074b1c2e774-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.441006 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nz7nl\" (UniqueName: \"kubernetes.io/projected/32dc2556-2ee6-4b11-90c0-c074b1c2e774-kube-api-access-nz7nl\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.441015 4766 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/32dc2556-2ee6-4b11-90c0-c074b1c2e774-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.445863 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.448295 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.460723 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.466235 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.466435 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.466555 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b156a1fc-37a3-45d3-999d-55852f10bcc3-config\") pod \"neutron-db-sync-8bfjz\" (UID: \"b156a1fc-37a3-45d3-999d-55852f10bcc3\") " pod="openstack/neutron-db-sync-8bfjz" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.467104 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66feb239-5790-4cfe-96db-e9dd6413bb09-db-sync-config-data\") pod \"barbican-db-sync-pd2t2\" (UID: \"66feb239-5790-4cfe-96db-e9dd6413bb09\") " pod="openstack/barbican-db-sync-pd2t2" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.467257 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b156a1fc-37a3-45d3-999d-55852f10bcc3-combined-ca-bundle\") pod \"neutron-db-sync-8bfjz\" (UID: \"b156a1fc-37a3-45d3-999d-55852f10bcc3\") " pod="openstack/neutron-db-sync-8bfjz" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.467606 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66feb239-5790-4cfe-96db-e9dd6413bb09-combined-ca-bundle\") pod \"barbican-db-sync-pd2t2\" (UID: \"66feb239-5790-4cfe-96db-e9dd6413bb09\") " pod="openstack/barbican-db-sync-pd2t2" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.470827 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.473891 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-9259k" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.479146 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r568f\" (UniqueName: \"kubernetes.io/projected/66feb239-5790-4cfe-96db-e9dd6413bb09-kube-api-access-r568f\") pod \"barbican-db-sync-pd2t2\" (UID: \"66feb239-5790-4cfe-96db-e9dd6413bb09\") " pod="openstack/barbican-db-sync-pd2t2" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.484878 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-lvp7w"] Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.486554 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.488944 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8g58\" (UniqueName: \"kubernetes.io/projected/b156a1fc-37a3-45d3-999d-55852f10bcc3-kube-api-access-t8g58\") pod \"neutron-db-sync-8bfjz\" (UID: \"b156a1fc-37a3-45d3-999d-55852f10bcc3\") " pod="openstack/neutron-db-sync-8bfjz" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.505468 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-lvp7w"] Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.510490 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.543845 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-logs\") pod \"placement-db-sync-7rfnj\" (UID: \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\") " pod="openstack/placement-db-sync-7rfnj" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.543975 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-scripts\") pod \"placement-db-sync-7rfnj\" (UID: \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\") " pod="openstack/placement-db-sync-7rfnj" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.544009 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-config-data\") pod \"placement-db-sync-7rfnj\" (UID: \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\") " pod="openstack/placement-db-sync-7rfnj" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.544034 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-combined-ca-bundle\") pod \"placement-db-sync-7rfnj\" (UID: \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\") " pod="openstack/placement-db-sync-7rfnj" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.544067 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28v2q\" (UniqueName: \"kubernetes.io/projected/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-kube-api-access-28v2q\") pod \"placement-db-sync-7rfnj\" (UID: \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\") " pod="openstack/placement-db-sync-7rfnj" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.584961 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" event={"ID":"eb6091b5-cc82-44cf-93cd-825dd952dbb3","Type":"ContainerStarted","Data":"77371af89f48688fc7cf2d1aa9c52f2692ea30ca7db0d6479b26509911a3c9b0"} Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.585502 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.585090 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" podUID="eb6091b5-cc82-44cf-93cd-825dd952dbb3" containerName="dnsmasq-dns" containerID="cri-o://77371af89f48688fc7cf2d1aa9c52f2692ea30ca7db0d6479b26509911a3c9b0" gracePeriod=10 Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.595808 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg" event={"ID":"32dc2556-2ee6-4b11-90c0-c074b1c2e774","Type":"ContainerDied","Data":"dba76ac70a983dbece745e7f42480484eeb256530355906665b95def6c9d9e49"} Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.595845 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dba76ac70a983dbece745e7f42480484eeb256530355906665b95def6c9d9e49" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.595909 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.605324 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-8bfjz" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.607848 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pd2t2" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.651552 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-config-data\") pod \"ceilometer-0\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.651612 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.651683 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-config\") pod \"dnsmasq-dns-56df8fb6b7-lvp7w\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.651728 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-lvp7w\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.651815 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-scripts\") pod \"ceilometer-0\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.651847 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55ec5b52-2985-4607-8c89-590f43a3096c-log-httpd\") pod \"ceilometer-0\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.651881 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-scripts\") pod \"placement-db-sync-7rfnj\" (UID: \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\") " pod="openstack/placement-db-sync-7rfnj" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.651926 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-config-data\") pod \"placement-db-sync-7rfnj\" (UID: \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\") " pod="openstack/placement-db-sync-7rfnj" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.651968 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-combined-ca-bundle\") pod \"placement-db-sync-7rfnj\" (UID: \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\") " pod="openstack/placement-db-sync-7rfnj" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.652022 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-lvp7w\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.652085 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28v2q\" (UniqueName: \"kubernetes.io/projected/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-kube-api-access-28v2q\") pod \"placement-db-sync-7rfnj\" (UID: \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\") " pod="openstack/placement-db-sync-7rfnj" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.652155 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-lvp7w\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.652205 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.652277 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-lvp7w\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.652304 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55ec5b52-2985-4607-8c89-590f43a3096c-run-httpd\") pod \"ceilometer-0\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.652363 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq4qd\" (UniqueName: \"kubernetes.io/projected/cbf480e0-f3e9-445c-a174-db6ce0967ed7-kube-api-access-sq4qd\") pod \"dnsmasq-dns-56df8fb6b7-lvp7w\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.655550 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" podStartSLOduration=3.6555275849999997 podStartE2EDuration="3.655527585s" podCreationTimestamp="2025-11-26 00:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:45:03.604911905 +0000 UTC m=+1284.453682345" watchObservedRunningTime="2025-11-26 00:45:03.655527585 +0000 UTC m=+1284.504298015" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.659856 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-logs\") pod \"placement-db-sync-7rfnj\" (UID: \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\") " pod="openstack/placement-db-sync-7rfnj" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.663971 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-logs\") pod \"placement-db-sync-7rfnj\" (UID: \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\") " pod="openstack/placement-db-sync-7rfnj" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.664570 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrq6q\" (UniqueName: \"kubernetes.io/projected/55ec5b52-2985-4607-8c89-590f43a3096c-kube-api-access-wrq6q\") pod \"ceilometer-0\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.669172 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-config-data\") pod \"placement-db-sync-7rfnj\" (UID: \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\") " pod="openstack/placement-db-sync-7rfnj" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.669410 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-combined-ca-bundle\") pod \"placement-db-sync-7rfnj\" (UID: \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\") " pod="openstack/placement-db-sync-7rfnj" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.672223 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-scripts\") pod \"placement-db-sync-7rfnj\" (UID: \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\") " pod="openstack/placement-db-sync-7rfnj" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.714177 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28v2q\" (UniqueName: \"kubernetes.io/projected/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-kube-api-access-28v2q\") pod \"placement-db-sync-7rfnj\" (UID: \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\") " pod="openstack/placement-db-sync-7rfnj" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.766618 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-lvp7w\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.767077 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-lvp7w\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.767270 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.767342 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-lvp7w\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.767367 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55ec5b52-2985-4607-8c89-590f43a3096c-run-httpd\") pod \"ceilometer-0\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.767393 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq4qd\" (UniqueName: \"kubernetes.io/projected/cbf480e0-f3e9-445c-a174-db6ce0967ed7-kube-api-access-sq4qd\") pod \"dnsmasq-dns-56df8fb6b7-lvp7w\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.768095 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrq6q\" (UniqueName: \"kubernetes.io/projected/55ec5b52-2985-4607-8c89-590f43a3096c-kube-api-access-wrq6q\") pod \"ceilometer-0\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.768186 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-config-data\") pod \"ceilometer-0\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.768204 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.768234 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-config\") pod \"dnsmasq-dns-56df8fb6b7-lvp7w\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.768261 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-lvp7w\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.768318 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-scripts\") pod \"ceilometer-0\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.768333 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55ec5b52-2985-4607-8c89-590f43a3096c-log-httpd\") pod \"ceilometer-0\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.768924 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55ec5b52-2985-4607-8c89-590f43a3096c-log-httpd\") pod \"ceilometer-0\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.769416 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-lvp7w\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.770106 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-lvp7w\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.771765 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-lvp7w\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.772684 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55ec5b52-2985-4607-8c89-590f43a3096c-run-httpd\") pod \"ceilometer-0\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.773067 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-config-data\") pod \"ceilometer-0\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.773356 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-lvp7w\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.776504 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-config\") pod \"dnsmasq-dns-56df8fb6b7-lvp7w\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.780874 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.781196 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.782694 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-scripts\") pod \"ceilometer-0\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.791423 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq4qd\" (UniqueName: \"kubernetes.io/projected/cbf480e0-f3e9-445c-a174-db6ce0967ed7-kube-api-access-sq4qd\") pod \"dnsmasq-dns-56df8fb6b7-lvp7w\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.792396 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrq6q\" (UniqueName: \"kubernetes.io/projected/55ec5b52-2985-4607-8c89-590f43a3096c-kube-api-access-wrq6q\") pod \"ceilometer-0\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " pod="openstack/ceilometer-0" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.863478 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b89002b9-75d6-4868-9148-d3d9d8c71db7" path="/var/lib/kubelet/pods/b89002b9-75d6-4868-9148-d3d9d8c71db7/volumes" Nov 26 00:45:03 crc kubenswrapper[4766]: I1126 00:45:03.994664 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-sbhxb"] Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:03.997955 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-7rfnj" Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.009305 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.022328 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.231007 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rg6jq"] Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.356962 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-8sbgx"] Nov 26 00:45:04 crc kubenswrapper[4766]: W1126 00:45:04.395265 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fd45623_8db1_440a_960a_a848abbb0c90.slice/crio-b7e5dce76135b8a76f33cb2f4036060c4d8bb466f033813e20b84ca3202f7d5d WatchSource:0}: Error finding container b7e5dce76135b8a76f33cb2f4036060c4d8bb466f033813e20b84ca3202f7d5d: Status 404 returned error can't find the container with id b7e5dce76135b8a76f33cb2f4036060c4d8bb466f033813e20b84ca3202f7d5d Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.409505 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-9259k"] Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.627751 4766 generic.go:334] "Generic (PLEG): container finished" podID="edf21578-364f-472d-99bd-ec283f2fb585" containerID="a84c1bafa4f7d273b2c90f4e3a2fa8e419d41d8435b7399ac6df6cee0ee73e4e" exitCode=0 Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.628208 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" event={"ID":"edf21578-364f-472d-99bd-ec283f2fb585","Type":"ContainerDied","Data":"a84c1bafa4f7d273b2c90f4e3a2fa8e419d41d8435b7399ac6df6cee0ee73e4e"} Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.628247 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" event={"ID":"edf21578-364f-472d-99bd-ec283f2fb585","Type":"ContainerStarted","Data":"9f56553f6d699a5d2a649b714b87251ac73120ba5b79310aac335cb8c01c139b"} Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.629806 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.632673 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rg6jq" event={"ID":"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a","Type":"ContainerStarted","Data":"a27ff9610e9d1fa93dfe0a0cc7de5373411722db90194c0496848270840c5d20"} Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.635213 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-9259k" event={"ID":"2fd45623-8db1-440a-960a-a848abbb0c90","Type":"ContainerStarted","Data":"b7e5dce76135b8a76f33cb2f4036060c4d8bb466f033813e20b84ca3202f7d5d"} Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.649610 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8sbgx" event={"ID":"ffaf6d04-ed5d-4438-a285-bc4981448be5","Type":"ContainerStarted","Data":"b67e7262cd6bc785308ebef5bbf9eb60677a4390794ea24c07d3fe0790938608"} Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.657461 4766 generic.go:334] "Generic (PLEG): container finished" podID="eb6091b5-cc82-44cf-93cd-825dd952dbb3" containerID="77371af89f48688fc7cf2d1aa9c52f2692ea30ca7db0d6479b26509911a3c9b0" exitCode=0 Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.657530 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" event={"ID":"eb6091b5-cc82-44cf-93cd-825dd952dbb3","Type":"ContainerDied","Data":"77371af89f48688fc7cf2d1aa9c52f2692ea30ca7db0d6479b26509911a3c9b0"} Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.657576 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" event={"ID":"eb6091b5-cc82-44cf-93cd-825dd952dbb3","Type":"ContainerDied","Data":"0a315a25146024c38c57e119df498cbf09855ad8c814e61d8d24fec0d578e933"} Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.657599 4766 scope.go:117] "RemoveContainer" containerID="77371af89f48688fc7cf2d1aa9c52f2692ea30ca7db0d6479b26509911a3c9b0" Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.657829 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-pj25t" Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.690770 4766 scope.go:117] "RemoveContainer" containerID="3d37dbe375032d57ea58c3c0cb53002b299b4aa1db9eaf7cb6ad313046dc6c6e" Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.703867 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-ovsdbserver-sb\") pod \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.703973 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-dns-swift-storage-0\") pod \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.704004 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-dns-svc\") pod \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.704318 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26x8f\" (UniqueName: \"kubernetes.io/projected/eb6091b5-cc82-44cf-93cd-825dd952dbb3-kube-api-access-26x8f\") pod \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.704435 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-ovsdbserver-nb\") pod \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.704559 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-config\") pod \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\" (UID: \"eb6091b5-cc82-44cf-93cd-825dd952dbb3\") " Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.727783 4766 scope.go:117] "RemoveContainer" containerID="77371af89f48688fc7cf2d1aa9c52f2692ea30ca7db0d6479b26509911a3c9b0" Nov 26 00:45:04 crc kubenswrapper[4766]: E1126 00:45:04.728327 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77371af89f48688fc7cf2d1aa9c52f2692ea30ca7db0d6479b26509911a3c9b0\": container with ID starting with 77371af89f48688fc7cf2d1aa9c52f2692ea30ca7db0d6479b26509911a3c9b0 not found: ID does not exist" containerID="77371af89f48688fc7cf2d1aa9c52f2692ea30ca7db0d6479b26509911a3c9b0" Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.728365 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77371af89f48688fc7cf2d1aa9c52f2692ea30ca7db0d6479b26509911a3c9b0"} err="failed to get container status \"77371af89f48688fc7cf2d1aa9c52f2692ea30ca7db0d6479b26509911a3c9b0\": rpc error: code = NotFound desc = could not find container \"77371af89f48688fc7cf2d1aa9c52f2692ea30ca7db0d6479b26509911a3c9b0\": container with ID starting with 77371af89f48688fc7cf2d1aa9c52f2692ea30ca7db0d6479b26509911a3c9b0 not found: ID does not exist" Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.728390 4766 scope.go:117] "RemoveContainer" containerID="3d37dbe375032d57ea58c3c0cb53002b299b4aa1db9eaf7cb6ad313046dc6c6e" Nov 26 00:45:04 crc kubenswrapper[4766]: E1126 00:45:04.728943 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d37dbe375032d57ea58c3c0cb53002b299b4aa1db9eaf7cb6ad313046dc6c6e\": container with ID starting with 3d37dbe375032d57ea58c3c0cb53002b299b4aa1db9eaf7cb6ad313046dc6c6e not found: ID does not exist" containerID="3d37dbe375032d57ea58c3c0cb53002b299b4aa1db9eaf7cb6ad313046dc6c6e" Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.729018 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d37dbe375032d57ea58c3c0cb53002b299b4aa1db9eaf7cb6ad313046dc6c6e"} err="failed to get container status \"3d37dbe375032d57ea58c3c0cb53002b299b4aa1db9eaf7cb6ad313046dc6c6e\": rpc error: code = NotFound desc = could not find container \"3d37dbe375032d57ea58c3c0cb53002b299b4aa1db9eaf7cb6ad313046dc6c6e\": container with ID starting with 3d37dbe375032d57ea58c3c0cb53002b299b4aa1db9eaf7cb6ad313046dc6c6e not found: ID does not exist" Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.734104 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb6091b5-cc82-44cf-93cd-825dd952dbb3-kube-api-access-26x8f" (OuterVolumeSpecName: "kube-api-access-26x8f") pod "eb6091b5-cc82-44cf-93cd-825dd952dbb3" (UID: "eb6091b5-cc82-44cf-93cd-825dd952dbb3"). InnerVolumeSpecName "kube-api-access-26x8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.779200 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-8bfjz"] Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.792696 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-pd2t2"] Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.808097 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26x8f\" (UniqueName: \"kubernetes.io/projected/eb6091b5-cc82-44cf-93cd-825dd952dbb3-kube-api-access-26x8f\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.952153 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-config" (OuterVolumeSpecName: "config") pod "eb6091b5-cc82-44cf-93cd-825dd952dbb3" (UID: "eb6091b5-cc82-44cf-93cd-825dd952dbb3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.976256 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eb6091b5-cc82-44cf-93cd-825dd952dbb3" (UID: "eb6091b5-cc82-44cf-93cd-825dd952dbb3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.978849 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eb6091b5-cc82-44cf-93cd-825dd952dbb3" (UID: "eb6091b5-cc82-44cf-93cd-825dd952dbb3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:04 crc kubenswrapper[4766]: I1126 00:45:04.993231 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eb6091b5-cc82-44cf-93cd-825dd952dbb3" (UID: "eb6091b5-cc82-44cf-93cd-825dd952dbb3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.008415 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "eb6091b5-cc82-44cf-93cd-825dd952dbb3" (UID: "eb6091b5-cc82-44cf-93cd-825dd952dbb3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.015767 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.015802 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.015814 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.015825 4766 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.015841 4766 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb6091b5-cc82-44cf-93cd-825dd952dbb3-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.134798 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-7rfnj"] Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.147308 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.194247 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-lvp7w"] Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.303143 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.319866 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-dns-swift-storage-0\") pod \"edf21578-364f-472d-99bd-ec283f2fb585\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.321302 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-config\") pod \"edf21578-364f-472d-99bd-ec283f2fb585\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.321366 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-dns-svc\") pod \"edf21578-364f-472d-99bd-ec283f2fb585\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.321887 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-ovsdbserver-sb\") pod \"edf21578-364f-472d-99bd-ec283f2fb585\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.322096 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqcmd\" (UniqueName: \"kubernetes.io/projected/edf21578-364f-472d-99bd-ec283f2fb585-kube-api-access-wqcmd\") pod \"edf21578-364f-472d-99bd-ec283f2fb585\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.322117 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-ovsdbserver-nb\") pod \"edf21578-364f-472d-99bd-ec283f2fb585\" (UID: \"edf21578-364f-472d-99bd-ec283f2fb585\") " Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.338493 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-pj25t"] Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.345232 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edf21578-364f-472d-99bd-ec283f2fb585-kube-api-access-wqcmd" (OuterVolumeSpecName: "kube-api-access-wqcmd") pod "edf21578-364f-472d-99bd-ec283f2fb585" (UID: "edf21578-364f-472d-99bd-ec283f2fb585"). InnerVolumeSpecName "kube-api-access-wqcmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.369617 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "edf21578-364f-472d-99bd-ec283f2fb585" (UID: "edf21578-364f-472d-99bd-ec283f2fb585"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.388168 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-config" (OuterVolumeSpecName: "config") pod "edf21578-364f-472d-99bd-ec283f2fb585" (UID: "edf21578-364f-472d-99bd-ec283f2fb585"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.390165 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "edf21578-364f-472d-99bd-ec283f2fb585" (UID: "edf21578-364f-472d-99bd-ec283f2fb585"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.399681 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "edf21578-364f-472d-99bd-ec283f2fb585" (UID: "edf21578-364f-472d-99bd-ec283f2fb585"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.408921 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "edf21578-364f-472d-99bd-ec283f2fb585" (UID: "edf21578-364f-472d-99bd-ec283f2fb585"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.415263 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-pj25t"] Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.427258 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqcmd\" (UniqueName: \"kubernetes.io/projected/edf21578-364f-472d-99bd-ec283f2fb585-kube-api-access-wqcmd\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.427296 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.427309 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.427320 4766 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.427333 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.427343 4766 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edf21578-364f-472d-99bd-ec283f2fb585-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.698465 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" event={"ID":"edf21578-364f-472d-99bd-ec283f2fb585","Type":"ContainerDied","Data":"9f56553f6d699a5d2a649b714b87251ac73120ba5b79310aac335cb8c01c139b"} Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.698816 4766 scope.go:117] "RemoveContainer" containerID="a84c1bafa4f7d273b2c90f4e3a2fa8e419d41d8435b7399ac6df6cee0ee73e4e" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.698855 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-sbhxb" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.713228 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rg6jq" event={"ID":"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a","Type":"ContainerStarted","Data":"7f8c888a919d7a6d73f8981f0fc18c2ff35464b5c11779cedf8e8fa696efdfd9"} Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.719467 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-8bfjz" event={"ID":"b156a1fc-37a3-45d3-999d-55852f10bcc3","Type":"ContainerStarted","Data":"3718f9376f2bea3d70776252e88a7eb5fcc27e2ec33341a258d4bd09f79aac4c"} Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.719523 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-8bfjz" event={"ID":"b156a1fc-37a3-45d3-999d-55852f10bcc3","Type":"ContainerStarted","Data":"5d4426e8d391ff8b8db03240735addc9f82377051e936ec61fb3302e82a9f27d"} Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.734848 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55ec5b52-2985-4607-8c89-590f43a3096c","Type":"ContainerStarted","Data":"8fd5a36c2222314544c095bc82df4d571ffde4f52c7fc264fa3943fbc8a09444"} Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.737421 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-7rfnj" event={"ID":"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8","Type":"ContainerStarted","Data":"2ad21a7440118caad6624fadba329df97067115512ca455069184e249155de32"} Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.740271 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pd2t2" event={"ID":"66feb239-5790-4cfe-96db-e9dd6413bb09","Type":"ContainerStarted","Data":"7517af61f45f00e8296e55632d0df0eb107f2dae81291e52ecf13564d8e891a2"} Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.802959 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-rg6jq" podStartSLOduration=3.802933208 podStartE2EDuration="3.802933208s" podCreationTimestamp="2025-11-26 00:45:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:45:05.794953378 +0000 UTC m=+1286.643723808" watchObservedRunningTime="2025-11-26 00:45:05.802933208 +0000 UTC m=+1286.651703648" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.822368 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" event={"ID":"cbf480e0-f3e9-445c-a174-db6ce0967ed7","Type":"ContainerStarted","Data":"eccfbc38abe6f8428984fb53b569287a7eb4bd7a3e2d408b5b1a65fbef90408c"} Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.822411 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" event={"ID":"cbf480e0-f3e9-445c-a174-db6ce0967ed7","Type":"ContainerStarted","Data":"570236219f9aaae0cbf842f0e3c63696c270c9112022bb6fb595755785a0facd"} Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.842408 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-8bfjz" podStartSLOduration=2.842389258 podStartE2EDuration="2.842389258s" podCreationTimestamp="2025-11-26 00:45:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:45:05.835280699 +0000 UTC m=+1286.684051129" watchObservedRunningTime="2025-11-26 00:45:05.842389258 +0000 UTC m=+1286.691159688" Nov 26 00:45:05 crc kubenswrapper[4766]: I1126 00:45:05.887221 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb6091b5-cc82-44cf-93cd-825dd952dbb3" path="/var/lib/kubelet/pods/eb6091b5-cc82-44cf-93cd-825dd952dbb3/volumes" Nov 26 00:45:06 crc kubenswrapper[4766]: I1126 00:45:06.058980 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-sbhxb"] Nov 26 00:45:06 crc kubenswrapper[4766]: I1126 00:45:06.073916 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-sbhxb"] Nov 26 00:45:06 crc kubenswrapper[4766]: I1126 00:45:06.329896 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:45:06 crc kubenswrapper[4766]: I1126 00:45:06.844459 4766 generic.go:334] "Generic (PLEG): container finished" podID="cbf480e0-f3e9-445c-a174-db6ce0967ed7" containerID="eccfbc38abe6f8428984fb53b569287a7eb4bd7a3e2d408b5b1a65fbef90408c" exitCode=0 Nov 26 00:45:06 crc kubenswrapper[4766]: I1126 00:45:06.844911 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" event={"ID":"cbf480e0-f3e9-445c-a174-db6ce0967ed7","Type":"ContainerDied","Data":"eccfbc38abe6f8428984fb53b569287a7eb4bd7a3e2d408b5b1a65fbef90408c"} Nov 26 00:45:06 crc kubenswrapper[4766]: I1126 00:45:06.844944 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" event={"ID":"cbf480e0-f3e9-445c-a174-db6ce0967ed7","Type":"ContainerStarted","Data":"467596d52fa9c032644d706448c18ccfc8ef04c98600e4de8d35a521162c60b2"} Nov 26 00:45:06 crc kubenswrapper[4766]: I1126 00:45:06.846174 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:06 crc kubenswrapper[4766]: I1126 00:45:06.868370 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" podStartSLOduration=3.868352217 podStartE2EDuration="3.868352217s" podCreationTimestamp="2025-11-26 00:45:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:45:06.866300586 +0000 UTC m=+1287.715071016" watchObservedRunningTime="2025-11-26 00:45:06.868352217 +0000 UTC m=+1287.717122647" Nov 26 00:45:07 crc kubenswrapper[4766]: I1126 00:45:07.848051 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edf21578-364f-472d-99bd-ec283f2fb585" path="/var/lib/kubelet/pods/edf21578-364f-472d-99bd-ec283f2fb585/volumes" Nov 26 00:45:09 crc kubenswrapper[4766]: I1126 00:45:09.882178 4766 generic.go:334] "Generic (PLEG): container finished" podID="0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a" containerID="7f8c888a919d7a6d73f8981f0fc18c2ff35464b5c11779cedf8e8fa696efdfd9" exitCode=0 Nov 26 00:45:09 crc kubenswrapper[4766]: I1126 00:45:09.882208 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rg6jq" event={"ID":"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a","Type":"ContainerDied","Data":"7f8c888a919d7a6d73f8981f0fc18c2ff35464b5c11779cedf8e8fa696efdfd9"} Nov 26 00:45:11 crc kubenswrapper[4766]: I1126 00:45:11.479935 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:45:11 crc kubenswrapper[4766]: I1126 00:45:11.480302 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:45:11 crc kubenswrapper[4766]: I1126 00:45:11.849062 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 26 00:45:11 crc kubenswrapper[4766]: I1126 00:45:11.856550 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 26 00:45:11 crc kubenswrapper[4766]: I1126 00:45:11.907391 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 26 00:45:14 crc kubenswrapper[4766]: I1126 00:45:14.024489 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:14 crc kubenswrapper[4766]: I1126 00:45:14.079570 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-2wx6s"] Nov 26 00:45:14 crc kubenswrapper[4766]: I1126 00:45:14.080185 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" podUID="c4c2347f-5488-4f85-ab8d-c0a9324c1179" containerName="dnsmasq-dns" containerID="cri-o://968b54683480167549f5590c4888d75550afea27a313610b67f89220484f4e73" gracePeriod=10 Nov 26 00:45:14 crc kubenswrapper[4766]: I1126 00:45:14.950403 4766 generic.go:334] "Generic (PLEG): container finished" podID="c4c2347f-5488-4f85-ab8d-c0a9324c1179" containerID="968b54683480167549f5590c4888d75550afea27a313610b67f89220484f4e73" exitCode=0 Nov 26 00:45:14 crc kubenswrapper[4766]: I1126 00:45:14.950444 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" event={"ID":"c4c2347f-5488-4f85-ab8d-c0a9324c1179","Type":"ContainerDied","Data":"968b54683480167549f5590c4888d75550afea27a313610b67f89220484f4e73"} Nov 26 00:45:15 crc kubenswrapper[4766]: I1126 00:45:15.677363 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" podUID="c4c2347f-5488-4f85-ab8d-c0a9324c1179" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.142:5353: connect: connection refused" Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.177125 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.325714 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-credential-keys\") pod \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.326158 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-scripts\") pod \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.326282 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-fernet-keys\") pod \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.326416 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-config-data\") pod \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.326609 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-combined-ca-bundle\") pod \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.326811 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbg98\" (UniqueName: \"kubernetes.io/projected/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-kube-api-access-wbg98\") pod \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\" (UID: \"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a\") " Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.331140 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a" (UID: "0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.331502 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-scripts" (OuterVolumeSpecName: "scripts") pod "0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a" (UID: "0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.333486 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a" (UID: "0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.335416 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-kube-api-access-wbg98" (OuterVolumeSpecName: "kube-api-access-wbg98") pod "0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a" (UID: "0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a"). InnerVolumeSpecName "kube-api-access-wbg98". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.358834 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a" (UID: "0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.362087 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-config-data" (OuterVolumeSpecName: "config-data") pod "0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a" (UID: "0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.429536 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbg98\" (UniqueName: \"kubernetes.io/projected/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-kube-api-access-wbg98\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.429571 4766 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.429584 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.429596 4766 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.429606 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.429621 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:18 crc kubenswrapper[4766]: E1126 00:45:18.652610 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Nov 26 00:45:18 crc kubenswrapper[4766]: E1126 00:45:18.653461 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r568f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-pd2t2_openstack(66feb239-5790-4cfe-96db-e9dd6413bb09): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 00:45:18 crc kubenswrapper[4766]: E1126 00:45:18.654744 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-pd2t2" podUID="66feb239-5790-4cfe-96db-e9dd6413bb09" Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.996203 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rg6jq" Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.996291 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rg6jq" event={"ID":"0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a","Type":"ContainerDied","Data":"a27ff9610e9d1fa93dfe0a0cc7de5373411722db90194c0496848270840c5d20"} Nov 26 00:45:18 crc kubenswrapper[4766]: I1126 00:45:18.996342 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a27ff9610e9d1fa93dfe0a0cc7de5373411722db90194c0496848270840c5d20" Nov 26 00:45:18 crc kubenswrapper[4766]: E1126 00:45:18.997717 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-pd2t2" podUID="66feb239-5790-4cfe-96db-e9dd6413bb09" Nov 26 00:45:19 crc kubenswrapper[4766]: E1126 00:45:19.080635 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Nov 26 00:45:19 crc kubenswrapper[4766]: E1126 00:45:19.080829 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n678h698h555h54ch548h669h56dh9bh698h657h5b4h59hc7h5bfh6bh574h89h9bh6h56ch7hf5h64h576hb9h646h574hb8h588hd9hc6h669q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wrq6q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(55ec5b52-2985-4607-8c89-590f43a3096c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.263362 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-rg6jq"] Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.270629 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-rg6jq"] Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.363283 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-d26z8"] Nov 26 00:45:19 crc kubenswrapper[4766]: E1126 00:45:19.363983 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edf21578-364f-472d-99bd-ec283f2fb585" containerName="init" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.363999 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="edf21578-364f-472d-99bd-ec283f2fb585" containerName="init" Nov 26 00:45:19 crc kubenswrapper[4766]: E1126 00:45:19.364018 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb6091b5-cc82-44cf-93cd-825dd952dbb3" containerName="init" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.364025 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb6091b5-cc82-44cf-93cd-825dd952dbb3" containerName="init" Nov 26 00:45:19 crc kubenswrapper[4766]: E1126 00:45:19.364036 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb6091b5-cc82-44cf-93cd-825dd952dbb3" containerName="dnsmasq-dns" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.364042 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb6091b5-cc82-44cf-93cd-825dd952dbb3" containerName="dnsmasq-dns" Nov 26 00:45:19 crc kubenswrapper[4766]: E1126 00:45:19.364063 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a" containerName="keystone-bootstrap" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.364069 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a" containerName="keystone-bootstrap" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.364237 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="edf21578-364f-472d-99bd-ec283f2fb585" containerName="init" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.364258 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a" containerName="keystone-bootstrap" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.364269 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb6091b5-cc82-44cf-93cd-825dd952dbb3" containerName="dnsmasq-dns" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.364994 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.367168 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.367194 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ksmzm" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.370538 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-d26z8"] Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.417025 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.417175 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.417220 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.461249 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-config-data\") pod \"keystone-bootstrap-d26z8\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.461362 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-combined-ca-bundle\") pod \"keystone-bootstrap-d26z8\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.461394 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-credential-keys\") pod \"keystone-bootstrap-d26z8\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.461460 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-fernet-keys\") pod \"keystone-bootstrap-d26z8\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.461493 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-scripts\") pod \"keystone-bootstrap-d26z8\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.461606 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znqbw\" (UniqueName: \"kubernetes.io/projected/6969780f-beaf-4216-8516-1eea9445f0c4-kube-api-access-znqbw\") pod \"keystone-bootstrap-d26z8\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.563406 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-fernet-keys\") pod \"keystone-bootstrap-d26z8\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.563475 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-scripts\") pod \"keystone-bootstrap-d26z8\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.563543 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znqbw\" (UniqueName: \"kubernetes.io/projected/6969780f-beaf-4216-8516-1eea9445f0c4-kube-api-access-znqbw\") pod \"keystone-bootstrap-d26z8\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.563948 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-config-data\") pod \"keystone-bootstrap-d26z8\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.564038 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-combined-ca-bundle\") pod \"keystone-bootstrap-d26z8\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.564065 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-credential-keys\") pod \"keystone-bootstrap-d26z8\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.570271 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-combined-ca-bundle\") pod \"keystone-bootstrap-d26z8\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.570632 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-fernet-keys\") pod \"keystone-bootstrap-d26z8\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.571139 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-scripts\") pod \"keystone-bootstrap-d26z8\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.573050 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-credential-keys\") pod \"keystone-bootstrap-d26z8\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.581060 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-config-data\") pod \"keystone-bootstrap-d26z8\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.581309 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znqbw\" (UniqueName: \"kubernetes.io/projected/6969780f-beaf-4216-8516-1eea9445f0c4-kube-api-access-znqbw\") pod \"keystone-bootstrap-d26z8\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.735735 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:19 crc kubenswrapper[4766]: I1126 00:45:19.839210 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a" path="/var/lib/kubelet/pods/0fbfc54b-a51b-4cab-95dd-4d031bf8aa1a/volumes" Nov 26 00:45:25 crc kubenswrapper[4766]: I1126 00:45:25.677661 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" podUID="c4c2347f-5488-4f85-ab8d-c0a9324c1179" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.142:5353: i/o timeout" Nov 26 00:45:28 crc kubenswrapper[4766]: I1126 00:45:28.312319 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" Nov 26 00:45:28 crc kubenswrapper[4766]: I1126 00:45:28.435643 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-ovsdbserver-nb\") pod \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\" (UID: \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\") " Nov 26 00:45:28 crc kubenswrapper[4766]: I1126 00:45:28.436185 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-ovsdbserver-sb\") pod \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\" (UID: \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\") " Nov 26 00:45:28 crc kubenswrapper[4766]: I1126 00:45:28.436240 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxwmw\" (UniqueName: \"kubernetes.io/projected/c4c2347f-5488-4f85-ab8d-c0a9324c1179-kube-api-access-fxwmw\") pod \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\" (UID: \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\") " Nov 26 00:45:28 crc kubenswrapper[4766]: I1126 00:45:28.436268 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-config\") pod \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\" (UID: \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\") " Nov 26 00:45:28 crc kubenswrapper[4766]: I1126 00:45:28.436318 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-dns-svc\") pod \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\" (UID: \"c4c2347f-5488-4f85-ab8d-c0a9324c1179\") " Nov 26 00:45:28 crc kubenswrapper[4766]: I1126 00:45:28.440421 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4c2347f-5488-4f85-ab8d-c0a9324c1179-kube-api-access-fxwmw" (OuterVolumeSpecName: "kube-api-access-fxwmw") pod "c4c2347f-5488-4f85-ab8d-c0a9324c1179" (UID: "c4c2347f-5488-4f85-ab8d-c0a9324c1179"). InnerVolumeSpecName "kube-api-access-fxwmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:45:28 crc kubenswrapper[4766]: I1126 00:45:28.482569 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c4c2347f-5488-4f85-ab8d-c0a9324c1179" (UID: "c4c2347f-5488-4f85-ab8d-c0a9324c1179"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:28 crc kubenswrapper[4766]: I1126 00:45:28.483617 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c4c2347f-5488-4f85-ab8d-c0a9324c1179" (UID: "c4c2347f-5488-4f85-ab8d-c0a9324c1179"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:28 crc kubenswrapper[4766]: I1126 00:45:28.488916 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-config" (OuterVolumeSpecName: "config") pod "c4c2347f-5488-4f85-ab8d-c0a9324c1179" (UID: "c4c2347f-5488-4f85-ab8d-c0a9324c1179"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:28 crc kubenswrapper[4766]: I1126 00:45:28.491107 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c4c2347f-5488-4f85-ab8d-c0a9324c1179" (UID: "c4c2347f-5488-4f85-ab8d-c0a9324c1179"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:28 crc kubenswrapper[4766]: I1126 00:45:28.538464 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:28 crc kubenswrapper[4766]: I1126 00:45:28.538508 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxwmw\" (UniqueName: \"kubernetes.io/projected/c4c2347f-5488-4f85-ab8d-c0a9324c1179-kube-api-access-fxwmw\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:28 crc kubenswrapper[4766]: I1126 00:45:28.538527 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:28 crc kubenswrapper[4766]: I1126 00:45:28.538566 4766 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:28 crc kubenswrapper[4766]: I1126 00:45:28.538583 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4c2347f-5488-4f85-ab8d-c0a9324c1179-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:29 crc kubenswrapper[4766]: I1126 00:45:29.099704 4766 generic.go:334] "Generic (PLEG): container finished" podID="b156a1fc-37a3-45d3-999d-55852f10bcc3" containerID="3718f9376f2bea3d70776252e88a7eb5fcc27e2ec33341a258d4bd09f79aac4c" exitCode=0 Nov 26 00:45:29 crc kubenswrapper[4766]: I1126 00:45:29.099755 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-8bfjz" event={"ID":"b156a1fc-37a3-45d3-999d-55852f10bcc3","Type":"ContainerDied","Data":"3718f9376f2bea3d70776252e88a7eb5fcc27e2ec33341a258d4bd09f79aac4c"} Nov 26 00:45:29 crc kubenswrapper[4766]: I1126 00:45:29.102635 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" event={"ID":"c4c2347f-5488-4f85-ab8d-c0a9324c1179","Type":"ContainerDied","Data":"c648fdbb68fe5ed5ad4a8b58910220f7c023bc3cc36cb6abe2d7a11f373afae4"} Nov 26 00:45:29 crc kubenswrapper[4766]: I1126 00:45:29.102746 4766 scope.go:117] "RemoveContainer" containerID="968b54683480167549f5590c4888d75550afea27a313610b67f89220484f4e73" Nov 26 00:45:29 crc kubenswrapper[4766]: I1126 00:45:29.102994 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" Nov 26 00:45:29 crc kubenswrapper[4766]: I1126 00:45:29.150000 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-2wx6s"] Nov 26 00:45:29 crc kubenswrapper[4766]: I1126 00:45:29.161235 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-2wx6s"] Nov 26 00:45:29 crc kubenswrapper[4766]: I1126 00:45:29.470118 4766 scope.go:117] "RemoveContainer" containerID="c21a4ede7b6d1408e7ae84444e13c7a43bac3b60340457338c7d5780e812e108" Nov 26 00:45:29 crc kubenswrapper[4766]: E1126 00:45:29.510339 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 26 00:45:29 crc kubenswrapper[4766]: E1126 00:45:29.510476 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jgxb5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-8sbgx_openstack(ffaf6d04-ed5d-4438-a285-bc4981448be5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 00:45:29 crc kubenswrapper[4766]: E1126 00:45:29.512364 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-8sbgx" podUID="ffaf6d04-ed5d-4438-a285-bc4981448be5" Nov 26 00:45:29 crc kubenswrapper[4766]: I1126 00:45:29.845409 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4c2347f-5488-4f85-ab8d-c0a9324c1179" path="/var/lib/kubelet/pods/c4c2347f-5488-4f85-ab8d-c0a9324c1179/volumes" Nov 26 00:45:30 crc kubenswrapper[4766]: I1126 00:45:30.111832 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-9259k" event={"ID":"2fd45623-8db1-440a-960a-a848abbb0c90","Type":"ContainerStarted","Data":"dd0a910ba1ed6f7f52274d8aba3d44c06ccdc43b8bfc50f731aad0b2322b2405"} Nov 26 00:45:30 crc kubenswrapper[4766]: I1126 00:45:30.114827 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55ec5b52-2985-4607-8c89-590f43a3096c","Type":"ContainerStarted","Data":"fe436118dd6795c5727832d6353c09cd384307b2bc3e597fdd9639d3635552cd"} Nov 26 00:45:30 crc kubenswrapper[4766]: I1126 00:45:30.115794 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-7rfnj" event={"ID":"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8","Type":"ContainerStarted","Data":"530f9b909755d33d9f92a02281b30b40fd91c52515c3b64e26f9b0b5e00801c5"} Nov 26 00:45:30 crc kubenswrapper[4766]: E1126 00:45:30.117075 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-8sbgx" podUID="ffaf6d04-ed5d-4438-a285-bc4981448be5" Nov 26 00:45:30 crc kubenswrapper[4766]: I1126 00:45:30.133341 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-9259k" podStartSLOduration=3.048876812 podStartE2EDuration="28.133323339s" podCreationTimestamp="2025-11-26 00:45:02 +0000 UTC" firstStartedPulling="2025-11-26 00:45:04.407338399 +0000 UTC m=+1285.256108829" lastFinishedPulling="2025-11-26 00:45:29.491784926 +0000 UTC m=+1310.340555356" observedRunningTime="2025-11-26 00:45:30.129459382 +0000 UTC m=+1310.978229812" watchObservedRunningTime="2025-11-26 00:45:30.133323339 +0000 UTC m=+1310.982093759" Nov 26 00:45:30 crc kubenswrapper[4766]: I1126 00:45:30.154228 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-7rfnj" podStartSLOduration=2.801774564 podStartE2EDuration="27.154204833s" podCreationTimestamp="2025-11-26 00:45:03 +0000 UTC" firstStartedPulling="2025-11-26 00:45:05.104422841 +0000 UTC m=+1285.953193271" lastFinishedPulling="2025-11-26 00:45:29.45685311 +0000 UTC m=+1310.305623540" observedRunningTime="2025-11-26 00:45:30.149471924 +0000 UTC m=+1310.998242354" watchObservedRunningTime="2025-11-26 00:45:30.154204833 +0000 UTC m=+1311.002975263" Nov 26 00:45:30 crc kubenswrapper[4766]: I1126 00:45:30.216333 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-d26z8"] Nov 26 00:45:30 crc kubenswrapper[4766]: I1126 00:45:30.434563 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-8bfjz" Nov 26 00:45:30 crc kubenswrapper[4766]: I1126 00:45:30.599140 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b156a1fc-37a3-45d3-999d-55852f10bcc3-config\") pod \"b156a1fc-37a3-45d3-999d-55852f10bcc3\" (UID: \"b156a1fc-37a3-45d3-999d-55852f10bcc3\") " Nov 26 00:45:30 crc kubenswrapper[4766]: I1126 00:45:30.599260 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8g58\" (UniqueName: \"kubernetes.io/projected/b156a1fc-37a3-45d3-999d-55852f10bcc3-kube-api-access-t8g58\") pod \"b156a1fc-37a3-45d3-999d-55852f10bcc3\" (UID: \"b156a1fc-37a3-45d3-999d-55852f10bcc3\") " Nov 26 00:45:30 crc kubenswrapper[4766]: I1126 00:45:30.599355 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b156a1fc-37a3-45d3-999d-55852f10bcc3-combined-ca-bundle\") pod \"b156a1fc-37a3-45d3-999d-55852f10bcc3\" (UID: \"b156a1fc-37a3-45d3-999d-55852f10bcc3\") " Nov 26 00:45:30 crc kubenswrapper[4766]: I1126 00:45:30.604765 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b156a1fc-37a3-45d3-999d-55852f10bcc3-kube-api-access-t8g58" (OuterVolumeSpecName: "kube-api-access-t8g58") pod "b156a1fc-37a3-45d3-999d-55852f10bcc3" (UID: "b156a1fc-37a3-45d3-999d-55852f10bcc3"). InnerVolumeSpecName "kube-api-access-t8g58". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:45:30 crc kubenswrapper[4766]: I1126 00:45:30.630894 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b156a1fc-37a3-45d3-999d-55852f10bcc3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b156a1fc-37a3-45d3-999d-55852f10bcc3" (UID: "b156a1fc-37a3-45d3-999d-55852f10bcc3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:30 crc kubenswrapper[4766]: I1126 00:45:30.633153 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b156a1fc-37a3-45d3-999d-55852f10bcc3-config" (OuterVolumeSpecName: "config") pod "b156a1fc-37a3-45d3-999d-55852f10bcc3" (UID: "b156a1fc-37a3-45d3-999d-55852f10bcc3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:30 crc kubenswrapper[4766]: I1126 00:45:30.678962 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-2wx6s" podUID="c4c2347f-5488-4f85-ab8d-c0a9324c1179" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.142:5353: i/o timeout" Nov 26 00:45:30 crc kubenswrapper[4766]: I1126 00:45:30.701709 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b156a1fc-37a3-45d3-999d-55852f10bcc3-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:30 crc kubenswrapper[4766]: I1126 00:45:30.701744 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8g58\" (UniqueName: \"kubernetes.io/projected/b156a1fc-37a3-45d3-999d-55852f10bcc3-kube-api-access-t8g58\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:30 crc kubenswrapper[4766]: I1126 00:45:30.701756 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b156a1fc-37a3-45d3-999d-55852f10bcc3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.128044 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-d26z8" event={"ID":"6969780f-beaf-4216-8516-1eea9445f0c4","Type":"ContainerStarted","Data":"e0731b84189b5cf1d81b460c924c87afd8f2e54e09bd83d8ffc0da67f42e28fc"} Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.128603 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-d26z8" event={"ID":"6969780f-beaf-4216-8516-1eea9445f0c4","Type":"ContainerStarted","Data":"5589d7d2d769622a9977455a4bec4f19e6717ea21f785383fd6507dd30acf973"} Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.137228 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-8bfjz" event={"ID":"b156a1fc-37a3-45d3-999d-55852f10bcc3","Type":"ContainerDied","Data":"5d4426e8d391ff8b8db03240735addc9f82377051e936ec61fb3302e82a9f27d"} Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.137281 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d4426e8d391ff8b8db03240735addc9f82377051e936ec61fb3302e82a9f27d" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.137284 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-8bfjz" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.164513 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-d26z8" podStartSLOduration=12.164491879 podStartE2EDuration="12.164491879s" podCreationTimestamp="2025-11-26 00:45:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:45:31.149183845 +0000 UTC m=+1311.997954275" watchObservedRunningTime="2025-11-26 00:45:31.164491879 +0000 UTC m=+1312.013262309" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.446352 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-vtkhw"] Nov 26 00:45:31 crc kubenswrapper[4766]: E1126 00:45:31.448216 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4c2347f-5488-4f85-ab8d-c0a9324c1179" containerName="dnsmasq-dns" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.448301 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4c2347f-5488-4f85-ab8d-c0a9324c1179" containerName="dnsmasq-dns" Nov 26 00:45:31 crc kubenswrapper[4766]: E1126 00:45:31.448384 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b156a1fc-37a3-45d3-999d-55852f10bcc3" containerName="neutron-db-sync" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.448461 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="b156a1fc-37a3-45d3-999d-55852f10bcc3" containerName="neutron-db-sync" Nov 26 00:45:31 crc kubenswrapper[4766]: E1126 00:45:31.448518 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4c2347f-5488-4f85-ab8d-c0a9324c1179" containerName="init" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.448578 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4c2347f-5488-4f85-ab8d-c0a9324c1179" containerName="init" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.448910 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="b156a1fc-37a3-45d3-999d-55852f10bcc3" containerName="neutron-db-sync" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.448994 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4c2347f-5488-4f85-ab8d-c0a9324c1179" containerName="dnsmasq-dns" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.450094 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.462240 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-vtkhw"] Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.523542 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-vtkhw\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.523890 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-dns-svc\") pod \"dnsmasq-dns-6b7b667979-vtkhw\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.523923 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-vtkhw\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.523955 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-vtkhw\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.524038 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-config\") pod \"dnsmasq-dns-6b7b667979-vtkhw\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.524209 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f285\" (UniqueName: \"kubernetes.io/projected/5db2e99c-4e43-41e6-9ba2-331990b516bd-kube-api-access-8f285\") pod \"dnsmasq-dns-6b7b667979-vtkhw\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.555351 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7bbf79c884-hg8wv"] Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.561545 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bbf79c884-hg8wv" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.568414 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.568695 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-8w9h8" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.569249 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.569361 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.600410 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7bbf79c884-hg8wv"] Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.631354 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-combined-ca-bundle\") pod \"neutron-7bbf79c884-hg8wv\" (UID: \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\") " pod="openstack/neutron-7bbf79c884-hg8wv" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.631410 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-vtkhw\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.631446 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87f9q\" (UniqueName: \"kubernetes.io/projected/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-kube-api-access-87f9q\") pod \"neutron-7bbf79c884-hg8wv\" (UID: \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\") " pod="openstack/neutron-7bbf79c884-hg8wv" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.631473 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-dns-svc\") pod \"dnsmasq-dns-6b7b667979-vtkhw\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.631499 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-vtkhw\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.631529 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-vtkhw\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.631556 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-ovndb-tls-certs\") pod \"neutron-7bbf79c884-hg8wv\" (UID: \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\") " pod="openstack/neutron-7bbf79c884-hg8wv" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.631596 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-config\") pod \"dnsmasq-dns-6b7b667979-vtkhw\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.631689 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-config\") pod \"neutron-7bbf79c884-hg8wv\" (UID: \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\") " pod="openstack/neutron-7bbf79c884-hg8wv" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.631733 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f285\" (UniqueName: \"kubernetes.io/projected/5db2e99c-4e43-41e6-9ba2-331990b516bd-kube-api-access-8f285\") pod \"dnsmasq-dns-6b7b667979-vtkhw\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.631802 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-httpd-config\") pod \"neutron-7bbf79c884-hg8wv\" (UID: \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\") " pod="openstack/neutron-7bbf79c884-hg8wv" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.632724 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-vtkhw\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.634630 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-vtkhw\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.635903 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-dns-svc\") pod \"dnsmasq-dns-6b7b667979-vtkhw\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.636303 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-vtkhw\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.636615 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-config\") pod \"dnsmasq-dns-6b7b667979-vtkhw\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.658974 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f285\" (UniqueName: \"kubernetes.io/projected/5db2e99c-4e43-41e6-9ba2-331990b516bd-kube-api-access-8f285\") pod \"dnsmasq-dns-6b7b667979-vtkhw\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.733958 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-config\") pod \"neutron-7bbf79c884-hg8wv\" (UID: \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\") " pod="openstack/neutron-7bbf79c884-hg8wv" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.734166 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-httpd-config\") pod \"neutron-7bbf79c884-hg8wv\" (UID: \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\") " pod="openstack/neutron-7bbf79c884-hg8wv" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.735223 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-combined-ca-bundle\") pod \"neutron-7bbf79c884-hg8wv\" (UID: \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\") " pod="openstack/neutron-7bbf79c884-hg8wv" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.735294 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87f9q\" (UniqueName: \"kubernetes.io/projected/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-kube-api-access-87f9q\") pod \"neutron-7bbf79c884-hg8wv\" (UID: \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\") " pod="openstack/neutron-7bbf79c884-hg8wv" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.735365 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-ovndb-tls-certs\") pod \"neutron-7bbf79c884-hg8wv\" (UID: \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\") " pod="openstack/neutron-7bbf79c884-hg8wv" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.742575 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-ovndb-tls-certs\") pod \"neutron-7bbf79c884-hg8wv\" (UID: \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\") " pod="openstack/neutron-7bbf79c884-hg8wv" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.743070 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-combined-ca-bundle\") pod \"neutron-7bbf79c884-hg8wv\" (UID: \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\") " pod="openstack/neutron-7bbf79c884-hg8wv" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.743412 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-httpd-config\") pod \"neutron-7bbf79c884-hg8wv\" (UID: \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\") " pod="openstack/neutron-7bbf79c884-hg8wv" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.755270 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-config\") pod \"neutron-7bbf79c884-hg8wv\" (UID: \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\") " pod="openstack/neutron-7bbf79c884-hg8wv" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.764029 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87f9q\" (UniqueName: \"kubernetes.io/projected/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-kube-api-access-87f9q\") pod \"neutron-7bbf79c884-hg8wv\" (UID: \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\") " pod="openstack/neutron-7bbf79c884-hg8wv" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.794590 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:31 crc kubenswrapper[4766]: I1126 00:45:31.914872 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bbf79c884-hg8wv" Nov 26 00:45:32 crc kubenswrapper[4766]: I1126 00:45:32.191043 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pd2t2" event={"ID":"66feb239-5790-4cfe-96db-e9dd6413bb09","Type":"ContainerStarted","Data":"2a9815005b07139bc4c1db61fdce11fd291bd00028e65ab55088bfc31dbe5001"} Nov 26 00:45:32 crc kubenswrapper[4766]: I1126 00:45:32.212667 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-pd2t2" podStartSLOduration=2.771620576 podStartE2EDuration="29.208633066s" podCreationTimestamp="2025-11-26 00:45:03 +0000 UTC" firstStartedPulling="2025-11-26 00:45:04.850436932 +0000 UTC m=+1285.699207362" lastFinishedPulling="2025-11-26 00:45:31.287449422 +0000 UTC m=+1312.136219852" observedRunningTime="2025-11-26 00:45:32.20799864 +0000 UTC m=+1313.056769070" watchObservedRunningTime="2025-11-26 00:45:32.208633066 +0000 UTC m=+1313.057403486" Nov 26 00:45:32 crc kubenswrapper[4766]: I1126 00:45:32.459225 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-vtkhw"] Nov 26 00:45:32 crc kubenswrapper[4766]: I1126 00:45:32.658601 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7bbf79c884-hg8wv"] Nov 26 00:45:33 crc kubenswrapper[4766]: I1126 00:45:33.213201 4766 generic.go:334] "Generic (PLEG): container finished" podID="7d2fbed9-754d-45c7-a8c1-7dc4828b26a8" containerID="530f9b909755d33d9f92a02281b30b40fd91c52515c3b64e26f9b0b5e00801c5" exitCode=0 Nov 26 00:45:33 crc kubenswrapper[4766]: I1126 00:45:33.213430 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-7rfnj" event={"ID":"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8","Type":"ContainerDied","Data":"530f9b909755d33d9f92a02281b30b40fd91c52515c3b64e26f9b0b5e00801c5"} Nov 26 00:45:33 crc kubenswrapper[4766]: I1126 00:45:33.219227 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" event={"ID":"5db2e99c-4e43-41e6-9ba2-331990b516bd","Type":"ContainerStarted","Data":"0eb801770935205934c7a12573e12dc3138bf6ef9e19d3ea069d51a57c45fe2e"} Nov 26 00:45:33 crc kubenswrapper[4766]: I1126 00:45:33.779677 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-756fcfc485-ktbfm"] Nov 26 00:45:33 crc kubenswrapper[4766]: I1126 00:45:33.781262 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:33 crc kubenswrapper[4766]: I1126 00:45:33.784798 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 26 00:45:33 crc kubenswrapper[4766]: I1126 00:45:33.785208 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 26 00:45:33 crc kubenswrapper[4766]: I1126 00:45:33.801258 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-756fcfc485-ktbfm"] Nov 26 00:45:33 crc kubenswrapper[4766]: I1126 00:45:33.906217 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8d171003-64ac-4108-957b-70f4225359c5-config\") pod \"neutron-756fcfc485-ktbfm\" (UID: \"8d171003-64ac-4108-957b-70f4225359c5\") " pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:33 crc kubenswrapper[4766]: I1126 00:45:33.906263 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z97sp\" (UniqueName: \"kubernetes.io/projected/8d171003-64ac-4108-957b-70f4225359c5-kube-api-access-z97sp\") pod \"neutron-756fcfc485-ktbfm\" (UID: \"8d171003-64ac-4108-957b-70f4225359c5\") " pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:33 crc kubenswrapper[4766]: I1126 00:45:33.906284 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d171003-64ac-4108-957b-70f4225359c5-public-tls-certs\") pod \"neutron-756fcfc485-ktbfm\" (UID: \"8d171003-64ac-4108-957b-70f4225359c5\") " pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:33 crc kubenswrapper[4766]: I1126 00:45:33.906330 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d171003-64ac-4108-957b-70f4225359c5-ovndb-tls-certs\") pod \"neutron-756fcfc485-ktbfm\" (UID: \"8d171003-64ac-4108-957b-70f4225359c5\") " pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:33 crc kubenswrapper[4766]: I1126 00:45:33.906375 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8d171003-64ac-4108-957b-70f4225359c5-httpd-config\") pod \"neutron-756fcfc485-ktbfm\" (UID: \"8d171003-64ac-4108-957b-70f4225359c5\") " pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:33 crc kubenswrapper[4766]: I1126 00:45:33.906407 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d171003-64ac-4108-957b-70f4225359c5-combined-ca-bundle\") pod \"neutron-756fcfc485-ktbfm\" (UID: \"8d171003-64ac-4108-957b-70f4225359c5\") " pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:33 crc kubenswrapper[4766]: I1126 00:45:33.906511 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d171003-64ac-4108-957b-70f4225359c5-internal-tls-certs\") pod \"neutron-756fcfc485-ktbfm\" (UID: \"8d171003-64ac-4108-957b-70f4225359c5\") " pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:34 crc kubenswrapper[4766]: I1126 00:45:34.008453 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8d171003-64ac-4108-957b-70f4225359c5-config\") pod \"neutron-756fcfc485-ktbfm\" (UID: \"8d171003-64ac-4108-957b-70f4225359c5\") " pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:34 crc kubenswrapper[4766]: I1126 00:45:34.008517 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z97sp\" (UniqueName: \"kubernetes.io/projected/8d171003-64ac-4108-957b-70f4225359c5-kube-api-access-z97sp\") pod \"neutron-756fcfc485-ktbfm\" (UID: \"8d171003-64ac-4108-957b-70f4225359c5\") " pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:34 crc kubenswrapper[4766]: I1126 00:45:34.008551 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d171003-64ac-4108-957b-70f4225359c5-public-tls-certs\") pod \"neutron-756fcfc485-ktbfm\" (UID: \"8d171003-64ac-4108-957b-70f4225359c5\") " pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:34 crc kubenswrapper[4766]: I1126 00:45:34.008631 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d171003-64ac-4108-957b-70f4225359c5-ovndb-tls-certs\") pod \"neutron-756fcfc485-ktbfm\" (UID: \"8d171003-64ac-4108-957b-70f4225359c5\") " pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:34 crc kubenswrapper[4766]: I1126 00:45:34.008732 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8d171003-64ac-4108-957b-70f4225359c5-httpd-config\") pod \"neutron-756fcfc485-ktbfm\" (UID: \"8d171003-64ac-4108-957b-70f4225359c5\") " pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:34 crc kubenswrapper[4766]: I1126 00:45:34.008787 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d171003-64ac-4108-957b-70f4225359c5-combined-ca-bundle\") pod \"neutron-756fcfc485-ktbfm\" (UID: \"8d171003-64ac-4108-957b-70f4225359c5\") " pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:34 crc kubenswrapper[4766]: I1126 00:45:34.008856 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d171003-64ac-4108-957b-70f4225359c5-internal-tls-certs\") pod \"neutron-756fcfc485-ktbfm\" (UID: \"8d171003-64ac-4108-957b-70f4225359c5\") " pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:34 crc kubenswrapper[4766]: I1126 00:45:34.017331 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d171003-64ac-4108-957b-70f4225359c5-internal-tls-certs\") pod \"neutron-756fcfc485-ktbfm\" (UID: \"8d171003-64ac-4108-957b-70f4225359c5\") " pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:34 crc kubenswrapper[4766]: I1126 00:45:34.017786 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8d171003-64ac-4108-957b-70f4225359c5-config\") pod \"neutron-756fcfc485-ktbfm\" (UID: \"8d171003-64ac-4108-957b-70f4225359c5\") " pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:34 crc kubenswrapper[4766]: I1126 00:45:34.019491 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8d171003-64ac-4108-957b-70f4225359c5-httpd-config\") pod \"neutron-756fcfc485-ktbfm\" (UID: \"8d171003-64ac-4108-957b-70f4225359c5\") " pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:34 crc kubenswrapper[4766]: I1126 00:45:34.024206 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d171003-64ac-4108-957b-70f4225359c5-ovndb-tls-certs\") pod \"neutron-756fcfc485-ktbfm\" (UID: \"8d171003-64ac-4108-957b-70f4225359c5\") " pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:34 crc kubenswrapper[4766]: I1126 00:45:34.024816 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d171003-64ac-4108-957b-70f4225359c5-combined-ca-bundle\") pod \"neutron-756fcfc485-ktbfm\" (UID: \"8d171003-64ac-4108-957b-70f4225359c5\") " pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:34 crc kubenswrapper[4766]: I1126 00:45:34.026351 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d171003-64ac-4108-957b-70f4225359c5-public-tls-certs\") pod \"neutron-756fcfc485-ktbfm\" (UID: \"8d171003-64ac-4108-957b-70f4225359c5\") " pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:34 crc kubenswrapper[4766]: I1126 00:45:34.033829 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z97sp\" (UniqueName: \"kubernetes.io/projected/8d171003-64ac-4108-957b-70f4225359c5-kube-api-access-z97sp\") pod \"neutron-756fcfc485-ktbfm\" (UID: \"8d171003-64ac-4108-957b-70f4225359c5\") " pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:34 crc kubenswrapper[4766]: I1126 00:45:34.111997 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:34 crc kubenswrapper[4766]: I1126 00:45:34.238744 4766 generic.go:334] "Generic (PLEG): container finished" podID="6969780f-beaf-4216-8516-1eea9445f0c4" containerID="e0731b84189b5cf1d81b460c924c87afd8f2e54e09bd83d8ffc0da67f42e28fc" exitCode=0 Nov 26 00:45:34 crc kubenswrapper[4766]: I1126 00:45:34.238976 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-d26z8" event={"ID":"6969780f-beaf-4216-8516-1eea9445f0c4","Type":"ContainerDied","Data":"e0731b84189b5cf1d81b460c924c87afd8f2e54e09bd83d8ffc0da67f42e28fc"} Nov 26 00:45:35 crc kubenswrapper[4766]: I1126 00:45:35.252394 4766 generic.go:334] "Generic (PLEG): container finished" podID="66feb239-5790-4cfe-96db-e9dd6413bb09" containerID="2a9815005b07139bc4c1db61fdce11fd291bd00028e65ab55088bfc31dbe5001" exitCode=0 Nov 26 00:45:35 crc kubenswrapper[4766]: I1126 00:45:35.252481 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pd2t2" event={"ID":"66feb239-5790-4cfe-96db-e9dd6413bb09","Type":"ContainerDied","Data":"2a9815005b07139bc4c1db61fdce11fd291bd00028e65ab55088bfc31dbe5001"} Nov 26 00:45:35 crc kubenswrapper[4766]: I1126 00:45:35.255232 4766 generic.go:334] "Generic (PLEG): container finished" podID="2fd45623-8db1-440a-960a-a848abbb0c90" containerID="dd0a910ba1ed6f7f52274d8aba3d44c06ccdc43b8bfc50f731aad0b2322b2405" exitCode=0 Nov 26 00:45:35 crc kubenswrapper[4766]: I1126 00:45:35.255334 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-9259k" event={"ID":"2fd45623-8db1-440a-960a-a848abbb0c90","Type":"ContainerDied","Data":"dd0a910ba1ed6f7f52274d8aba3d44c06ccdc43b8bfc50f731aad0b2322b2405"} Nov 26 00:45:36 crc kubenswrapper[4766]: W1126 00:45:36.117320 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ee29573_3f38_49d9_8c91_6c7ae3d2bd35.slice/crio-7169e74c49370232b54c4158d36d4795c8429b59f4af4a4c1a75d8fa1f2cafc0 WatchSource:0}: Error finding container 7169e74c49370232b54c4158d36d4795c8429b59f4af4a4c1a75d8fa1f2cafc0: Status 404 returned error can't find the container with id 7169e74c49370232b54c4158d36d4795c8429b59f4af4a4c1a75d8fa1f2cafc0 Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.272527 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-7rfnj" event={"ID":"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8","Type":"ContainerDied","Data":"2ad21a7440118caad6624fadba329df97067115512ca455069184e249155de32"} Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.272596 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ad21a7440118caad6624fadba329df97067115512ca455069184e249155de32" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.275804 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-d26z8" event={"ID":"6969780f-beaf-4216-8516-1eea9445f0c4","Type":"ContainerDied","Data":"5589d7d2d769622a9977455a4bec4f19e6717ea21f785383fd6507dd30acf973"} Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.275847 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5589d7d2d769622a9977455a4bec4f19e6717ea21f785383fd6507dd30acf973" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.277580 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bbf79c884-hg8wv" event={"ID":"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35","Type":"ContainerStarted","Data":"7169e74c49370232b54c4158d36d4795c8429b59f4af4a4c1a75d8fa1f2cafc0"} Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.526281 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.570700 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-7rfnj" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.663795 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-9259k" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.710078 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-config-data\") pod \"6969780f-beaf-4216-8516-1eea9445f0c4\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.710110 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-combined-ca-bundle\") pod \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\" (UID: \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\") " Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.710135 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-config-data\") pod \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\" (UID: \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\") " Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.710152 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-scripts\") pod \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\" (UID: \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\") " Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.710175 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-scripts\") pod \"6969780f-beaf-4216-8516-1eea9445f0c4\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.710204 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28v2q\" (UniqueName: \"kubernetes.io/projected/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-kube-api-access-28v2q\") pod \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\" (UID: \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\") " Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.710225 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fd45623-8db1-440a-960a-a848abbb0c90-config-data\") pod \"2fd45623-8db1-440a-960a-a848abbb0c90\" (UID: \"2fd45623-8db1-440a-960a-a848abbb0c90\") " Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.710263 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-logs\") pod \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\" (UID: \"7d2fbed9-754d-45c7-a8c1-7dc4828b26a8\") " Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.710279 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-fernet-keys\") pod \"6969780f-beaf-4216-8516-1eea9445f0c4\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.710309 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znqbw\" (UniqueName: \"kubernetes.io/projected/6969780f-beaf-4216-8516-1eea9445f0c4-kube-api-access-znqbw\") pod \"6969780f-beaf-4216-8516-1eea9445f0c4\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.710327 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-credential-keys\") pod \"6969780f-beaf-4216-8516-1eea9445f0c4\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.710364 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fd45623-8db1-440a-960a-a848abbb0c90-combined-ca-bundle\") pod \"2fd45623-8db1-440a-960a-a848abbb0c90\" (UID: \"2fd45623-8db1-440a-960a-a848abbb0c90\") " Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.710401 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wf9q8\" (UniqueName: \"kubernetes.io/projected/2fd45623-8db1-440a-960a-a848abbb0c90-kube-api-access-wf9q8\") pod \"2fd45623-8db1-440a-960a-a848abbb0c90\" (UID: \"2fd45623-8db1-440a-960a-a848abbb0c90\") " Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.710425 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-combined-ca-bundle\") pod \"6969780f-beaf-4216-8516-1eea9445f0c4\" (UID: \"6969780f-beaf-4216-8516-1eea9445f0c4\") " Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.713171 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-logs" (OuterVolumeSpecName: "logs") pod "7d2fbed9-754d-45c7-a8c1-7dc4828b26a8" (UID: "7d2fbed9-754d-45c7-a8c1-7dc4828b26a8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.719217 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6969780f-beaf-4216-8516-1eea9445f0c4" (UID: "6969780f-beaf-4216-8516-1eea9445f0c4"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.720450 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fd45623-8db1-440a-960a-a848abbb0c90-kube-api-access-wf9q8" (OuterVolumeSpecName: "kube-api-access-wf9q8") pod "2fd45623-8db1-440a-960a-a848abbb0c90" (UID: "2fd45623-8db1-440a-960a-a848abbb0c90"). InnerVolumeSpecName "kube-api-access-wf9q8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.721055 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6969780f-beaf-4216-8516-1eea9445f0c4" (UID: "6969780f-beaf-4216-8516-1eea9445f0c4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.721564 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-kube-api-access-28v2q" (OuterVolumeSpecName: "kube-api-access-28v2q") pod "7d2fbed9-754d-45c7-a8c1-7dc4828b26a8" (UID: "7d2fbed9-754d-45c7-a8c1-7dc4828b26a8"). InnerVolumeSpecName "kube-api-access-28v2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.722998 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-scripts" (OuterVolumeSpecName: "scripts") pod "6969780f-beaf-4216-8516-1eea9445f0c4" (UID: "6969780f-beaf-4216-8516-1eea9445f0c4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.727682 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-scripts" (OuterVolumeSpecName: "scripts") pod "7d2fbed9-754d-45c7-a8c1-7dc4828b26a8" (UID: "7d2fbed9-754d-45c7-a8c1-7dc4828b26a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.732290 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6969780f-beaf-4216-8516-1eea9445f0c4-kube-api-access-znqbw" (OuterVolumeSpecName: "kube-api-access-znqbw") pod "6969780f-beaf-4216-8516-1eea9445f0c4" (UID: "6969780f-beaf-4216-8516-1eea9445f0c4"). InnerVolumeSpecName "kube-api-access-znqbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.757006 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-config-data" (OuterVolumeSpecName: "config-data") pod "6969780f-beaf-4216-8516-1eea9445f0c4" (UID: "6969780f-beaf-4216-8516-1eea9445f0c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.764273 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fd45623-8db1-440a-960a-a848abbb0c90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2fd45623-8db1-440a-960a-a848abbb0c90" (UID: "2fd45623-8db1-440a-960a-a848abbb0c90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.768088 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-config-data" (OuterVolumeSpecName: "config-data") pod "7d2fbed9-754d-45c7-a8c1-7dc4828b26a8" (UID: "7d2fbed9-754d-45c7-a8c1-7dc4828b26a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.769108 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d2fbed9-754d-45c7-a8c1-7dc4828b26a8" (UID: "7d2fbed9-754d-45c7-a8c1-7dc4828b26a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.774890 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6969780f-beaf-4216-8516-1eea9445f0c4" (UID: "6969780f-beaf-4216-8516-1eea9445f0c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.794095 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pd2t2" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.814215 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fd45623-8db1-440a-960a-a848abbb0c90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.814494 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wf9q8\" (UniqueName: \"kubernetes.io/projected/2fd45623-8db1-440a-960a-a848abbb0c90-kube-api-access-wf9q8\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.814576 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.814690 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.814840 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.814934 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.815006 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.815077 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.815148 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28v2q\" (UniqueName: \"kubernetes.io/projected/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-kube-api-access-28v2q\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.815219 4766 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8-logs\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.815293 4766 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.815364 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znqbw\" (UniqueName: \"kubernetes.io/projected/6969780f-beaf-4216-8516-1eea9445f0c4-kube-api-access-znqbw\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.815586 4766 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6969780f-beaf-4216-8516-1eea9445f0c4-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.841121 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fd45623-8db1-440a-960a-a848abbb0c90-config-data" (OuterVolumeSpecName: "config-data") pod "2fd45623-8db1-440a-960a-a848abbb0c90" (UID: "2fd45623-8db1-440a-960a-a848abbb0c90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.917241 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66feb239-5790-4cfe-96db-e9dd6413bb09-db-sync-config-data\") pod \"66feb239-5790-4cfe-96db-e9dd6413bb09\" (UID: \"66feb239-5790-4cfe-96db-e9dd6413bb09\") " Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.917334 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r568f\" (UniqueName: \"kubernetes.io/projected/66feb239-5790-4cfe-96db-e9dd6413bb09-kube-api-access-r568f\") pod \"66feb239-5790-4cfe-96db-e9dd6413bb09\" (UID: \"66feb239-5790-4cfe-96db-e9dd6413bb09\") " Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.917364 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66feb239-5790-4cfe-96db-e9dd6413bb09-combined-ca-bundle\") pod \"66feb239-5790-4cfe-96db-e9dd6413bb09\" (UID: \"66feb239-5790-4cfe-96db-e9dd6413bb09\") " Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.917845 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fd45623-8db1-440a-960a-a848abbb0c90-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.921496 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66feb239-5790-4cfe-96db-e9dd6413bb09-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "66feb239-5790-4cfe-96db-e9dd6413bb09" (UID: "66feb239-5790-4cfe-96db-e9dd6413bb09"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.922926 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66feb239-5790-4cfe-96db-e9dd6413bb09-kube-api-access-r568f" (OuterVolumeSpecName: "kube-api-access-r568f") pod "66feb239-5790-4cfe-96db-e9dd6413bb09" (UID: "66feb239-5790-4cfe-96db-e9dd6413bb09"). InnerVolumeSpecName "kube-api-access-r568f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.943252 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-756fcfc485-ktbfm"] Nov 26 00:45:36 crc kubenswrapper[4766]: I1126 00:45:36.946602 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66feb239-5790-4cfe-96db-e9dd6413bb09-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66feb239-5790-4cfe-96db-e9dd6413bb09" (UID: "66feb239-5790-4cfe-96db-e9dd6413bb09"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:36 crc kubenswrapper[4766]: W1126 00:45:36.948289 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d171003_64ac_4108_957b_70f4225359c5.slice/crio-bc27db43d2f31b16c1e5e758f2d2af58072a7776aca22520c6da82c3cb2658c6 WatchSource:0}: Error finding container bc27db43d2f31b16c1e5e758f2d2af58072a7776aca22520c6da82c3cb2658c6: Status 404 returned error can't find the container with id bc27db43d2f31b16c1e5e758f2d2af58072a7776aca22520c6da82c3cb2658c6 Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.019689 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66feb239-5790-4cfe-96db-e9dd6413bb09-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.019729 4766 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66feb239-5790-4cfe-96db-e9dd6413bb09-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.019740 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r568f\" (UniqueName: \"kubernetes.io/projected/66feb239-5790-4cfe-96db-e9dd6413bb09-kube-api-access-r568f\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.301807 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bbf79c884-hg8wv" event={"ID":"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35","Type":"ContainerStarted","Data":"19e9b43327a1616f211162f6aa898a52ef234db2201e5cee50375cf75587de00"} Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.303466 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bbf79c884-hg8wv" event={"ID":"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35","Type":"ContainerStarted","Data":"e0db47f8b6f4b59ae417bc220def5e2f48777634002fe5de75bf3765bf0af817"} Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.304262 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7bbf79c884-hg8wv" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.307570 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55ec5b52-2985-4607-8c89-590f43a3096c","Type":"ContainerStarted","Data":"0d860a8dad08282eb7da2f8a9ca7f69f203d24091631f7cbd6c1aa9002e778f7"} Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.310814 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pd2t2" event={"ID":"66feb239-5790-4cfe-96db-e9dd6413bb09","Type":"ContainerDied","Data":"7517af61f45f00e8296e55632d0df0eb107f2dae81291e52ecf13564d8e891a2"} Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.311081 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7517af61f45f00e8296e55632d0df0eb107f2dae81291e52ecf13564d8e891a2" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.311225 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pd2t2" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.316116 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-756fcfc485-ktbfm" event={"ID":"8d171003-64ac-4108-957b-70f4225359c5","Type":"ContainerStarted","Data":"1ec5dc3fdb32b2d52deb3633150d635fe4739af8f54b682d3188c439de35a1da"} Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.316280 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-756fcfc485-ktbfm" event={"ID":"8d171003-64ac-4108-957b-70f4225359c5","Type":"ContainerStarted","Data":"bc27db43d2f31b16c1e5e758f2d2af58072a7776aca22520c6da82c3cb2658c6"} Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.320143 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-9259k" event={"ID":"2fd45623-8db1-440a-960a-a848abbb0c90","Type":"ContainerDied","Data":"b7e5dce76135b8a76f33cb2f4036060c4d8bb466f033813e20b84ca3202f7d5d"} Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.320177 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7e5dce76135b8a76f33cb2f4036060c4d8bb466f033813e20b84ca3202f7d5d" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.320253 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-9259k" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.323119 4766 generic.go:334] "Generic (PLEG): container finished" podID="5db2e99c-4e43-41e6-9ba2-331990b516bd" containerID="93b832a5096615e68ebbcf6c35a4ccfc44b471692741b1fa4601a189b8877dd0" exitCode=0 Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.323228 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-d26z8" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.323218 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" event={"ID":"5db2e99c-4e43-41e6-9ba2-331990b516bd","Type":"ContainerDied","Data":"93b832a5096615e68ebbcf6c35a4ccfc44b471692741b1fa4601a189b8877dd0"} Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.323350 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-7rfnj" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.335801 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7bbf79c884-hg8wv" podStartSLOduration=6.335783022 podStartE2EDuration="6.335783022s" podCreationTimestamp="2025-11-26 00:45:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:45:37.330436197 +0000 UTC m=+1318.179206628" watchObservedRunningTime="2025-11-26 00:45:37.335783022 +0000 UTC m=+1318.184553452" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.526392 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7fd5c59d54-6ccpg"] Nov 26 00:45:37 crc kubenswrapper[4766]: E1126 00:45:37.526969 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d2fbed9-754d-45c7-a8c1-7dc4828b26a8" containerName="placement-db-sync" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.526991 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d2fbed9-754d-45c7-a8c1-7dc4828b26a8" containerName="placement-db-sync" Nov 26 00:45:37 crc kubenswrapper[4766]: E1126 00:45:37.527030 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66feb239-5790-4cfe-96db-e9dd6413bb09" containerName="barbican-db-sync" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.527040 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="66feb239-5790-4cfe-96db-e9dd6413bb09" containerName="barbican-db-sync" Nov 26 00:45:37 crc kubenswrapper[4766]: E1126 00:45:37.527052 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6969780f-beaf-4216-8516-1eea9445f0c4" containerName="keystone-bootstrap" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.527061 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6969780f-beaf-4216-8516-1eea9445f0c4" containerName="keystone-bootstrap" Nov 26 00:45:37 crc kubenswrapper[4766]: E1126 00:45:37.527074 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fd45623-8db1-440a-960a-a848abbb0c90" containerName="heat-db-sync" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.527082 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fd45623-8db1-440a-960a-a848abbb0c90" containerName="heat-db-sync" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.527327 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fd45623-8db1-440a-960a-a848abbb0c90" containerName="heat-db-sync" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.527345 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="6969780f-beaf-4216-8516-1eea9445f0c4" containerName="keystone-bootstrap" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.527373 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d2fbed9-754d-45c7-a8c1-7dc4828b26a8" containerName="placement-db-sync" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.527392 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="66feb239-5790-4cfe-96db-e9dd6413bb09" containerName="barbican-db-sync" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.528810 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7fd5c59d54-6ccpg" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.533004 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.533406 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-87dbv" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.539582 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.554407 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7fd5c59d54-6ccpg"] Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.593717 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-68b9476655-7svbz"] Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.595841 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68b9476655-7svbz" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.601558 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.635906 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/81836728-9333-4234-8175-954682305e4e-config-data-custom\") pod \"barbican-keystone-listener-7fd5c59d54-6ccpg\" (UID: \"81836728-9333-4234-8175-954682305e4e\") " pod="openstack/barbican-keystone-listener-7fd5c59d54-6ccpg" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.636230 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81836728-9333-4234-8175-954682305e4e-combined-ca-bundle\") pod \"barbican-keystone-listener-7fd5c59d54-6ccpg\" (UID: \"81836728-9333-4234-8175-954682305e4e\") " pod="openstack/barbican-keystone-listener-7fd5c59d54-6ccpg" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.636439 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81836728-9333-4234-8175-954682305e4e-logs\") pod \"barbican-keystone-listener-7fd5c59d54-6ccpg\" (UID: \"81836728-9333-4234-8175-954682305e4e\") " pod="openstack/barbican-keystone-listener-7fd5c59d54-6ccpg" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.636635 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjf2c\" (UniqueName: \"kubernetes.io/projected/81836728-9333-4234-8175-954682305e4e-kube-api-access-mjf2c\") pod \"barbican-keystone-listener-7fd5c59d54-6ccpg\" (UID: \"81836728-9333-4234-8175-954682305e4e\") " pod="openstack/barbican-keystone-listener-7fd5c59d54-6ccpg" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.636775 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81836728-9333-4234-8175-954682305e4e-config-data\") pod \"barbican-keystone-listener-7fd5c59d54-6ccpg\" (UID: \"81836728-9333-4234-8175-954682305e4e\") " pod="openstack/barbican-keystone-listener-7fd5c59d54-6ccpg" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.654716 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-68b9476655-7svbz"] Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.719730 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-vtkhw"] Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.740099 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01cfcf71-49d4-498f-a9cc-5e9c65c08709-logs\") pod \"barbican-worker-68b9476655-7svbz\" (UID: \"01cfcf71-49d4-498f-a9cc-5e9c65c08709\") " pod="openstack/barbican-worker-68b9476655-7svbz" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.740173 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8twqx\" (UniqueName: \"kubernetes.io/projected/01cfcf71-49d4-498f-a9cc-5e9c65c08709-kube-api-access-8twqx\") pod \"barbican-worker-68b9476655-7svbz\" (UID: \"01cfcf71-49d4-498f-a9cc-5e9c65c08709\") " pod="openstack/barbican-worker-68b9476655-7svbz" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.740207 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81836728-9333-4234-8175-954682305e4e-logs\") pod \"barbican-keystone-listener-7fd5c59d54-6ccpg\" (UID: \"81836728-9333-4234-8175-954682305e4e\") " pod="openstack/barbican-keystone-listener-7fd5c59d54-6ccpg" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.740276 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjf2c\" (UniqueName: \"kubernetes.io/projected/81836728-9333-4234-8175-954682305e4e-kube-api-access-mjf2c\") pod \"barbican-keystone-listener-7fd5c59d54-6ccpg\" (UID: \"81836728-9333-4234-8175-954682305e4e\") " pod="openstack/barbican-keystone-listener-7fd5c59d54-6ccpg" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.740305 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81836728-9333-4234-8175-954682305e4e-config-data\") pod \"barbican-keystone-listener-7fd5c59d54-6ccpg\" (UID: \"81836728-9333-4234-8175-954682305e4e\") " pod="openstack/barbican-keystone-listener-7fd5c59d54-6ccpg" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.740347 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/01cfcf71-49d4-498f-a9cc-5e9c65c08709-config-data-custom\") pod \"barbican-worker-68b9476655-7svbz\" (UID: \"01cfcf71-49d4-498f-a9cc-5e9c65c08709\") " pod="openstack/barbican-worker-68b9476655-7svbz" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.740402 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01cfcf71-49d4-498f-a9cc-5e9c65c08709-combined-ca-bundle\") pod \"barbican-worker-68b9476655-7svbz\" (UID: \"01cfcf71-49d4-498f-a9cc-5e9c65c08709\") " pod="openstack/barbican-worker-68b9476655-7svbz" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.740427 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/81836728-9333-4234-8175-954682305e4e-config-data-custom\") pod \"barbican-keystone-listener-7fd5c59d54-6ccpg\" (UID: \"81836728-9333-4234-8175-954682305e4e\") " pod="openstack/barbican-keystone-listener-7fd5c59d54-6ccpg" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.740503 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81836728-9333-4234-8175-954682305e4e-combined-ca-bundle\") pod \"barbican-keystone-listener-7fd5c59d54-6ccpg\" (UID: \"81836728-9333-4234-8175-954682305e4e\") " pod="openstack/barbican-keystone-listener-7fd5c59d54-6ccpg" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.740566 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01cfcf71-49d4-498f-a9cc-5e9c65c08709-config-data\") pod \"barbican-worker-68b9476655-7svbz\" (UID: \"01cfcf71-49d4-498f-a9cc-5e9c65c08709\") " pod="openstack/barbican-worker-68b9476655-7svbz" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.743180 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81836728-9333-4234-8175-954682305e4e-logs\") pod \"barbican-keystone-listener-7fd5c59d54-6ccpg\" (UID: \"81836728-9333-4234-8175-954682305e4e\") " pod="openstack/barbican-keystone-listener-7fd5c59d54-6ccpg" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.746959 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-6mpmd"] Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.749173 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.754401 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81836728-9333-4234-8175-954682305e4e-config-data\") pod \"barbican-keystone-listener-7fd5c59d54-6ccpg\" (UID: \"81836728-9333-4234-8175-954682305e4e\") " pod="openstack/barbican-keystone-listener-7fd5c59d54-6ccpg" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.758444 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/81836728-9333-4234-8175-954682305e4e-config-data-custom\") pod \"barbican-keystone-listener-7fd5c59d54-6ccpg\" (UID: \"81836728-9333-4234-8175-954682305e4e\") " pod="openstack/barbican-keystone-listener-7fd5c59d54-6ccpg" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.767110 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81836728-9333-4234-8175-954682305e4e-combined-ca-bundle\") pod \"barbican-keystone-listener-7fd5c59d54-6ccpg\" (UID: \"81836728-9333-4234-8175-954682305e4e\") " pod="openstack/barbican-keystone-listener-7fd5c59d54-6ccpg" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.778030 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjf2c\" (UniqueName: \"kubernetes.io/projected/81836728-9333-4234-8175-954682305e4e-kube-api-access-mjf2c\") pod \"barbican-keystone-listener-7fd5c59d54-6ccpg\" (UID: \"81836728-9333-4234-8175-954682305e4e\") " pod="openstack/barbican-keystone-listener-7fd5c59d54-6ccpg" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.786872 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-6mpmd"] Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.802791 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6768b6c6f7-lm9pw"] Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.804618 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.809260 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ksmzm" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.809437 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.809539 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.809643 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.809794 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.815184 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.817619 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6768b6c6f7-lm9pw"] Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.851099 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-6mpmd\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.851245 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01cfcf71-49d4-498f-a9cc-5e9c65c08709-config-data\") pod \"barbican-worker-68b9476655-7svbz\" (UID: \"01cfcf71-49d4-498f-a9cc-5e9c65c08709\") " pod="openstack/barbican-worker-68b9476655-7svbz" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.851322 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-6mpmd\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.851352 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01cfcf71-49d4-498f-a9cc-5e9c65c08709-logs\") pod \"barbican-worker-68b9476655-7svbz\" (UID: \"01cfcf71-49d4-498f-a9cc-5e9c65c08709\") " pod="openstack/barbican-worker-68b9476655-7svbz" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.851417 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8twqx\" (UniqueName: \"kubernetes.io/projected/01cfcf71-49d4-498f-a9cc-5e9c65c08709-kube-api-access-8twqx\") pod \"barbican-worker-68b9476655-7svbz\" (UID: \"01cfcf71-49d4-498f-a9cc-5e9c65c08709\") " pod="openstack/barbican-worker-68b9476655-7svbz" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.851462 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-6mpmd\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.851505 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-config\") pod \"dnsmasq-dns-848cf88cfc-6mpmd\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.851643 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/01cfcf71-49d4-498f-a9cc-5e9c65c08709-config-data-custom\") pod \"barbican-worker-68b9476655-7svbz\" (UID: \"01cfcf71-49d4-498f-a9cc-5e9c65c08709\") " pod="openstack/barbican-worker-68b9476655-7svbz" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.851750 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01cfcf71-49d4-498f-a9cc-5e9c65c08709-combined-ca-bundle\") pod \"barbican-worker-68b9476655-7svbz\" (UID: \"01cfcf71-49d4-498f-a9cc-5e9c65c08709\") " pod="openstack/barbican-worker-68b9476655-7svbz" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.851875 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-6mpmd\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.851897 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z77t\" (UniqueName: \"kubernetes.io/projected/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-kube-api-access-8z77t\") pod \"dnsmasq-dns-848cf88cfc-6mpmd\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.852905 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01cfcf71-49d4-498f-a9cc-5e9c65c08709-logs\") pod \"barbican-worker-68b9476655-7svbz\" (UID: \"01cfcf71-49d4-498f-a9cc-5e9c65c08709\") " pod="openstack/barbican-worker-68b9476655-7svbz" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.868073 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01cfcf71-49d4-498f-a9cc-5e9c65c08709-combined-ca-bundle\") pod \"barbican-worker-68b9476655-7svbz\" (UID: \"01cfcf71-49d4-498f-a9cc-5e9c65c08709\") " pod="openstack/barbican-worker-68b9476655-7svbz" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.870283 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5fd6484f46-cqsxg"] Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.876466 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.882997 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01cfcf71-49d4-498f-a9cc-5e9c65c08709-config-data\") pod \"barbican-worker-68b9476655-7svbz\" (UID: \"01cfcf71-49d4-498f-a9cc-5e9c65c08709\") " pod="openstack/barbican-worker-68b9476655-7svbz" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.883942 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-44jtp" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.884644 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/01cfcf71-49d4-498f-a9cc-5e9c65c08709-config-data-custom\") pod \"barbican-worker-68b9476655-7svbz\" (UID: \"01cfcf71-49d4-498f-a9cc-5e9c65c08709\") " pod="openstack/barbican-worker-68b9476655-7svbz" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.883978 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.884795 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8twqx\" (UniqueName: \"kubernetes.io/projected/01cfcf71-49d4-498f-a9cc-5e9c65c08709-kube-api-access-8twqx\") pod \"barbican-worker-68b9476655-7svbz\" (UID: \"01cfcf71-49d4-498f-a9cc-5e9c65c08709\") " pod="openstack/barbican-worker-68b9476655-7svbz" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.884037 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.884042 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.884110 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.889172 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7fd5c59d54-6ccpg" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.890886 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5fd6484f46-cqsxg"] Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.918944 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-67fdfd759d-sqzs7"] Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.943229 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68b9476655-7svbz" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.945236 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.948421 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.954087 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fdd0908-10b2-437d-92bf-ccdc345a0d80-combined-ca-bundle\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.954147 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fdd0908-10b2-437d-92bf-ccdc345a0d80-public-tls-certs\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.954196 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-6mpmd\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.954236 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fdd0908-10b2-437d-92bf-ccdc345a0d80-internal-tls-certs\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.954259 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3fdd0908-10b2-437d-92bf-ccdc345a0d80-credential-keys\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.954305 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-6mpmd\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.954327 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-config\") pod \"dnsmasq-dns-848cf88cfc-6mpmd\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.954376 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3fdd0908-10b2-437d-92bf-ccdc345a0d80-fernet-keys\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.954413 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfhdg\" (UniqueName: \"kubernetes.io/projected/3fdd0908-10b2-437d-92bf-ccdc345a0d80-kube-api-access-gfhdg\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.954428 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fdd0908-10b2-437d-92bf-ccdc345a0d80-config-data\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.954520 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-6mpmd\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.954539 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z77t\" (UniqueName: \"kubernetes.io/projected/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-kube-api-access-8z77t\") pod \"dnsmasq-dns-848cf88cfc-6mpmd\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.954557 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-6mpmd\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.954592 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3fdd0908-10b2-437d-92bf-ccdc345a0d80-scripts\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.955942 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-6mpmd\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.957217 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-6mpmd\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.957919 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-config\") pod \"dnsmasq-dns-848cf88cfc-6mpmd\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.961799 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-6mpmd\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:37 crc kubenswrapper[4766]: I1126 00:45:37.962176 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-6mpmd\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.015409 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-67fdfd759d-sqzs7"] Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.049359 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z77t\" (UniqueName: \"kubernetes.io/projected/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-kube-api-access-8z77t\") pod \"dnsmasq-dns-848cf88cfc-6mpmd\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.075349 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fdd0908-10b2-437d-92bf-ccdc345a0d80-internal-tls-certs\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.075413 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3fdd0908-10b2-437d-92bf-ccdc345a0d80-credential-keys\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.075449 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be4073df-dee9-4a70-8911-b23a317a0a57-internal-tls-certs\") pod \"placement-5fd6484f46-cqsxg\" (UID: \"be4073df-dee9-4a70-8911-b23a317a0a57\") " pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.075478 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be4073df-dee9-4a70-8911-b23a317a0a57-public-tls-certs\") pod \"placement-5fd6484f46-cqsxg\" (UID: \"be4073df-dee9-4a70-8911-b23a317a0a57\") " pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.075582 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htvbp\" (UniqueName: \"kubernetes.io/projected/be4073df-dee9-4a70-8911-b23a317a0a57-kube-api-access-htvbp\") pod \"placement-5fd6484f46-cqsxg\" (UID: \"be4073df-dee9-4a70-8911-b23a317a0a57\") " pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.075631 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be4073df-dee9-4a70-8911-b23a317a0a57-scripts\") pod \"placement-5fd6484f46-cqsxg\" (UID: \"be4073df-dee9-4a70-8911-b23a317a0a57\") " pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.075665 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-logs\") pod \"barbican-api-67fdfd759d-sqzs7\" (UID: \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\") " pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.075684 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3fdd0908-10b2-437d-92bf-ccdc345a0d80-fernet-keys\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.075717 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mv6r4\" (UniqueName: \"kubernetes.io/projected/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-kube-api-access-mv6r4\") pod \"barbican-api-67fdfd759d-sqzs7\" (UID: \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\") " pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.084811 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfhdg\" (UniqueName: \"kubernetes.io/projected/3fdd0908-10b2-437d-92bf-ccdc345a0d80-kube-api-access-gfhdg\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.084933 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fdd0908-10b2-437d-92bf-ccdc345a0d80-config-data\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.085014 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-combined-ca-bundle\") pod \"barbican-api-67fdfd759d-sqzs7\" (UID: \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\") " pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.085092 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be4073df-dee9-4a70-8911-b23a317a0a57-logs\") pod \"placement-5fd6484f46-cqsxg\" (UID: \"be4073df-dee9-4a70-8911-b23a317a0a57\") " pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.085219 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4073df-dee9-4a70-8911-b23a317a0a57-combined-ca-bundle\") pod \"placement-5fd6484f46-cqsxg\" (UID: \"be4073df-dee9-4a70-8911-b23a317a0a57\") " pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.085289 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-config-data\") pod \"barbican-api-67fdfd759d-sqzs7\" (UID: \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\") " pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.085319 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-config-data-custom\") pod \"barbican-api-67fdfd759d-sqzs7\" (UID: \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\") " pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.085460 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3fdd0908-10b2-437d-92bf-ccdc345a0d80-scripts\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.085534 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fdd0908-10b2-437d-92bf-ccdc345a0d80-combined-ca-bundle\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.085624 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fdd0908-10b2-437d-92bf-ccdc345a0d80-public-tls-certs\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.085673 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be4073df-dee9-4a70-8911-b23a317a0a57-config-data\") pod \"placement-5fd6484f46-cqsxg\" (UID: \"be4073df-dee9-4a70-8911-b23a317a0a57\") " pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.087140 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fdd0908-10b2-437d-92bf-ccdc345a0d80-internal-tls-certs\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.088669 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3fdd0908-10b2-437d-92bf-ccdc345a0d80-credential-keys\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.093285 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fdd0908-10b2-437d-92bf-ccdc345a0d80-combined-ca-bundle\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.091471 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fdd0908-10b2-437d-92bf-ccdc345a0d80-public-tls-certs\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.089806 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3fdd0908-10b2-437d-92bf-ccdc345a0d80-scripts\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.094307 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3fdd0908-10b2-437d-92bf-ccdc345a0d80-fernet-keys\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.098851 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fdd0908-10b2-437d-92bf-ccdc345a0d80-config-data\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.104824 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.105204 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfhdg\" (UniqueName: \"kubernetes.io/projected/3fdd0908-10b2-437d-92bf-ccdc345a0d80-kube-api-access-gfhdg\") pod \"keystone-6768b6c6f7-lm9pw\" (UID: \"3fdd0908-10b2-437d-92bf-ccdc345a0d80\") " pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.166684 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.203795 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mv6r4\" (UniqueName: \"kubernetes.io/projected/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-kube-api-access-mv6r4\") pod \"barbican-api-67fdfd759d-sqzs7\" (UID: \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\") " pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.203895 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-combined-ca-bundle\") pod \"barbican-api-67fdfd759d-sqzs7\" (UID: \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\") " pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.203950 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be4073df-dee9-4a70-8911-b23a317a0a57-logs\") pod \"placement-5fd6484f46-cqsxg\" (UID: \"be4073df-dee9-4a70-8911-b23a317a0a57\") " pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.204024 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4073df-dee9-4a70-8911-b23a317a0a57-combined-ca-bundle\") pod \"placement-5fd6484f46-cqsxg\" (UID: \"be4073df-dee9-4a70-8911-b23a317a0a57\") " pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.204091 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-config-data\") pod \"barbican-api-67fdfd759d-sqzs7\" (UID: \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\") " pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.204111 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-config-data-custom\") pod \"barbican-api-67fdfd759d-sqzs7\" (UID: \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\") " pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.204243 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be4073df-dee9-4a70-8911-b23a317a0a57-config-data\") pod \"placement-5fd6484f46-cqsxg\" (UID: \"be4073df-dee9-4a70-8911-b23a317a0a57\") " pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.204336 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be4073df-dee9-4a70-8911-b23a317a0a57-internal-tls-certs\") pod \"placement-5fd6484f46-cqsxg\" (UID: \"be4073df-dee9-4a70-8911-b23a317a0a57\") " pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.204369 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be4073df-dee9-4a70-8911-b23a317a0a57-public-tls-certs\") pod \"placement-5fd6484f46-cqsxg\" (UID: \"be4073df-dee9-4a70-8911-b23a317a0a57\") " pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.204399 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htvbp\" (UniqueName: \"kubernetes.io/projected/be4073df-dee9-4a70-8911-b23a317a0a57-kube-api-access-htvbp\") pod \"placement-5fd6484f46-cqsxg\" (UID: \"be4073df-dee9-4a70-8911-b23a317a0a57\") " pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.204462 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be4073df-dee9-4a70-8911-b23a317a0a57-scripts\") pod \"placement-5fd6484f46-cqsxg\" (UID: \"be4073df-dee9-4a70-8911-b23a317a0a57\") " pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.204481 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-logs\") pod \"barbican-api-67fdfd759d-sqzs7\" (UID: \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\") " pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.214403 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be4073df-dee9-4a70-8911-b23a317a0a57-logs\") pod \"placement-5fd6484f46-cqsxg\" (UID: \"be4073df-dee9-4a70-8911-b23a317a0a57\") " pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.224943 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be4073df-dee9-4a70-8911-b23a317a0a57-config-data\") pod \"placement-5fd6484f46-cqsxg\" (UID: \"be4073df-dee9-4a70-8911-b23a317a0a57\") " pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.225842 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-config-data-custom\") pod \"barbican-api-67fdfd759d-sqzs7\" (UID: \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\") " pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.204999 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-logs\") pod \"barbican-api-67fdfd759d-sqzs7\" (UID: \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\") " pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.237691 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4073df-dee9-4a70-8911-b23a317a0a57-combined-ca-bundle\") pod \"placement-5fd6484f46-cqsxg\" (UID: \"be4073df-dee9-4a70-8911-b23a317a0a57\") " pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.240125 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be4073df-dee9-4a70-8911-b23a317a0a57-scripts\") pod \"placement-5fd6484f46-cqsxg\" (UID: \"be4073df-dee9-4a70-8911-b23a317a0a57\") " pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.240409 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be4073df-dee9-4a70-8911-b23a317a0a57-public-tls-certs\") pod \"placement-5fd6484f46-cqsxg\" (UID: \"be4073df-dee9-4a70-8911-b23a317a0a57\") " pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.243448 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-combined-ca-bundle\") pod \"barbican-api-67fdfd759d-sqzs7\" (UID: \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\") " pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.244844 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-config-data\") pod \"barbican-api-67fdfd759d-sqzs7\" (UID: \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\") " pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.246837 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mv6r4\" (UniqueName: \"kubernetes.io/projected/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-kube-api-access-mv6r4\") pod \"barbican-api-67fdfd759d-sqzs7\" (UID: \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\") " pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.251864 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htvbp\" (UniqueName: \"kubernetes.io/projected/be4073df-dee9-4a70-8911-b23a317a0a57-kube-api-access-htvbp\") pod \"placement-5fd6484f46-cqsxg\" (UID: \"be4073df-dee9-4a70-8911-b23a317a0a57\") " pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.252142 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be4073df-dee9-4a70-8911-b23a317a0a57-internal-tls-certs\") pod \"placement-5fd6484f46-cqsxg\" (UID: \"be4073df-dee9-4a70-8911-b23a317a0a57\") " pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.299272 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.407213 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" event={"ID":"5db2e99c-4e43-41e6-9ba2-331990b516bd","Type":"ContainerStarted","Data":"96f333884e24d31c81b9872ac56dae88010201f35ab3c5598b75a859cb7abba9"} Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.407599 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" podUID="5db2e99c-4e43-41e6-9ba2-331990b516bd" containerName="dnsmasq-dns" containerID="cri-o://96f333884e24d31c81b9872ac56dae88010201f35ab3c5598b75a859cb7abba9" gracePeriod=10 Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.407735 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.422486 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-756fcfc485-ktbfm" event={"ID":"8d171003-64ac-4108-957b-70f4225359c5","Type":"ContainerStarted","Data":"c1ef0e45e159ffd533666272ecdfbf4d7f49afc87b509822a15f0ed3f9f9d720"} Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.422632 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.461248 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" podStartSLOduration=7.461226035 podStartE2EDuration="7.461226035s" podCreationTimestamp="2025-11-26 00:45:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:45:38.447945662 +0000 UTC m=+1319.296716092" watchObservedRunningTime="2025-11-26 00:45:38.461226035 +0000 UTC m=+1319.309996455" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.480536 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-756fcfc485-ktbfm" podStartSLOduration=5.480517789 podStartE2EDuration="5.480517789s" podCreationTimestamp="2025-11-26 00:45:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:45:38.475223536 +0000 UTC m=+1319.323993966" watchObservedRunningTime="2025-11-26 00:45:38.480517789 +0000 UTC m=+1319.329288219" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.522179 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.555147 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7fd5c59d54-6ccpg"] Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.774808 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-68b9476655-7svbz"] Nov 26 00:45:38 crc kubenswrapper[4766]: I1126 00:45:38.929057 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-6mpmd"] Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.088321 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-67fdfd759d-sqzs7"] Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.101800 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6768b6c6f7-lm9pw"] Nov 26 00:45:39 crc kubenswrapper[4766]: W1126 00:45:39.139257 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9b3be2b_22dc_4660_9d9a_dde2bbf80a5d.slice/crio-c1a081c01d23e6f098de1d4221082f278f3ca37fce64196a3afe897391cee961 WatchSource:0}: Error finding container c1a081c01d23e6f098de1d4221082f278f3ca37fce64196a3afe897391cee961: Status 404 returned error can't find the container with id c1a081c01d23e6f098de1d4221082f278f3ca37fce64196a3afe897391cee961 Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.346538 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5fd6484f46-cqsxg"] Nov 26 00:45:39 crc kubenswrapper[4766]: W1126 00:45:39.373494 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe4073df_dee9_4a70_8911_b23a317a0a57.slice/crio-29731f725585979d6fc985f5e1191041b5e587c8dfaea8fc4f7598fff51c1a2c WatchSource:0}: Error finding container 29731f725585979d6fc985f5e1191041b5e587c8dfaea8fc4f7598fff51c1a2c: Status 404 returned error can't find the container with id 29731f725585979d6fc985f5e1191041b5e587c8dfaea8fc4f7598fff51c1a2c Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.438078 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.440244 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7fd5c59d54-6ccpg" event={"ID":"81836728-9333-4234-8175-954682305e4e","Type":"ContainerStarted","Data":"2055754ed1acef2813822be5b54b5750d7a855d2d6e635540cbbc5bad7d2932b"} Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.447774 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67fdfd759d-sqzs7" event={"ID":"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d","Type":"ContainerStarted","Data":"c1a081c01d23e6f098de1d4221082f278f3ca37fce64196a3afe897391cee961"} Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.450876 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6768b6c6f7-lm9pw" event={"ID":"3fdd0908-10b2-437d-92bf-ccdc345a0d80","Type":"ContainerStarted","Data":"65bb1a1490ca17f0ac73b42da423d2f85987441d9ec6c1612c64a91c9ab7570e"} Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.458693 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" event={"ID":"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf","Type":"ContainerStarted","Data":"b7d4b61a7aecfcfc928de430bc52b2afd98b9427c1f733399cf26530e913f479"} Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.472163 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5fd6484f46-cqsxg" event={"ID":"be4073df-dee9-4a70-8911-b23a317a0a57","Type":"ContainerStarted","Data":"29731f725585979d6fc985f5e1191041b5e587c8dfaea8fc4f7598fff51c1a2c"} Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.477899 4766 generic.go:334] "Generic (PLEG): container finished" podID="5db2e99c-4e43-41e6-9ba2-331990b516bd" containerID="96f333884e24d31c81b9872ac56dae88010201f35ab3c5598b75a859cb7abba9" exitCode=0 Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.477974 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" event={"ID":"5db2e99c-4e43-41e6-9ba2-331990b516bd","Type":"ContainerDied","Data":"96f333884e24d31c81b9872ac56dae88010201f35ab3c5598b75a859cb7abba9"} Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.478005 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" event={"ID":"5db2e99c-4e43-41e6-9ba2-331990b516bd","Type":"ContainerDied","Data":"0eb801770935205934c7a12573e12dc3138bf6ef9e19d3ea069d51a57c45fe2e"} Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.478025 4766 scope.go:117] "RemoveContainer" containerID="96f333884e24d31c81b9872ac56dae88010201f35ab3c5598b75a859cb7abba9" Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.478108 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-vtkhw" Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.489116 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68b9476655-7svbz" event={"ID":"01cfcf71-49d4-498f-a9cc-5e9c65c08709","Type":"ContainerStarted","Data":"134306ac43283ea341e23d042e74c2ea5b113a7dd5d5b76d87b08f4d2ebda536"} Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.544526 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-dns-svc\") pod \"5db2e99c-4e43-41e6-9ba2-331990b516bd\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.544586 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-dns-swift-storage-0\") pod \"5db2e99c-4e43-41e6-9ba2-331990b516bd\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.544618 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8f285\" (UniqueName: \"kubernetes.io/projected/5db2e99c-4e43-41e6-9ba2-331990b516bd-kube-api-access-8f285\") pod \"5db2e99c-4e43-41e6-9ba2-331990b516bd\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.544638 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-ovsdbserver-sb\") pod \"5db2e99c-4e43-41e6-9ba2-331990b516bd\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.544719 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-config\") pod \"5db2e99c-4e43-41e6-9ba2-331990b516bd\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.544765 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-ovsdbserver-nb\") pod \"5db2e99c-4e43-41e6-9ba2-331990b516bd\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.619404 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5db2e99c-4e43-41e6-9ba2-331990b516bd-kube-api-access-8f285" (OuterVolumeSpecName: "kube-api-access-8f285") pod "5db2e99c-4e43-41e6-9ba2-331990b516bd" (UID: "5db2e99c-4e43-41e6-9ba2-331990b516bd"). InnerVolumeSpecName "kube-api-access-8f285". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.648038 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8f285\" (UniqueName: \"kubernetes.io/projected/5db2e99c-4e43-41e6-9ba2-331990b516bd-kube-api-access-8f285\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.678951 4766 scope.go:117] "RemoveContainer" containerID="93b832a5096615e68ebbcf6c35a4ccfc44b471692741b1fa4601a189b8877dd0" Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.773717 4766 scope.go:117] "RemoveContainer" containerID="96f333884e24d31c81b9872ac56dae88010201f35ab3c5598b75a859cb7abba9" Nov 26 00:45:39 crc kubenswrapper[4766]: E1126 00:45:39.774522 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96f333884e24d31c81b9872ac56dae88010201f35ab3c5598b75a859cb7abba9\": container with ID starting with 96f333884e24d31c81b9872ac56dae88010201f35ab3c5598b75a859cb7abba9 not found: ID does not exist" containerID="96f333884e24d31c81b9872ac56dae88010201f35ab3c5598b75a859cb7abba9" Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.774626 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96f333884e24d31c81b9872ac56dae88010201f35ab3c5598b75a859cb7abba9"} err="failed to get container status \"96f333884e24d31c81b9872ac56dae88010201f35ab3c5598b75a859cb7abba9\": rpc error: code = NotFound desc = could not find container \"96f333884e24d31c81b9872ac56dae88010201f35ab3c5598b75a859cb7abba9\": container with ID starting with 96f333884e24d31c81b9872ac56dae88010201f35ab3c5598b75a859cb7abba9 not found: ID does not exist" Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.774707 4766 scope.go:117] "RemoveContainer" containerID="93b832a5096615e68ebbcf6c35a4ccfc44b471692741b1fa4601a189b8877dd0" Nov 26 00:45:39 crc kubenswrapper[4766]: E1126 00:45:39.775365 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93b832a5096615e68ebbcf6c35a4ccfc44b471692741b1fa4601a189b8877dd0\": container with ID starting with 93b832a5096615e68ebbcf6c35a4ccfc44b471692741b1fa4601a189b8877dd0 not found: ID does not exist" containerID="93b832a5096615e68ebbcf6c35a4ccfc44b471692741b1fa4601a189b8877dd0" Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.775392 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93b832a5096615e68ebbcf6c35a4ccfc44b471692741b1fa4601a189b8877dd0"} err="failed to get container status \"93b832a5096615e68ebbcf6c35a4ccfc44b471692741b1fa4601a189b8877dd0\": rpc error: code = NotFound desc = could not find container \"93b832a5096615e68ebbcf6c35a4ccfc44b471692741b1fa4601a189b8877dd0\": container with ID starting with 93b832a5096615e68ebbcf6c35a4ccfc44b471692741b1fa4601a189b8877dd0 not found: ID does not exist" Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.788563 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-config" (OuterVolumeSpecName: "config") pod "5db2e99c-4e43-41e6-9ba2-331990b516bd" (UID: "5db2e99c-4e43-41e6-9ba2-331990b516bd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.789366 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5db2e99c-4e43-41e6-9ba2-331990b516bd" (UID: "5db2e99c-4e43-41e6-9ba2-331990b516bd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.791985 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5db2e99c-4e43-41e6-9ba2-331990b516bd" (UID: "5db2e99c-4e43-41e6-9ba2-331990b516bd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:39 crc kubenswrapper[4766]: E1126 00:45:39.795051 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-ovsdbserver-sb podName:5db2e99c-4e43-41e6-9ba2-331990b516bd nodeName:}" failed. No retries permitted until 2025-11-26 00:45:40.295006443 +0000 UTC m=+1321.143776873 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ovsdbserver-sb" (UniqueName: "kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-ovsdbserver-sb") pod "5db2e99c-4e43-41e6-9ba2-331990b516bd" (UID: "5db2e99c-4e43-41e6-9ba2-331990b516bd") : error deleting /var/lib/kubelet/pods/5db2e99c-4e43-41e6-9ba2-331990b516bd/volume-subpaths: remove /var/lib/kubelet/pods/5db2e99c-4e43-41e6-9ba2-331990b516bd/volume-subpaths: no such file or directory Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.795969 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5db2e99c-4e43-41e6-9ba2-331990b516bd" (UID: "5db2e99c-4e43-41e6-9ba2-331990b516bd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.856153 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.856448 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.856460 4766 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:39 crc kubenswrapper[4766]: I1126 00:45:39.856473 4766 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:40 crc kubenswrapper[4766]: I1126 00:45:40.387436 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5db2e99c-4e43-41e6-9ba2-331990b516bd" (UID: "5db2e99c-4e43-41e6-9ba2-331990b516bd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:40 crc kubenswrapper[4766]: I1126 00:45:40.389196 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-ovsdbserver-sb\") pod \"5db2e99c-4e43-41e6-9ba2-331990b516bd\" (UID: \"5db2e99c-4e43-41e6-9ba2-331990b516bd\") " Nov 26 00:45:40 crc kubenswrapper[4766]: I1126 00:45:40.391750 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5db2e99c-4e43-41e6-9ba2-331990b516bd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:40 crc kubenswrapper[4766]: I1126 00:45:40.523010 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67fdfd759d-sqzs7" event={"ID":"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d","Type":"ContainerStarted","Data":"0c2c7807bed04eacaab51baab389f6ad6e8d873b0bbbe0bac9f40f9297331099"} Nov 26 00:45:40 crc kubenswrapper[4766]: I1126 00:45:40.529468 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" event={"ID":"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf","Type":"ContainerStarted","Data":"83579b0bcf4b669c84c274a4beac436450174217ba2feda87383e0ff170b31d5"} Nov 26 00:45:40 crc kubenswrapper[4766]: I1126 00:45:40.722111 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-vtkhw"] Nov 26 00:45:40 crc kubenswrapper[4766]: I1126 00:45:40.733636 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-vtkhw"] Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.476048 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6b7ccbcb5b-rf9v9"] Nov 26 00:45:41 crc kubenswrapper[4766]: E1126 00:45:41.477644 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5db2e99c-4e43-41e6-9ba2-331990b516bd" containerName="init" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.477682 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5db2e99c-4e43-41e6-9ba2-331990b516bd" containerName="init" Nov 26 00:45:41 crc kubenswrapper[4766]: E1126 00:45:41.477693 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5db2e99c-4e43-41e6-9ba2-331990b516bd" containerName="dnsmasq-dns" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.477701 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5db2e99c-4e43-41e6-9ba2-331990b516bd" containerName="dnsmasq-dns" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.477964 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="5db2e99c-4e43-41e6-9ba2-331990b516bd" containerName="dnsmasq-dns" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.479359 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.479752 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.479811 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.486582 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.486924 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.508627 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6b7ccbcb5b-rf9v9"] Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.530705 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5655d266-3ea5-404f-901e-80da609b0319-logs\") pod \"barbican-api-6b7ccbcb5b-rf9v9\" (UID: \"5655d266-3ea5-404f-901e-80da609b0319\") " pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.532992 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5655d266-3ea5-404f-901e-80da609b0319-internal-tls-certs\") pod \"barbican-api-6b7ccbcb5b-rf9v9\" (UID: \"5655d266-3ea5-404f-901e-80da609b0319\") " pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.533060 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5655d266-3ea5-404f-901e-80da609b0319-config-data\") pod \"barbican-api-6b7ccbcb5b-rf9v9\" (UID: \"5655d266-3ea5-404f-901e-80da609b0319\") " pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.533191 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5655d266-3ea5-404f-901e-80da609b0319-combined-ca-bundle\") pod \"barbican-api-6b7ccbcb5b-rf9v9\" (UID: \"5655d266-3ea5-404f-901e-80da609b0319\") " pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.533240 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5655d266-3ea5-404f-901e-80da609b0319-config-data-custom\") pod \"barbican-api-6b7ccbcb5b-rf9v9\" (UID: \"5655d266-3ea5-404f-901e-80da609b0319\") " pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.533254 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5655d266-3ea5-404f-901e-80da609b0319-public-tls-certs\") pod \"barbican-api-6b7ccbcb5b-rf9v9\" (UID: \"5655d266-3ea5-404f-901e-80da609b0319\") " pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.533333 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h6tf\" (UniqueName: \"kubernetes.io/projected/5655d266-3ea5-404f-901e-80da609b0319-kube-api-access-9h6tf\") pod \"barbican-api-6b7ccbcb5b-rf9v9\" (UID: \"5655d266-3ea5-404f-901e-80da609b0319\") " pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.635826 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5655d266-3ea5-404f-901e-80da609b0319-internal-tls-certs\") pod \"barbican-api-6b7ccbcb5b-rf9v9\" (UID: \"5655d266-3ea5-404f-901e-80da609b0319\") " pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.635897 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5655d266-3ea5-404f-901e-80da609b0319-config-data\") pod \"barbican-api-6b7ccbcb5b-rf9v9\" (UID: \"5655d266-3ea5-404f-901e-80da609b0319\") " pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.635978 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5655d266-3ea5-404f-901e-80da609b0319-combined-ca-bundle\") pod \"barbican-api-6b7ccbcb5b-rf9v9\" (UID: \"5655d266-3ea5-404f-901e-80da609b0319\") " pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.636019 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5655d266-3ea5-404f-901e-80da609b0319-config-data-custom\") pod \"barbican-api-6b7ccbcb5b-rf9v9\" (UID: \"5655d266-3ea5-404f-901e-80da609b0319\") " pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.636038 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5655d266-3ea5-404f-901e-80da609b0319-public-tls-certs\") pod \"barbican-api-6b7ccbcb5b-rf9v9\" (UID: \"5655d266-3ea5-404f-901e-80da609b0319\") " pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.636092 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h6tf\" (UniqueName: \"kubernetes.io/projected/5655d266-3ea5-404f-901e-80da609b0319-kube-api-access-9h6tf\") pod \"barbican-api-6b7ccbcb5b-rf9v9\" (UID: \"5655d266-3ea5-404f-901e-80da609b0319\") " pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.636170 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5655d266-3ea5-404f-901e-80da609b0319-logs\") pod \"barbican-api-6b7ccbcb5b-rf9v9\" (UID: \"5655d266-3ea5-404f-901e-80da609b0319\") " pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.636641 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5655d266-3ea5-404f-901e-80da609b0319-logs\") pod \"barbican-api-6b7ccbcb5b-rf9v9\" (UID: \"5655d266-3ea5-404f-901e-80da609b0319\") " pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.643257 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5655d266-3ea5-404f-901e-80da609b0319-internal-tls-certs\") pod \"barbican-api-6b7ccbcb5b-rf9v9\" (UID: \"5655d266-3ea5-404f-901e-80da609b0319\") " pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.644086 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5655d266-3ea5-404f-901e-80da609b0319-config-data\") pod \"barbican-api-6b7ccbcb5b-rf9v9\" (UID: \"5655d266-3ea5-404f-901e-80da609b0319\") " pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.645064 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5655d266-3ea5-404f-901e-80da609b0319-config-data-custom\") pod \"barbican-api-6b7ccbcb5b-rf9v9\" (UID: \"5655d266-3ea5-404f-901e-80da609b0319\") " pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.646424 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5655d266-3ea5-404f-901e-80da609b0319-combined-ca-bundle\") pod \"barbican-api-6b7ccbcb5b-rf9v9\" (UID: \"5655d266-3ea5-404f-901e-80da609b0319\") " pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.647757 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5655d266-3ea5-404f-901e-80da609b0319-public-tls-certs\") pod \"barbican-api-6b7ccbcb5b-rf9v9\" (UID: \"5655d266-3ea5-404f-901e-80da609b0319\") " pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.658584 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h6tf\" (UniqueName: \"kubernetes.io/projected/5655d266-3ea5-404f-901e-80da609b0319-kube-api-access-9h6tf\") pod \"barbican-api-6b7ccbcb5b-rf9v9\" (UID: \"5655d266-3ea5-404f-901e-80da609b0319\") " pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.818344 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:41 crc kubenswrapper[4766]: I1126 00:45:41.843431 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5db2e99c-4e43-41e6-9ba2-331990b516bd" path="/var/lib/kubelet/pods/5db2e99c-4e43-41e6-9ba2-331990b516bd/volumes" Nov 26 00:45:42 crc kubenswrapper[4766]: I1126 00:45:42.554174 4766 generic.go:334] "Generic (PLEG): container finished" podID="de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf" containerID="83579b0bcf4b669c84c274a4beac436450174217ba2feda87383e0ff170b31d5" exitCode=0 Nov 26 00:45:42 crc kubenswrapper[4766]: I1126 00:45:42.554226 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" event={"ID":"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf","Type":"ContainerDied","Data":"83579b0bcf4b669c84c274a4beac436450174217ba2feda87383e0ff170b31d5"} Nov 26 00:45:49 crc kubenswrapper[4766]: I1126 00:45:49.623918 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6768b6c6f7-lm9pw" event={"ID":"3fdd0908-10b2-437d-92bf-ccdc345a0d80","Type":"ContainerStarted","Data":"ba503f01907c201051864ed941206e73a9a40bae28037dad9b3f7a5ae362d172"} Nov 26 00:45:49 crc kubenswrapper[4766]: I1126 00:45:49.624420 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:45:49 crc kubenswrapper[4766]: I1126 00:45:49.625307 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5fd6484f46-cqsxg" event={"ID":"be4073df-dee9-4a70-8911-b23a317a0a57","Type":"ContainerStarted","Data":"90dc77e70c594437586ef3767dffa9714e4a31dd93c0403310df701b82e0fa6c"} Nov 26 00:45:49 crc kubenswrapper[4766]: I1126 00:45:49.639853 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6768b6c6f7-lm9pw" podStartSLOduration=12.639837488 podStartE2EDuration="12.639837488s" podCreationTimestamp="2025-11-26 00:45:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:45:49.639331176 +0000 UTC m=+1330.488101606" watchObservedRunningTime="2025-11-26 00:45:49.639837488 +0000 UTC m=+1330.488607918" Nov 26 00:45:52 crc kubenswrapper[4766]: I1126 00:45:52.713756 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6b7ccbcb5b-rf9v9"] Nov 26 00:45:53 crc kubenswrapper[4766]: I1126 00:45:53.669209 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" event={"ID":"5655d266-3ea5-404f-901e-80da609b0319","Type":"ContainerStarted","Data":"e27792576fe55a0cbeddab5eeb51b4b860cd9fa26354c2e755d83273dd8a9701"} Nov 26 00:45:54 crc kubenswrapper[4766]: E1126 00:45:54.022045 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="55ec5b52-2985-4607-8c89-590f43a3096c" Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.679552 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" event={"ID":"5655d266-3ea5-404f-901e-80da609b0319","Type":"ContainerStarted","Data":"f58c09cf2902d9ae494fc6827e4cf1eb8a98436fc97542023024d3c1a191d697"} Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.680091 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" event={"ID":"5655d266-3ea5-404f-901e-80da609b0319","Type":"ContainerStarted","Data":"fc643aee064b98501c2a49c7d67401a25688052bc82a38e58ddb15a5e3286d71"} Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.681386 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.681486 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.684209 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7fd5c59d54-6ccpg" event={"ID":"81836728-9333-4234-8175-954682305e4e","Type":"ContainerStarted","Data":"fbed1025da7eb5efb3baf66a5d20c0c7a806f1c61d1a3ac50abcf4010959a011"} Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.684253 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7fd5c59d54-6ccpg" event={"ID":"81836728-9333-4234-8175-954682305e4e","Type":"ContainerStarted","Data":"d5250aa53957ec6326e8ec5b254c5bb1a4ece2daef19a9dc3f65601c7c358b49"} Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.686104 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" event={"ID":"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf","Type":"ContainerStarted","Data":"05a7e8985febc475efe59ece6286b70097ffaa8b37feb30ded6e1f4efbd2a67c"} Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.686308 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.687682 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5fd6484f46-cqsxg" event={"ID":"be4073df-dee9-4a70-8911-b23a317a0a57","Type":"ContainerStarted","Data":"3a44a973765d8f1d5ff1f73d5e17572c2942c4bec2ac26da8c5e4010252c5928"} Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.687865 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.689286 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8sbgx" event={"ID":"ffaf6d04-ed5d-4438-a285-bc4981448be5","Type":"ContainerStarted","Data":"e18a280668d8cba4f4a4e353024890c172f809b1751ab6ef5b6847fd11f4e2e3"} Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.689373 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.690961 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68b9476655-7svbz" event={"ID":"01cfcf71-49d4-498f-a9cc-5e9c65c08709","Type":"ContainerStarted","Data":"8376977924be450e65465a9e1bd002097032c20b68f0d7f329db63ef4b3287d4"} Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.691055 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68b9476655-7svbz" event={"ID":"01cfcf71-49d4-498f-a9cc-5e9c65c08709","Type":"ContainerStarted","Data":"9ed6c08b63e51a4f1c6b76d7ff4be4d02eb5cbf4d69ae429369b5b5d0ce1cd00"} Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.695963 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55ec5b52-2985-4607-8c89-590f43a3096c","Type":"ContainerStarted","Data":"837ce3c458b3e5b8741c04b01e433bd4697f9c705a967612d5e9d5d0f61170d2"} Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.696103 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="55ec5b52-2985-4607-8c89-590f43a3096c" containerName="ceilometer-notification-agent" containerID="cri-o://fe436118dd6795c5727832d6353c09cd384307b2bc3e597fdd9639d3635552cd" gracePeriod=30 Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.696121 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.696230 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="55ec5b52-2985-4607-8c89-590f43a3096c" containerName="proxy-httpd" containerID="cri-o://837ce3c458b3e5b8741c04b01e433bd4697f9c705a967612d5e9d5d0f61170d2" gracePeriod=30 Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.696304 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="55ec5b52-2985-4607-8c89-590f43a3096c" containerName="sg-core" containerID="cri-o://0d860a8dad08282eb7da2f8a9ca7f69f203d24091631f7cbd6c1aa9002e778f7" gracePeriod=30 Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.701210 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67fdfd759d-sqzs7" event={"ID":"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d","Type":"ContainerStarted","Data":"abaec1e74c1d959f142b7ae8c9c578a7be50a03c4e6f9a706debca9c5b7619b4"} Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.702536 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.703183 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.718644 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" podStartSLOduration=13.718627302 podStartE2EDuration="13.718627302s" podCreationTimestamp="2025-11-26 00:45:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:45:54.704145799 +0000 UTC m=+1335.552916229" watchObservedRunningTime="2025-11-26 00:45:54.718627302 +0000 UTC m=+1335.567397732" Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.760422 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-8sbgx" podStartSLOduration=3.447676721 podStartE2EDuration="52.760402019s" podCreationTimestamp="2025-11-26 00:45:02 +0000 UTC" firstStartedPulling="2025-11-26 00:45:04.386256551 +0000 UTC m=+1285.235026981" lastFinishedPulling="2025-11-26 00:45:53.698981849 +0000 UTC m=+1334.547752279" observedRunningTime="2025-11-26 00:45:54.748691205 +0000 UTC m=+1335.597461635" watchObservedRunningTime="2025-11-26 00:45:54.760402019 +0000 UTC m=+1335.609172449" Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.776262 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-68b9476655-7svbz" podStartSLOduration=2.994423402 podStartE2EDuration="17.776247576s" podCreationTimestamp="2025-11-26 00:45:37 +0000 UTC" firstStartedPulling="2025-11-26 00:45:38.807994209 +0000 UTC m=+1319.656764639" lastFinishedPulling="2025-11-26 00:45:53.589818383 +0000 UTC m=+1334.438588813" observedRunningTime="2025-11-26 00:45:54.769728243 +0000 UTC m=+1335.618498693" watchObservedRunningTime="2025-11-26 00:45:54.776247576 +0000 UTC m=+1335.625018006" Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.795889 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5fd6484f46-cqsxg" podStartSLOduration=17.795872338 podStartE2EDuration="17.795872338s" podCreationTimestamp="2025-11-26 00:45:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:45:54.792282778 +0000 UTC m=+1335.641053218" watchObservedRunningTime="2025-11-26 00:45:54.795872338 +0000 UTC m=+1335.644642768" Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.824401 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" podStartSLOduration=17.824378293 podStartE2EDuration="17.824378293s" podCreationTimestamp="2025-11-26 00:45:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:45:54.812041294 +0000 UTC m=+1335.660811724" watchObservedRunningTime="2025-11-26 00:45:54.824378293 +0000 UTC m=+1335.673148723" Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.861215 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-67fdfd759d-sqzs7" podStartSLOduration=17.861196756 podStartE2EDuration="17.861196756s" podCreationTimestamp="2025-11-26 00:45:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:45:54.836770954 +0000 UTC m=+1335.685541414" watchObservedRunningTime="2025-11-26 00:45:54.861196756 +0000 UTC m=+1335.709967186" Nov 26 00:45:54 crc kubenswrapper[4766]: I1126 00:45:54.868501 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7fd5c59d54-6ccpg" podStartSLOduration=2.862116146 podStartE2EDuration="17.868480939s" podCreationTimestamp="2025-11-26 00:45:37 +0000 UTC" firstStartedPulling="2025-11-26 00:45:38.598387544 +0000 UTC m=+1319.447157974" lastFinishedPulling="2025-11-26 00:45:53.604752347 +0000 UTC m=+1334.453522767" observedRunningTime="2025-11-26 00:45:54.861161935 +0000 UTC m=+1335.709932375" watchObservedRunningTime="2025-11-26 00:45:54.868480939 +0000 UTC m=+1335.717251369" Nov 26 00:45:55 crc kubenswrapper[4766]: I1126 00:45:55.712141 4766 generic.go:334] "Generic (PLEG): container finished" podID="55ec5b52-2985-4607-8c89-590f43a3096c" containerID="837ce3c458b3e5b8741c04b01e433bd4697f9c705a967612d5e9d5d0f61170d2" exitCode=0 Nov 26 00:45:55 crc kubenswrapper[4766]: I1126 00:45:55.712173 4766 generic.go:334] "Generic (PLEG): container finished" podID="55ec5b52-2985-4607-8c89-590f43a3096c" containerID="0d860a8dad08282eb7da2f8a9ca7f69f203d24091631f7cbd6c1aa9002e778f7" exitCode=2 Nov 26 00:45:55 crc kubenswrapper[4766]: I1126 00:45:55.712885 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55ec5b52-2985-4607-8c89-590f43a3096c","Type":"ContainerDied","Data":"837ce3c458b3e5b8741c04b01e433bd4697f9c705a967612d5e9d5d0f61170d2"} Nov 26 00:45:55 crc kubenswrapper[4766]: I1126 00:45:55.712910 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55ec5b52-2985-4607-8c89-590f43a3096c","Type":"ContainerDied","Data":"0d860a8dad08282eb7da2f8a9ca7f69f203d24091631f7cbd6c1aa9002e778f7"} Nov 26 00:45:56 crc kubenswrapper[4766]: I1126 00:45:56.743190 4766 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 00:45:56 crc kubenswrapper[4766]: I1126 00:45:56.743482 4766 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 00:45:56 crc kubenswrapper[4766]: I1126 00:45:56.994216 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:45:57 crc kubenswrapper[4766]: I1126 00:45:57.036392 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.107866 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.187206 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-lvp7w"] Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.187622 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" podUID="cbf480e0-f3e9-445c-a174-db6ce0967ed7" containerName="dnsmasq-dns" containerID="cri-o://467596d52fa9c032644d706448c18ccfc8ef04c98600e4de8d35a521162c60b2" gracePeriod=10 Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.405643 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.543245 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-config-data\") pod \"55ec5b52-2985-4607-8c89-590f43a3096c\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.543330 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-combined-ca-bundle\") pod \"55ec5b52-2985-4607-8c89-590f43a3096c\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.543361 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-sg-core-conf-yaml\") pod \"55ec5b52-2985-4607-8c89-590f43a3096c\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.543390 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-scripts\") pod \"55ec5b52-2985-4607-8c89-590f43a3096c\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.543594 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55ec5b52-2985-4607-8c89-590f43a3096c-log-httpd\") pod \"55ec5b52-2985-4607-8c89-590f43a3096c\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.543674 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrq6q\" (UniqueName: \"kubernetes.io/projected/55ec5b52-2985-4607-8c89-590f43a3096c-kube-api-access-wrq6q\") pod \"55ec5b52-2985-4607-8c89-590f43a3096c\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.543744 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55ec5b52-2985-4607-8c89-590f43a3096c-run-httpd\") pod \"55ec5b52-2985-4607-8c89-590f43a3096c\" (UID: \"55ec5b52-2985-4607-8c89-590f43a3096c\") " Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.545147 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55ec5b52-2985-4607-8c89-590f43a3096c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "55ec5b52-2985-4607-8c89-590f43a3096c" (UID: "55ec5b52-2985-4607-8c89-590f43a3096c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.546988 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55ec5b52-2985-4607-8c89-590f43a3096c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "55ec5b52-2985-4607-8c89-590f43a3096c" (UID: "55ec5b52-2985-4607-8c89-590f43a3096c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.558224 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55ec5b52-2985-4607-8c89-590f43a3096c-kube-api-access-wrq6q" (OuterVolumeSpecName: "kube-api-access-wrq6q") pod "55ec5b52-2985-4607-8c89-590f43a3096c" (UID: "55ec5b52-2985-4607-8c89-590f43a3096c"). InnerVolumeSpecName "kube-api-access-wrq6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.568910 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-scripts" (OuterVolumeSpecName: "scripts") pod "55ec5b52-2985-4607-8c89-590f43a3096c" (UID: "55ec5b52-2985-4607-8c89-590f43a3096c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.630440 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "55ec5b52-2985-4607-8c89-590f43a3096c" (UID: "55ec5b52-2985-4607-8c89-590f43a3096c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.633864 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "55ec5b52-2985-4607-8c89-590f43a3096c" (UID: "55ec5b52-2985-4607-8c89-590f43a3096c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.647250 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.647321 4766 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.647333 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.647344 4766 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55ec5b52-2985-4607-8c89-590f43a3096c-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.647381 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrq6q\" (UniqueName: \"kubernetes.io/projected/55ec5b52-2985-4607-8c89-590f43a3096c-kube-api-access-wrq6q\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.647390 4766 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55ec5b52-2985-4607-8c89-590f43a3096c-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.667606 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-config-data" (OuterVolumeSpecName: "config-data") pod "55ec5b52-2985-4607-8c89-590f43a3096c" (UID: "55ec5b52-2985-4607-8c89-590f43a3096c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.683131 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.749088 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sq4qd\" (UniqueName: \"kubernetes.io/projected/cbf480e0-f3e9-445c-a174-db6ce0967ed7-kube-api-access-sq4qd\") pod \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.749141 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-ovsdbserver-nb\") pod \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.749163 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-dns-swift-storage-0\") pod \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.749198 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-config\") pod \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.749244 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-dns-svc\") pod \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.749376 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-ovsdbserver-sb\") pod \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\" (UID: \"cbf480e0-f3e9-445c-a174-db6ce0967ed7\") " Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.749931 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55ec5b52-2985-4607-8c89-590f43a3096c-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.776949 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbf480e0-f3e9-445c-a174-db6ce0967ed7-kube-api-access-sq4qd" (OuterVolumeSpecName: "kube-api-access-sq4qd") pod "cbf480e0-f3e9-445c-a174-db6ce0967ed7" (UID: "cbf480e0-f3e9-445c-a174-db6ce0967ed7"). InnerVolumeSpecName "kube-api-access-sq4qd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.789705 4766 generic.go:334] "Generic (PLEG): container finished" podID="55ec5b52-2985-4607-8c89-590f43a3096c" containerID="fe436118dd6795c5727832d6353c09cd384307b2bc3e597fdd9639d3635552cd" exitCode=0 Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.789925 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.789982 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55ec5b52-2985-4607-8c89-590f43a3096c","Type":"ContainerDied","Data":"fe436118dd6795c5727832d6353c09cd384307b2bc3e597fdd9639d3635552cd"} Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.790054 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55ec5b52-2985-4607-8c89-590f43a3096c","Type":"ContainerDied","Data":"8fd5a36c2222314544c095bc82df4d571ffde4f52c7fc264fa3943fbc8a09444"} Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.790078 4766 scope.go:117] "RemoveContainer" containerID="837ce3c458b3e5b8741c04b01e433bd4697f9c705a967612d5e9d5d0f61170d2" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.794423 4766 generic.go:334] "Generic (PLEG): container finished" podID="cbf480e0-f3e9-445c-a174-db6ce0967ed7" containerID="467596d52fa9c032644d706448c18ccfc8ef04c98600e4de8d35a521162c60b2" exitCode=0 Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.794470 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" event={"ID":"cbf480e0-f3e9-445c-a174-db6ce0967ed7","Type":"ContainerDied","Data":"467596d52fa9c032644d706448c18ccfc8ef04c98600e4de8d35a521162c60b2"} Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.794502 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" event={"ID":"cbf480e0-f3e9-445c-a174-db6ce0967ed7","Type":"ContainerDied","Data":"570236219f9aaae0cbf842f0e3c63696c270c9112022bb6fb595755785a0facd"} Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.794552 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-lvp7w" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.812826 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cbf480e0-f3e9-445c-a174-db6ce0967ed7" (UID: "cbf480e0-f3e9-445c-a174-db6ce0967ed7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.823378 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cbf480e0-f3e9-445c-a174-db6ce0967ed7" (UID: "cbf480e0-f3e9-445c-a174-db6ce0967ed7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.842926 4766 scope.go:117] "RemoveContainer" containerID="0d860a8dad08282eb7da2f8a9ca7f69f203d24091631f7cbd6c1aa9002e778f7" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.851404 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sq4qd\" (UniqueName: \"kubernetes.io/projected/cbf480e0-f3e9-445c-a174-db6ce0967ed7-kube-api-access-sq4qd\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.851439 4766 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.851451 4766 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.876751 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.878570 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cbf480e0-f3e9-445c-a174-db6ce0967ed7" (UID: "cbf480e0-f3e9-445c-a174-db6ce0967ed7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.884262 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-config" (OuterVolumeSpecName: "config") pod "cbf480e0-f3e9-445c-a174-db6ce0967ed7" (UID: "cbf480e0-f3e9-445c-a174-db6ce0967ed7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.887681 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.902147 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:45:58 crc kubenswrapper[4766]: E1126 00:45:58.902557 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55ec5b52-2985-4607-8c89-590f43a3096c" containerName="proxy-httpd" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.902571 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="55ec5b52-2985-4607-8c89-590f43a3096c" containerName="proxy-httpd" Nov 26 00:45:58 crc kubenswrapper[4766]: E1126 00:45:58.902596 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbf480e0-f3e9-445c-a174-db6ce0967ed7" containerName="init" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.902602 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbf480e0-f3e9-445c-a174-db6ce0967ed7" containerName="init" Nov 26 00:45:58 crc kubenswrapper[4766]: E1126 00:45:58.902609 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55ec5b52-2985-4607-8c89-590f43a3096c" containerName="sg-core" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.902615 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="55ec5b52-2985-4607-8c89-590f43a3096c" containerName="sg-core" Nov 26 00:45:58 crc kubenswrapper[4766]: E1126 00:45:58.902641 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbf480e0-f3e9-445c-a174-db6ce0967ed7" containerName="dnsmasq-dns" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.902664 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbf480e0-f3e9-445c-a174-db6ce0967ed7" containerName="dnsmasq-dns" Nov 26 00:45:58 crc kubenswrapper[4766]: E1126 00:45:58.902672 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55ec5b52-2985-4607-8c89-590f43a3096c" containerName="ceilometer-notification-agent" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.902678 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="55ec5b52-2985-4607-8c89-590f43a3096c" containerName="ceilometer-notification-agent" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.902850 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="55ec5b52-2985-4607-8c89-590f43a3096c" containerName="ceilometer-notification-agent" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.902861 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbf480e0-f3e9-445c-a174-db6ce0967ed7" containerName="dnsmasq-dns" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.902879 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="55ec5b52-2985-4607-8c89-590f43a3096c" containerName="sg-core" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.902891 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="55ec5b52-2985-4607-8c89-590f43a3096c" containerName="proxy-httpd" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.904561 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.911762 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.913620 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.913760 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.929260 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cbf480e0-f3e9-445c-a174-db6ce0967ed7" (UID: "cbf480e0-f3e9-445c-a174-db6ce0967ed7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.941966 4766 scope.go:117] "RemoveContainer" containerID="fe436118dd6795c5727832d6353c09cd384307b2bc3e597fdd9639d3635552cd" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.952828 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.952866 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:58 crc kubenswrapper[4766]: I1126 00:45:58.952877 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbf480e0-f3e9-445c-a174-db6ce0967ed7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.039438 4766 scope.go:117] "RemoveContainer" containerID="837ce3c458b3e5b8741c04b01e433bd4697f9c705a967612d5e9d5d0f61170d2" Nov 26 00:45:59 crc kubenswrapper[4766]: E1126 00:45:59.044153 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"837ce3c458b3e5b8741c04b01e433bd4697f9c705a967612d5e9d5d0f61170d2\": container with ID starting with 837ce3c458b3e5b8741c04b01e433bd4697f9c705a967612d5e9d5d0f61170d2 not found: ID does not exist" containerID="837ce3c458b3e5b8741c04b01e433bd4697f9c705a967612d5e9d5d0f61170d2" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.044202 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"837ce3c458b3e5b8741c04b01e433bd4697f9c705a967612d5e9d5d0f61170d2"} err="failed to get container status \"837ce3c458b3e5b8741c04b01e433bd4697f9c705a967612d5e9d5d0f61170d2\": rpc error: code = NotFound desc = could not find container \"837ce3c458b3e5b8741c04b01e433bd4697f9c705a967612d5e9d5d0f61170d2\": container with ID starting with 837ce3c458b3e5b8741c04b01e433bd4697f9c705a967612d5e9d5d0f61170d2 not found: ID does not exist" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.044226 4766 scope.go:117] "RemoveContainer" containerID="0d860a8dad08282eb7da2f8a9ca7f69f203d24091631f7cbd6c1aa9002e778f7" Nov 26 00:45:59 crc kubenswrapper[4766]: E1126 00:45:59.044603 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d860a8dad08282eb7da2f8a9ca7f69f203d24091631f7cbd6c1aa9002e778f7\": container with ID starting with 0d860a8dad08282eb7da2f8a9ca7f69f203d24091631f7cbd6c1aa9002e778f7 not found: ID does not exist" containerID="0d860a8dad08282eb7da2f8a9ca7f69f203d24091631f7cbd6c1aa9002e778f7" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.044624 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d860a8dad08282eb7da2f8a9ca7f69f203d24091631f7cbd6c1aa9002e778f7"} err="failed to get container status \"0d860a8dad08282eb7da2f8a9ca7f69f203d24091631f7cbd6c1aa9002e778f7\": rpc error: code = NotFound desc = could not find container \"0d860a8dad08282eb7da2f8a9ca7f69f203d24091631f7cbd6c1aa9002e778f7\": container with ID starting with 0d860a8dad08282eb7da2f8a9ca7f69f203d24091631f7cbd6c1aa9002e778f7 not found: ID does not exist" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.044637 4766 scope.go:117] "RemoveContainer" containerID="fe436118dd6795c5727832d6353c09cd384307b2bc3e597fdd9639d3635552cd" Nov 26 00:45:59 crc kubenswrapper[4766]: E1126 00:45:59.044866 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe436118dd6795c5727832d6353c09cd384307b2bc3e597fdd9639d3635552cd\": container with ID starting with fe436118dd6795c5727832d6353c09cd384307b2bc3e597fdd9639d3635552cd not found: ID does not exist" containerID="fe436118dd6795c5727832d6353c09cd384307b2bc3e597fdd9639d3635552cd" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.044887 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe436118dd6795c5727832d6353c09cd384307b2bc3e597fdd9639d3635552cd"} err="failed to get container status \"fe436118dd6795c5727832d6353c09cd384307b2bc3e597fdd9639d3635552cd\": rpc error: code = NotFound desc = could not find container \"fe436118dd6795c5727832d6353c09cd384307b2bc3e597fdd9639d3635552cd\": container with ID starting with fe436118dd6795c5727832d6353c09cd384307b2bc3e597fdd9639d3635552cd not found: ID does not exist" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.044905 4766 scope.go:117] "RemoveContainer" containerID="467596d52fa9c032644d706448c18ccfc8ef04c98600e4de8d35a521162c60b2" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.054180 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.054231 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlf77\" (UniqueName: \"kubernetes.io/projected/85f22975-acba-4af1-8787-124a69cb68e1-kube-api-access-hlf77\") pod \"ceilometer-0\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.054260 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.054383 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-config-data\") pod \"ceilometer-0\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.054434 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85f22975-acba-4af1-8787-124a69cb68e1-run-httpd\") pod \"ceilometer-0\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.054472 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85f22975-acba-4af1-8787-124a69cb68e1-log-httpd\") pod \"ceilometer-0\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.054595 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-scripts\") pod \"ceilometer-0\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.064870 4766 scope.go:117] "RemoveContainer" containerID="eccfbc38abe6f8428984fb53b569287a7eb4bd7a3e2d408b5b1a65fbef90408c" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.085942 4766 scope.go:117] "RemoveContainer" containerID="467596d52fa9c032644d706448c18ccfc8ef04c98600e4de8d35a521162c60b2" Nov 26 00:45:59 crc kubenswrapper[4766]: E1126 00:45:59.086342 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"467596d52fa9c032644d706448c18ccfc8ef04c98600e4de8d35a521162c60b2\": container with ID starting with 467596d52fa9c032644d706448c18ccfc8ef04c98600e4de8d35a521162c60b2 not found: ID does not exist" containerID="467596d52fa9c032644d706448c18ccfc8ef04c98600e4de8d35a521162c60b2" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.086387 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"467596d52fa9c032644d706448c18ccfc8ef04c98600e4de8d35a521162c60b2"} err="failed to get container status \"467596d52fa9c032644d706448c18ccfc8ef04c98600e4de8d35a521162c60b2\": rpc error: code = NotFound desc = could not find container \"467596d52fa9c032644d706448c18ccfc8ef04c98600e4de8d35a521162c60b2\": container with ID starting with 467596d52fa9c032644d706448c18ccfc8ef04c98600e4de8d35a521162c60b2 not found: ID does not exist" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.086403 4766 scope.go:117] "RemoveContainer" containerID="eccfbc38abe6f8428984fb53b569287a7eb4bd7a3e2d408b5b1a65fbef90408c" Nov 26 00:45:59 crc kubenswrapper[4766]: E1126 00:45:59.086830 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eccfbc38abe6f8428984fb53b569287a7eb4bd7a3e2d408b5b1a65fbef90408c\": container with ID starting with eccfbc38abe6f8428984fb53b569287a7eb4bd7a3e2d408b5b1a65fbef90408c not found: ID does not exist" containerID="eccfbc38abe6f8428984fb53b569287a7eb4bd7a3e2d408b5b1a65fbef90408c" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.086876 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eccfbc38abe6f8428984fb53b569287a7eb4bd7a3e2d408b5b1a65fbef90408c"} err="failed to get container status \"eccfbc38abe6f8428984fb53b569287a7eb4bd7a3e2d408b5b1a65fbef90408c\": rpc error: code = NotFound desc = could not find container \"eccfbc38abe6f8428984fb53b569287a7eb4bd7a3e2d408b5b1a65fbef90408c\": container with ID starting with eccfbc38abe6f8428984fb53b569287a7eb4bd7a3e2d408b5b1a65fbef90408c not found: ID does not exist" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.141637 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-lvp7w"] Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.153363 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-lvp7w"] Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.167921 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.167986 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlf77\" (UniqueName: \"kubernetes.io/projected/85f22975-acba-4af1-8787-124a69cb68e1-kube-api-access-hlf77\") pod \"ceilometer-0\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.168040 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.168166 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-config-data\") pod \"ceilometer-0\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.168333 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85f22975-acba-4af1-8787-124a69cb68e1-run-httpd\") pod \"ceilometer-0\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.168702 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85f22975-acba-4af1-8787-124a69cb68e1-log-httpd\") pod \"ceilometer-0\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.168845 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85f22975-acba-4af1-8787-124a69cb68e1-run-httpd\") pod \"ceilometer-0\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.169031 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-scripts\") pod \"ceilometer-0\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.169056 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85f22975-acba-4af1-8787-124a69cb68e1-log-httpd\") pod \"ceilometer-0\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.172818 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-config-data\") pod \"ceilometer-0\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.172824 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.173462 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-scripts\") pod \"ceilometer-0\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.193041 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.195457 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlf77\" (UniqueName: \"kubernetes.io/projected/85f22975-acba-4af1-8787-124a69cb68e1-kube-api-access-hlf77\") pod \"ceilometer-0\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.324375 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.803686 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.810091 4766 generic.go:334] "Generic (PLEG): container finished" podID="ffaf6d04-ed5d-4438-a285-bc4981448be5" containerID="e18a280668d8cba4f4a4e353024890c172f809b1751ab6ef5b6847fd11f4e2e3" exitCode=0 Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.810167 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8sbgx" event={"ID":"ffaf6d04-ed5d-4438-a285-bc4981448be5","Type":"ContainerDied","Data":"e18a280668d8cba4f4a4e353024890c172f809b1751ab6ef5b6847fd11f4e2e3"} Nov 26 00:45:59 crc kubenswrapper[4766]: W1126 00:45:59.816761 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85f22975_acba_4af1_8787_124a69cb68e1.slice/crio-ad698c6d5cf8113a583acc6ceeddde765e77c38680960fd925d00c490af7d4c9 WatchSource:0}: Error finding container ad698c6d5cf8113a583acc6ceeddde765e77c38680960fd925d00c490af7d4c9: Status 404 returned error can't find the container with id ad698c6d5cf8113a583acc6ceeddde765e77c38680960fd925d00c490af7d4c9 Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.842406 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55ec5b52-2985-4607-8c89-590f43a3096c" path="/var/lib/kubelet/pods/55ec5b52-2985-4607-8c89-590f43a3096c/volumes" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.843170 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbf480e0-f3e9-445c-a174-db6ce0967ed7" path="/var/lib/kubelet/pods/cbf480e0-f3e9-445c-a174-db6ce0967ed7/volumes" Nov 26 00:45:59 crc kubenswrapper[4766]: I1126 00:45:59.957276 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:46:00 crc kubenswrapper[4766]: I1126 00:46:00.828137 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85f22975-acba-4af1-8787-124a69cb68e1","Type":"ContainerStarted","Data":"8fb365e52e88bc33ca948883a43b83a045fad7d8bb2ec15511f497bc20abefe7"} Nov 26 00:46:00 crc kubenswrapper[4766]: I1126 00:46:00.828176 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85f22975-acba-4af1-8787-124a69cb68e1","Type":"ContainerStarted","Data":"ad698c6d5cf8113a583acc6ceeddde765e77c38680960fd925d00c490af7d4c9"} Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.175295 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.312443 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ffaf6d04-ed5d-4438-a285-bc4981448be5-etc-machine-id\") pod \"ffaf6d04-ed5d-4438-a285-bc4981448be5\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.312544 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgxb5\" (UniqueName: \"kubernetes.io/projected/ffaf6d04-ed5d-4438-a285-bc4981448be5-kube-api-access-jgxb5\") pod \"ffaf6d04-ed5d-4438-a285-bc4981448be5\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.312562 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffaf6d04-ed5d-4438-a285-bc4981448be5-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ffaf6d04-ed5d-4438-a285-bc4981448be5" (UID: "ffaf6d04-ed5d-4438-a285-bc4981448be5"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.312703 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-db-sync-config-data\") pod \"ffaf6d04-ed5d-4438-a285-bc4981448be5\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.312743 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-config-data\") pod \"ffaf6d04-ed5d-4438-a285-bc4981448be5\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.312785 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-scripts\") pod \"ffaf6d04-ed5d-4438-a285-bc4981448be5\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.312912 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-combined-ca-bundle\") pod \"ffaf6d04-ed5d-4438-a285-bc4981448be5\" (UID: \"ffaf6d04-ed5d-4438-a285-bc4981448be5\") " Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.313591 4766 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ffaf6d04-ed5d-4438-a285-bc4981448be5-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.317535 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-scripts" (OuterVolumeSpecName: "scripts") pod "ffaf6d04-ed5d-4438-a285-bc4981448be5" (UID: "ffaf6d04-ed5d-4438-a285-bc4981448be5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.318040 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ffaf6d04-ed5d-4438-a285-bc4981448be5" (UID: "ffaf6d04-ed5d-4438-a285-bc4981448be5"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.320744 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffaf6d04-ed5d-4438-a285-bc4981448be5-kube-api-access-jgxb5" (OuterVolumeSpecName: "kube-api-access-jgxb5") pod "ffaf6d04-ed5d-4438-a285-bc4981448be5" (UID: "ffaf6d04-ed5d-4438-a285-bc4981448be5"). InnerVolumeSpecName "kube-api-access-jgxb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.379791 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ffaf6d04-ed5d-4438-a285-bc4981448be5" (UID: "ffaf6d04-ed5d-4438-a285-bc4981448be5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.404947 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-config-data" (OuterVolumeSpecName: "config-data") pod "ffaf6d04-ed5d-4438-a285-bc4981448be5" (UID: "ffaf6d04-ed5d-4438-a285-bc4981448be5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.415285 4766 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.415347 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.415361 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.415371 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffaf6d04-ed5d-4438-a285-bc4981448be5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.415380 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgxb5\" (UniqueName: \"kubernetes.io/projected/ffaf6d04-ed5d-4438-a285-bc4981448be5-kube-api-access-jgxb5\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.843519 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8sbgx" Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.846601 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8sbgx" event={"ID":"ffaf6d04-ed5d-4438-a285-bc4981448be5","Type":"ContainerDied","Data":"b67e7262cd6bc785308ebef5bbf9eb60677a4390794ea24c07d3fe0790938608"} Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.846672 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b67e7262cd6bc785308ebef5bbf9eb60677a4390794ea24c07d3fe0790938608" Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.846687 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85f22975-acba-4af1-8787-124a69cb68e1","Type":"ContainerStarted","Data":"48d5602c1bf28f3d5d1828b4c020f553a9f7122a9bc64b06339dbed445c30f1a"} Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.846703 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85f22975-acba-4af1-8787-124a69cb68e1","Type":"ContainerStarted","Data":"2b52ab898b9b6da7a8b73b6a7b2491a7f5ae1f9e232a31b3bad2d13612f5f31d"} Nov 26 00:46:01 crc kubenswrapper[4766]: I1126 00:46:01.926067 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7bbf79c884-hg8wv" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.273581 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 26 00:46:02 crc kubenswrapper[4766]: E1126 00:46:02.274237 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffaf6d04-ed5d-4438-a285-bc4981448be5" containerName="cinder-db-sync" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.274254 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffaf6d04-ed5d-4438-a285-bc4981448be5" containerName="cinder-db-sync" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.274447 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffaf6d04-ed5d-4438-a285-bc4981448be5" containerName="cinder-db-sync" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.275476 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.297236 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.298785 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.299000 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-wcjrw" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.311732 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.311884 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.333746 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-config-data\") pod \"cinder-scheduler-0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.333846 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-scripts\") pod \"cinder-scheduler-0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.333885 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.333909 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcn86\" (UniqueName: \"kubernetes.io/projected/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-kube-api-access-gcn86\") pod \"cinder-scheduler-0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.333926 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.333941 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.345215 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-dpnb6"] Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.347283 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.377111 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-dpnb6"] Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.437438 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-dns-svc\") pod \"dnsmasq-dns-6578955fd5-dpnb6\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.437483 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-config\") pod \"dnsmasq-dns-6578955fd5-dpnb6\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.437506 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fwcc\" (UniqueName: \"kubernetes.io/projected/7cb97caa-8678-4b20-96e7-0ddf811bcc38-kube-api-access-5fwcc\") pod \"dnsmasq-dns-6578955fd5-dpnb6\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.437541 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-dpnb6\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.437568 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-scripts\") pod \"cinder-scheduler-0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.437595 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-dpnb6\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.437636 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.437677 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcn86\" (UniqueName: \"kubernetes.io/projected/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-kube-api-access-gcn86\") pod \"cinder-scheduler-0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.437695 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.437712 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.438096 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-dpnb6\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.438162 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-config-data\") pod \"cinder-scheduler-0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.441922 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.444563 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-scripts\") pod \"cinder-scheduler-0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.472767 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-config-data\") pod \"cinder-scheduler-0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.474369 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.475767 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.494104 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcn86\" (UniqueName: \"kubernetes.io/projected/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-kube-api-access-gcn86\") pod \"cinder-scheduler-0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.540099 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-dpnb6\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.540209 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-dns-svc\") pod \"dnsmasq-dns-6578955fd5-dpnb6\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.540235 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-config\") pod \"dnsmasq-dns-6578955fd5-dpnb6\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.540258 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fwcc\" (UniqueName: \"kubernetes.io/projected/7cb97caa-8678-4b20-96e7-0ddf811bcc38-kube-api-access-5fwcc\") pod \"dnsmasq-dns-6578955fd5-dpnb6\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.540306 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-dpnb6\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.540325 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-dpnb6\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.541244 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-dpnb6\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.541347 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-config\") pod \"dnsmasq-dns-6578955fd5-dpnb6\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.542001 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-dpnb6\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.542417 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-dpnb6\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.543786 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-dns-svc\") pod \"dnsmasq-dns-6578955fd5-dpnb6\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.580436 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fwcc\" (UniqueName: \"kubernetes.io/projected/7cb97caa-8678-4b20-96e7-0ddf811bcc38-kube-api-access-5fwcc\") pod \"dnsmasq-dns-6578955fd5-dpnb6\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.607701 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.608269 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.609421 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.612540 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.617180 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.706546 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.745009 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rf9g\" (UniqueName: \"kubernetes.io/projected/0f76ca58-391f-43c9-810b-f80f94213a31-kube-api-access-7rf9g\") pod \"cinder-api-0\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.745169 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-config-data\") pod \"cinder-api-0\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.745222 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0f76ca58-391f-43c9-810b-f80f94213a31-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.745324 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-config-data-custom\") pod \"cinder-api-0\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.745448 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-scripts\") pod \"cinder-api-0\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.745490 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f76ca58-391f-43c9-810b-f80f94213a31-logs\") pod \"cinder-api-0\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.745519 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.846991 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f76ca58-391f-43c9-810b-f80f94213a31-logs\") pod \"cinder-api-0\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.847331 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.847365 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rf9g\" (UniqueName: \"kubernetes.io/projected/0f76ca58-391f-43c9-810b-f80f94213a31-kube-api-access-7rf9g\") pod \"cinder-api-0\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.847400 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-config-data\") pod \"cinder-api-0\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.847436 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0f76ca58-391f-43c9-810b-f80f94213a31-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.847503 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-config-data-custom\") pod \"cinder-api-0\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.847616 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-scripts\") pod \"cinder-api-0\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.853112 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0f76ca58-391f-43c9-810b-f80f94213a31-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.853296 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f76ca58-391f-43c9-810b-f80f94213a31-logs\") pod \"cinder-api-0\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.854428 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-scripts\") pod \"cinder-api-0\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.863296 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-config-data-custom\") pod \"cinder-api-0\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.864230 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.869698 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-config-data\") pod \"cinder-api-0\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.878324 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rf9g\" (UniqueName: \"kubernetes.io/projected/0f76ca58-391f-43c9-810b-f80f94213a31-kube-api-access-7rf9g\") pod \"cinder-api-0\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " pod="openstack/cinder-api-0" Nov 26 00:46:02 crc kubenswrapper[4766]: I1126 00:46:02.932552 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 26 00:46:03 crc kubenswrapper[4766]: W1126 00:46:03.208232 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod645620df_fc9e_44ec_b035_a2b0b6a3b7d0.slice/crio-5d11aec0f8bc21aebde7b77384683b9547928ccebcf394522fd2f3d83f0eb8bc WatchSource:0}: Error finding container 5d11aec0f8bc21aebde7b77384683b9547928ccebcf394522fd2f3d83f0eb8bc: Status 404 returned error can't find the container with id 5d11aec0f8bc21aebde7b77384683b9547928ccebcf394522fd2f3d83f0eb8bc Nov 26 00:46:03 crc kubenswrapper[4766]: I1126 00:46:03.209634 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 26 00:46:03 crc kubenswrapper[4766]: I1126 00:46:03.353328 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-dpnb6"] Nov 26 00:46:03 crc kubenswrapper[4766]: I1126 00:46:03.567697 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 26 00:46:03 crc kubenswrapper[4766]: W1126 00:46:03.579719 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f76ca58_391f_43c9_810b_f80f94213a31.slice/crio-b04115395f9da31610eb16d10f8701131c2b93ad59ab125138b8538f04740754 WatchSource:0}: Error finding container b04115395f9da31610eb16d10f8701131c2b93ad59ab125138b8538f04740754: Status 404 returned error can't find the container with id b04115395f9da31610eb16d10f8701131c2b93ad59ab125138b8538f04740754 Nov 26 00:46:03 crc kubenswrapper[4766]: I1126 00:46:03.917536 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85f22975-acba-4af1-8787-124a69cb68e1","Type":"ContainerStarted","Data":"d24fe490bba51a44c3ebde3231849642624486ad1a4ea2d892d19770d1a77039"} Nov 26 00:46:03 crc kubenswrapper[4766]: I1126 00:46:03.918072 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 26 00:46:03 crc kubenswrapper[4766]: I1126 00:46:03.927832 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0f76ca58-391f-43c9-810b-f80f94213a31","Type":"ContainerStarted","Data":"b04115395f9da31610eb16d10f8701131c2b93ad59ab125138b8538f04740754"} Nov 26 00:46:03 crc kubenswrapper[4766]: I1126 00:46:03.937628 4766 generic.go:334] "Generic (PLEG): container finished" podID="7cb97caa-8678-4b20-96e7-0ddf811bcc38" containerID="9a6723f51bd3ee62f15560e4742b57b6c094e8ef945f848ce06b0cdab422ef9f" exitCode=0 Nov 26 00:46:03 crc kubenswrapper[4766]: I1126 00:46:03.937731 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" event={"ID":"7cb97caa-8678-4b20-96e7-0ddf811bcc38","Type":"ContainerDied","Data":"9a6723f51bd3ee62f15560e4742b57b6c094e8ef945f848ce06b0cdab422ef9f"} Nov 26 00:46:03 crc kubenswrapper[4766]: I1126 00:46:03.937771 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" event={"ID":"7cb97caa-8678-4b20-96e7-0ddf811bcc38","Type":"ContainerStarted","Data":"9298e63d560510a6bd5d7f9cb76b9fb9d4403eb49c2c756321f02de08ed029ef"} Nov 26 00:46:03 crc kubenswrapper[4766]: I1126 00:46:03.941141 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"645620df-fc9e-44ec-b035-a2b0b6a3b7d0","Type":"ContainerStarted","Data":"5d11aec0f8bc21aebde7b77384683b9547928ccebcf394522fd2f3d83f0eb8bc"} Nov 26 00:46:03 crc kubenswrapper[4766]: I1126 00:46:03.959207 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.79208954 podStartE2EDuration="5.959183068s" podCreationTimestamp="2025-11-26 00:45:58 +0000 UTC" firstStartedPulling="2025-11-26 00:45:59.819043936 +0000 UTC m=+1340.667814386" lastFinishedPulling="2025-11-26 00:46:02.986137484 +0000 UTC m=+1343.834907914" observedRunningTime="2025-11-26 00:46:03.951237579 +0000 UTC m=+1344.800008009" watchObservedRunningTime="2025-11-26 00:46:03.959183068 +0000 UTC m=+1344.807953498" Nov 26 00:46:04 crc kubenswrapper[4766]: I1126 00:46:04.146235 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-756fcfc485-ktbfm" Nov 26 00:46:04 crc kubenswrapper[4766]: I1126 00:46:04.280295 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7bbf79c884-hg8wv"] Nov 26 00:46:04 crc kubenswrapper[4766]: I1126 00:46:04.280836 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7bbf79c884-hg8wv" podUID="7ee29573-3f38-49d9-8c91-6c7ae3d2bd35" containerName="neutron-api" containerID="cri-o://e0db47f8b6f4b59ae417bc220def5e2f48777634002fe5de75bf3765bf0af817" gracePeriod=30 Nov 26 00:46:04 crc kubenswrapper[4766]: I1126 00:46:04.281380 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7bbf79c884-hg8wv" podUID="7ee29573-3f38-49d9-8c91-6c7ae3d2bd35" containerName="neutron-httpd" containerID="cri-o://19e9b43327a1616f211162f6aa898a52ef234db2201e5cee50375cf75587de00" gracePeriod=30 Nov 26 00:46:04 crc kubenswrapper[4766]: I1126 00:46:04.758592 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:46:04 crc kubenswrapper[4766]: I1126 00:46:04.890962 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 26 00:46:04 crc kubenswrapper[4766]: I1126 00:46:04.974925 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0f76ca58-391f-43c9-810b-f80f94213a31","Type":"ContainerStarted","Data":"76a3fa02501f1c62ee260c108e730ca16307f8c74245cc6f3b2f7a91e1e162c7"} Nov 26 00:46:04 crc kubenswrapper[4766]: I1126 00:46:04.979961 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" event={"ID":"7cb97caa-8678-4b20-96e7-0ddf811bcc38","Type":"ContainerStarted","Data":"cff0ffc937f2588a8daa1295e1eaefe470b14765cc13965c55dba86981dfc9fc"} Nov 26 00:46:04 crc kubenswrapper[4766]: I1126 00:46:04.981739 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:04 crc kubenswrapper[4766]: I1126 00:46:04.999177 4766 generic.go:334] "Generic (PLEG): container finished" podID="7ee29573-3f38-49d9-8c91-6c7ae3d2bd35" containerID="19e9b43327a1616f211162f6aa898a52ef234db2201e5cee50375cf75587de00" exitCode=0 Nov 26 00:46:05 crc kubenswrapper[4766]: I1126 00:46:05.000027 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bbf79c884-hg8wv" event={"ID":"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35","Type":"ContainerDied","Data":"19e9b43327a1616f211162f6aa898a52ef234db2201e5cee50375cf75587de00"} Nov 26 00:46:05 crc kubenswrapper[4766]: I1126 00:46:05.019508 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" podStartSLOduration=3.019478959 podStartE2EDuration="3.019478959s" podCreationTimestamp="2025-11-26 00:46:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:46:05.011091579 +0000 UTC m=+1345.859862009" watchObservedRunningTime="2025-11-26 00:46:05.019478959 +0000 UTC m=+1345.868249379" Nov 26 00:46:05 crc kubenswrapper[4766]: I1126 00:46:05.306505 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6b7ccbcb5b-rf9v9" Nov 26 00:46:05 crc kubenswrapper[4766]: I1126 00:46:05.404989 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-67fdfd759d-sqzs7"] Nov 26 00:46:05 crc kubenswrapper[4766]: I1126 00:46:05.405437 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-67fdfd759d-sqzs7" podUID="d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d" containerName="barbican-api-log" containerID="cri-o://0c2c7807bed04eacaab51baab389f6ad6e8d873b0bbbe0bac9f40f9297331099" gracePeriod=30 Nov 26 00:46:05 crc kubenswrapper[4766]: I1126 00:46:05.405787 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-67fdfd759d-sqzs7" podUID="d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d" containerName="barbican-api" containerID="cri-o://abaec1e74c1d959f142b7ae8c9c578a7be50a03c4e6f9a706debca9c5b7619b4" gracePeriod=30 Nov 26 00:46:06 crc kubenswrapper[4766]: I1126 00:46:06.012209 4766 generic.go:334] "Generic (PLEG): container finished" podID="d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d" containerID="0c2c7807bed04eacaab51baab389f6ad6e8d873b0bbbe0bac9f40f9297331099" exitCode=143 Nov 26 00:46:06 crc kubenswrapper[4766]: I1126 00:46:06.012547 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67fdfd759d-sqzs7" event={"ID":"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d","Type":"ContainerDied","Data":"0c2c7807bed04eacaab51baab389f6ad6e8d873b0bbbe0bac9f40f9297331099"} Nov 26 00:46:06 crc kubenswrapper[4766]: I1126 00:46:06.014797 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0f76ca58-391f-43c9-810b-f80f94213a31","Type":"ContainerStarted","Data":"d1c39cb41e9ec9c525837a97a31e77e47f4ff563f15ae600a44c0e8b141327d1"} Nov 26 00:46:06 crc kubenswrapper[4766]: I1126 00:46:06.014957 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0f76ca58-391f-43c9-810b-f80f94213a31" containerName="cinder-api-log" containerID="cri-o://76a3fa02501f1c62ee260c108e730ca16307f8c74245cc6f3b2f7a91e1e162c7" gracePeriod=30 Nov 26 00:46:06 crc kubenswrapper[4766]: I1126 00:46:06.015200 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 26 00:46:06 crc kubenswrapper[4766]: I1126 00:46:06.015450 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0f76ca58-391f-43c9-810b-f80f94213a31" containerName="cinder-api" containerID="cri-o://d1c39cb41e9ec9c525837a97a31e77e47f4ff563f15ae600a44c0e8b141327d1" gracePeriod=30 Nov 26 00:46:06 crc kubenswrapper[4766]: I1126 00:46:06.024933 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"645620df-fc9e-44ec-b035-a2b0b6a3b7d0","Type":"ContainerStarted","Data":"f41589082a4ba0a4af8f39f96c8a863e223f64645064fa84e742af76a2734a0d"} Nov 26 00:46:06 crc kubenswrapper[4766]: I1126 00:46:06.039003 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.038978388 podStartE2EDuration="4.038978388s" podCreationTimestamp="2025-11-26 00:46:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:46:06.033016979 +0000 UTC m=+1346.881787409" watchObservedRunningTime="2025-11-26 00:46:06.038978388 +0000 UTC m=+1346.887748818" Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.037964 4766 generic.go:334] "Generic (PLEG): container finished" podID="7ee29573-3f38-49d9-8c91-6c7ae3d2bd35" containerID="e0db47f8b6f4b59ae417bc220def5e2f48777634002fe5de75bf3765bf0af817" exitCode=0 Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.038046 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bbf79c884-hg8wv" event={"ID":"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35","Type":"ContainerDied","Data":"e0db47f8b6f4b59ae417bc220def5e2f48777634002fe5de75bf3765bf0af817"} Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.040618 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"645620df-fc9e-44ec-b035-a2b0b6a3b7d0","Type":"ContainerStarted","Data":"cccee7808c8f489ebb681b487a50eb11e480e8c26e414b30f31dbf4dfce60b22"} Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.042880 4766 generic.go:334] "Generic (PLEG): container finished" podID="0f76ca58-391f-43c9-810b-f80f94213a31" containerID="76a3fa02501f1c62ee260c108e730ca16307f8c74245cc6f3b2f7a91e1e162c7" exitCode=143 Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.042965 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0f76ca58-391f-43c9-810b-f80f94213a31","Type":"ContainerDied","Data":"76a3fa02501f1c62ee260c108e730ca16307f8c74245cc6f3b2f7a91e1e162c7"} Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.072825 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.187379278 podStartE2EDuration="5.072798725s" podCreationTimestamp="2025-11-26 00:46:02 +0000 UTC" firstStartedPulling="2025-11-26 00:46:03.234208853 +0000 UTC m=+1344.082979273" lastFinishedPulling="2025-11-26 00:46:04.11962829 +0000 UTC m=+1344.968398720" observedRunningTime="2025-11-26 00:46:07.061146473 +0000 UTC m=+1347.909916903" watchObservedRunningTime="2025-11-26 00:46:07.072798725 +0000 UTC m=+1347.921569155" Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.474945 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bbf79c884-hg8wv" Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.608414 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.615394 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-combined-ca-bundle\") pod \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\" (UID: \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\") " Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.615506 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87f9q\" (UniqueName: \"kubernetes.io/projected/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-kube-api-access-87f9q\") pod \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\" (UID: \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\") " Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.615788 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-ovndb-tls-certs\") pod \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\" (UID: \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\") " Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.615993 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-httpd-config\") pod \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\" (UID: \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\") " Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.616120 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-config\") pod \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\" (UID: \"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35\") " Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.623605 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-kube-api-access-87f9q" (OuterVolumeSpecName: "kube-api-access-87f9q") pod "7ee29573-3f38-49d9-8c91-6c7ae3d2bd35" (UID: "7ee29573-3f38-49d9-8c91-6c7ae3d2bd35"). InnerVolumeSpecName "kube-api-access-87f9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.626784 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "7ee29573-3f38-49d9-8c91-6c7ae3d2bd35" (UID: "7ee29573-3f38-49d9-8c91-6c7ae3d2bd35"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.698404 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ee29573-3f38-49d9-8c91-6c7ae3d2bd35" (UID: "7ee29573-3f38-49d9-8c91-6c7ae3d2bd35"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.708499 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-config" (OuterVolumeSpecName: "config") pod "7ee29573-3f38-49d9-8c91-6c7ae3d2bd35" (UID: "7ee29573-3f38-49d9-8c91-6c7ae3d2bd35"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.719047 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.720448 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.720526 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87f9q\" (UniqueName: \"kubernetes.io/projected/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-kube-api-access-87f9q\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.720585 4766 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.768139 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "7ee29573-3f38-49d9-8c91-6c7ae3d2bd35" (UID: "7ee29573-3f38-49d9-8c91-6c7ae3d2bd35"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:07 crc kubenswrapper[4766]: I1126 00:46:07.822858 4766 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:08 crc kubenswrapper[4766]: I1126 00:46:08.054877 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bbf79c884-hg8wv" Nov 26 00:46:08 crc kubenswrapper[4766]: I1126 00:46:08.054916 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bbf79c884-hg8wv" event={"ID":"7ee29573-3f38-49d9-8c91-6c7ae3d2bd35","Type":"ContainerDied","Data":"7169e74c49370232b54c4158d36d4795c8429b59f4af4a4c1a75d8fa1f2cafc0"} Nov 26 00:46:08 crc kubenswrapper[4766]: I1126 00:46:08.054952 4766 scope.go:117] "RemoveContainer" containerID="19e9b43327a1616f211162f6aa898a52ef234db2201e5cee50375cf75587de00" Nov 26 00:46:08 crc kubenswrapper[4766]: I1126 00:46:08.077013 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7bbf79c884-hg8wv"] Nov 26 00:46:08 crc kubenswrapper[4766]: I1126 00:46:08.081102 4766 scope.go:117] "RemoveContainer" containerID="e0db47f8b6f4b59ae417bc220def5e2f48777634002fe5de75bf3765bf0af817" Nov 26 00:46:08 crc kubenswrapper[4766]: I1126 00:46:08.084931 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7bbf79c884-hg8wv"] Nov 26 00:46:08 crc kubenswrapper[4766]: I1126 00:46:08.535756 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5fd6484f46-cqsxg" Nov 26 00:46:08 crc kubenswrapper[4766]: I1126 00:46:08.646279 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-67fdfd759d-sqzs7" podUID="d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.185:9311/healthcheck\": read tcp 10.217.0.2:33648->10.217.0.185:9311: read: connection reset by peer" Nov 26 00:46:08 crc kubenswrapper[4766]: I1126 00:46:08.646564 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-67fdfd759d-sqzs7" podUID="d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.185:9311/healthcheck\": read tcp 10.217.0.2:33664->10.217.0.185:9311: read: connection reset by peer" Nov 26 00:46:09 crc kubenswrapper[4766]: I1126 00:46:09.084026 4766 generic.go:334] "Generic (PLEG): container finished" podID="d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d" containerID="abaec1e74c1d959f142b7ae8c9c578a7be50a03c4e6f9a706debca9c5b7619b4" exitCode=0 Nov 26 00:46:09 crc kubenswrapper[4766]: I1126 00:46:09.084080 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67fdfd759d-sqzs7" event={"ID":"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d","Type":"ContainerDied","Data":"abaec1e74c1d959f142b7ae8c9c578a7be50a03c4e6f9a706debca9c5b7619b4"} Nov 26 00:46:09 crc kubenswrapper[4766]: I1126 00:46:09.253471 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:46:09 crc kubenswrapper[4766]: I1126 00:46:09.353531 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-combined-ca-bundle\") pod \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\" (UID: \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\") " Nov 26 00:46:09 crc kubenswrapper[4766]: I1126 00:46:09.353610 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-config-data-custom\") pod \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\" (UID: \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\") " Nov 26 00:46:09 crc kubenswrapper[4766]: I1126 00:46:09.353706 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mv6r4\" (UniqueName: \"kubernetes.io/projected/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-kube-api-access-mv6r4\") pod \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\" (UID: \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\") " Nov 26 00:46:09 crc kubenswrapper[4766]: I1126 00:46:09.353735 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-logs\") pod \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\" (UID: \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\") " Nov 26 00:46:09 crc kubenswrapper[4766]: I1126 00:46:09.353784 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-config-data\") pod \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\" (UID: \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\") " Nov 26 00:46:09 crc kubenswrapper[4766]: I1126 00:46:09.354234 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-logs" (OuterVolumeSpecName: "logs") pod "d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d" (UID: "d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:46:09 crc kubenswrapper[4766]: I1126 00:46:09.386822 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d" (UID: "d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:09 crc kubenswrapper[4766]: I1126 00:46:09.388541 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-kube-api-access-mv6r4" (OuterVolumeSpecName: "kube-api-access-mv6r4") pod "d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d" (UID: "d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d"). InnerVolumeSpecName "kube-api-access-mv6r4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:46:09 crc kubenswrapper[4766]: E1126 00:46:09.415001 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-combined-ca-bundle podName:d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d nodeName:}" failed. No retries permitted until 2025-11-26 00:46:09.914969243 +0000 UTC m=+1350.763739673 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-combined-ca-bundle") pod "d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d" (UID: "d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d") : error deleting /var/lib/kubelet/pods/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d/volume-subpaths: remove /var/lib/kubelet/pods/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d/volume-subpaths: no such file or directory Nov 26 00:46:09 crc kubenswrapper[4766]: I1126 00:46:09.420904 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-config-data" (OuterVolumeSpecName: "config-data") pod "d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d" (UID: "d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:09 crc kubenswrapper[4766]: I1126 00:46:09.455991 4766 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:09 crc kubenswrapper[4766]: I1126 00:46:09.456037 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mv6r4\" (UniqueName: \"kubernetes.io/projected/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-kube-api-access-mv6r4\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:09 crc kubenswrapper[4766]: I1126 00:46:09.456047 4766 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-logs\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:09 crc kubenswrapper[4766]: I1126 00:46:09.456058 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:09 crc kubenswrapper[4766]: I1126 00:46:09.840926 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ee29573-3f38-49d9-8c91-6c7ae3d2bd35" path="/var/lib/kubelet/pods/7ee29573-3f38-49d9-8c91-6c7ae3d2bd35/volumes" Nov 26 00:46:09 crc kubenswrapper[4766]: I1126 00:46:09.967700 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-combined-ca-bundle\") pod \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\" (UID: \"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d\") " Nov 26 00:46:09 crc kubenswrapper[4766]: I1126 00:46:09.972272 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d" (UID: "d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:10 crc kubenswrapper[4766]: I1126 00:46:10.071361 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:10 crc kubenswrapper[4766]: I1126 00:46:10.098882 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67fdfd759d-sqzs7" event={"ID":"d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d","Type":"ContainerDied","Data":"c1a081c01d23e6f098de1d4221082f278f3ca37fce64196a3afe897391cee961"} Nov 26 00:46:10 crc kubenswrapper[4766]: I1126 00:46:10.099121 4766 scope.go:117] "RemoveContainer" containerID="abaec1e74c1d959f142b7ae8c9c578a7be50a03c4e6f9a706debca9c5b7619b4" Nov 26 00:46:10 crc kubenswrapper[4766]: I1126 00:46:10.099221 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-67fdfd759d-sqzs7" Nov 26 00:46:10 crc kubenswrapper[4766]: I1126 00:46:10.165464 4766 scope.go:117] "RemoveContainer" containerID="0c2c7807bed04eacaab51baab389f6ad6e8d873b0bbbe0bac9f40f9297331099" Nov 26 00:46:10 crc kubenswrapper[4766]: I1126 00:46:10.167842 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-67fdfd759d-sqzs7"] Nov 26 00:46:10 crc kubenswrapper[4766]: I1126 00:46:10.186257 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-67fdfd759d-sqzs7"] Nov 26 00:46:10 crc kubenswrapper[4766]: I1126 00:46:10.757470 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6768b6c6f7-lm9pw" Nov 26 00:46:11 crc kubenswrapper[4766]: I1126 00:46:11.480005 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:46:11 crc kubenswrapper[4766]: I1126 00:46:11.480598 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:46:11 crc kubenswrapper[4766]: I1126 00:46:11.480716 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:46:11 crc kubenswrapper[4766]: I1126 00:46:11.481988 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8a2702c35239a61bcf6c33a8721b25d955ee16f1fd71838bddfceaf1a1a62d99"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 00:46:11 crc kubenswrapper[4766]: I1126 00:46:11.482059 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://8a2702c35239a61bcf6c33a8721b25d955ee16f1fd71838bddfceaf1a1a62d99" gracePeriod=600 Nov 26 00:46:11 crc kubenswrapper[4766]: I1126 00:46:11.839407 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d" path="/var/lib/kubelet/pods/d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d/volumes" Nov 26 00:46:12 crc kubenswrapper[4766]: I1126 00:46:12.121964 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="8a2702c35239a61bcf6c33a8721b25d955ee16f1fd71838bddfceaf1a1a62d99" exitCode=0 Nov 26 00:46:12 crc kubenswrapper[4766]: I1126 00:46:12.122008 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"8a2702c35239a61bcf6c33a8721b25d955ee16f1fd71838bddfceaf1a1a62d99"} Nov 26 00:46:12 crc kubenswrapper[4766]: I1126 00:46:12.122042 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc"} Nov 26 00:46:12 crc kubenswrapper[4766]: I1126 00:46:12.122059 4766 scope.go:117] "RemoveContainer" containerID="cef0ca8e4338b5dbd39bc68daed18f6f5044b46711899f9b65da90146983e184" Nov 26 00:46:12 crc kubenswrapper[4766]: I1126 00:46:12.709864 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:12 crc kubenswrapper[4766]: I1126 00:46:12.776601 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-6mpmd"] Nov 26 00:46:12 crc kubenswrapper[4766]: I1126 00:46:12.776852 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" podUID="de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf" containerName="dnsmasq-dns" containerID="cri-o://05a7e8985febc475efe59ece6286b70097ffaa8b37feb30ded6e1f4efbd2a67c" gracePeriod=10 Nov 26 00:46:12 crc kubenswrapper[4766]: I1126 00:46:12.948002 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.004952 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.137859 4766 generic.go:334] "Generic (PLEG): container finished" podID="de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf" containerID="05a7e8985febc475efe59ece6286b70097ffaa8b37feb30ded6e1f4efbd2a67c" exitCode=0 Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.138249 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" event={"ID":"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf","Type":"ContainerDied","Data":"05a7e8985febc475efe59ece6286b70097ffaa8b37feb30ded6e1f4efbd2a67c"} Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.152457 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="645620df-fc9e-44ec-b035-a2b0b6a3b7d0" containerName="cinder-scheduler" containerID="cri-o://f41589082a4ba0a4af8f39f96c8a863e223f64645064fa84e742af76a2734a0d" gracePeriod=30 Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.152880 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="645620df-fc9e-44ec-b035-a2b0b6a3b7d0" containerName="probe" containerID="cri-o://cccee7808c8f489ebb681b487a50eb11e480e8c26e414b30f31dbf4dfce60b22" gracePeriod=30 Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.360073 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.545060 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-dns-svc\") pod \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.545383 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-config\") pod \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.545411 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-dns-swift-storage-0\") pod \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.545476 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-ovsdbserver-sb\") pod \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.545551 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-ovsdbserver-nb\") pod \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.545598 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8z77t\" (UniqueName: \"kubernetes.io/projected/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-kube-api-access-8z77t\") pod \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\" (UID: \"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf\") " Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.550837 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-kube-api-access-8z77t" (OuterVolumeSpecName: "kube-api-access-8z77t") pod "de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf" (UID: "de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf"). InnerVolumeSpecName "kube-api-access-8z77t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.629121 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf" (UID: "de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.637882 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-config" (OuterVolumeSpecName: "config") pod "de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf" (UID: "de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.655480 4766 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.655523 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.655538 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8z77t\" (UniqueName: \"kubernetes.io/projected/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-kube-api-access-8z77t\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.656803 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf" (UID: "de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.708361 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf" (UID: "de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.740275 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf" (UID: "de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.757215 4766 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.757259 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:13 crc kubenswrapper[4766]: I1126 00:46:13.757271 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.165951 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" event={"ID":"de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf","Type":"ContainerDied","Data":"b7d4b61a7aecfcfc928de430bc52b2afd98b9427c1f733399cf26530e913f479"} Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.166287 4766 scope.go:117] "RemoveContainer" containerID="05a7e8985febc475efe59ece6286b70097ffaa8b37feb30ded6e1f4efbd2a67c" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.166445 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.222103 4766 scope.go:117] "RemoveContainer" containerID="83579b0bcf4b669c84c274a4beac436450174217ba2feda87383e0ff170b31d5" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.228554 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-6mpmd"] Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.237600 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-6mpmd"] Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.507988 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 26 00:46:14 crc kubenswrapper[4766]: E1126 00:46:14.508362 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d" containerName="barbican-api-log" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.508378 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d" containerName="barbican-api-log" Nov 26 00:46:14 crc kubenswrapper[4766]: E1126 00:46:14.508393 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee29573-3f38-49d9-8c91-6c7ae3d2bd35" containerName="neutron-api" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.508400 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee29573-3f38-49d9-8c91-6c7ae3d2bd35" containerName="neutron-api" Nov 26 00:46:14 crc kubenswrapper[4766]: E1126 00:46:14.508439 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d" containerName="barbican-api" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.508445 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d" containerName="barbican-api" Nov 26 00:46:14 crc kubenswrapper[4766]: E1126 00:46:14.508453 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf" containerName="dnsmasq-dns" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.508459 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf" containerName="dnsmasq-dns" Nov 26 00:46:14 crc kubenswrapper[4766]: E1126 00:46:14.508468 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf" containerName="init" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.508474 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf" containerName="init" Nov 26 00:46:14 crc kubenswrapper[4766]: E1126 00:46:14.508487 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee29573-3f38-49d9-8c91-6c7ae3d2bd35" containerName="neutron-httpd" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.508493 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee29573-3f38-49d9-8c91-6c7ae3d2bd35" containerName="neutron-httpd" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.509461 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ee29573-3f38-49d9-8c91-6c7ae3d2bd35" containerName="neutron-httpd" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.509490 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ee29573-3f38-49d9-8c91-6c7ae3d2bd35" containerName="neutron-api" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.509502 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf" containerName="dnsmasq-dns" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.509516 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d" containerName="barbican-api" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.509537 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9b3be2b-22dc-4660-9d9a-dde2bbf80a5d" containerName="barbican-api-log" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.510355 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.513361 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.513520 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-cwfw4" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.514333 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.523458 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.690424 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69c78e84-84eb-4f2f-a58b-1f6654142b26-combined-ca-bundle\") pod \"openstackclient\" (UID: \"69c78e84-84eb-4f2f-a58b-1f6654142b26\") " pod="openstack/openstackclient" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.690505 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sr6k\" (UniqueName: \"kubernetes.io/projected/69c78e84-84eb-4f2f-a58b-1f6654142b26-kube-api-access-2sr6k\") pod \"openstackclient\" (UID: \"69c78e84-84eb-4f2f-a58b-1f6654142b26\") " pod="openstack/openstackclient" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.690671 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/69c78e84-84eb-4f2f-a58b-1f6654142b26-openstack-config-secret\") pod \"openstackclient\" (UID: \"69c78e84-84eb-4f2f-a58b-1f6654142b26\") " pod="openstack/openstackclient" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.690737 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/69c78e84-84eb-4f2f-a58b-1f6654142b26-openstack-config\") pod \"openstackclient\" (UID: \"69c78e84-84eb-4f2f-a58b-1f6654142b26\") " pod="openstack/openstackclient" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.741620 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 26 00:46:14 crc kubenswrapper[4766]: E1126 00:46:14.744313 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-2sr6k openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="69c78e84-84eb-4f2f-a58b-1f6654142b26" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.757033 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.782680 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.783876 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.792450 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69c78e84-84eb-4f2f-a58b-1f6654142b26-combined-ca-bundle\") pod \"openstackclient\" (UID: \"69c78e84-84eb-4f2f-a58b-1f6654142b26\") " pod="openstack/openstackclient" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.792763 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sr6k\" (UniqueName: \"kubernetes.io/projected/69c78e84-84eb-4f2f-a58b-1f6654142b26-kube-api-access-2sr6k\") pod \"openstackclient\" (UID: \"69c78e84-84eb-4f2f-a58b-1f6654142b26\") " pod="openstack/openstackclient" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.792929 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/69c78e84-84eb-4f2f-a58b-1f6654142b26-openstack-config-secret\") pod \"openstackclient\" (UID: \"69c78e84-84eb-4f2f-a58b-1f6654142b26\") " pod="openstack/openstackclient" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.793038 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/69c78e84-84eb-4f2f-a58b-1f6654142b26-openstack-config\") pod \"openstackclient\" (UID: \"69c78e84-84eb-4f2f-a58b-1f6654142b26\") " pod="openstack/openstackclient" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.794102 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/69c78e84-84eb-4f2f-a58b-1f6654142b26-openstack-config\") pod \"openstackclient\" (UID: \"69c78e84-84eb-4f2f-a58b-1f6654142b26\") " pod="openstack/openstackclient" Nov 26 00:46:14 crc kubenswrapper[4766]: E1126 00:46:14.796220 4766 projected.go:194] Error preparing data for projected volume kube-api-access-2sr6k for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (69c78e84-84eb-4f2f-a58b-1f6654142b26) does not match the UID in record. The object might have been deleted and then recreated Nov 26 00:46:14 crc kubenswrapper[4766]: E1126 00:46:14.796302 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/69c78e84-84eb-4f2f-a58b-1f6654142b26-kube-api-access-2sr6k podName:69c78e84-84eb-4f2f-a58b-1f6654142b26 nodeName:}" failed. No retries permitted until 2025-11-26 00:46:15.296279179 +0000 UTC m=+1356.145049699 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-2sr6k" (UniqueName: "kubernetes.io/projected/69c78e84-84eb-4f2f-a58b-1f6654142b26-kube-api-access-2sr6k") pod "openstackclient" (UID: "69c78e84-84eb-4f2f-a58b-1f6654142b26") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (69c78e84-84eb-4f2f-a58b-1f6654142b26) does not match the UID in record. The object might have been deleted and then recreated Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.799981 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/69c78e84-84eb-4f2f-a58b-1f6654142b26-openstack-config-secret\") pod \"openstackclient\" (UID: \"69c78e84-84eb-4f2f-a58b-1f6654142b26\") " pod="openstack/openstackclient" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.802242 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.825324 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69c78e84-84eb-4f2f-a58b-1f6654142b26-combined-ca-bundle\") pod \"openstackclient\" (UID: \"69c78e84-84eb-4f2f-a58b-1f6654142b26\") " pod="openstack/openstackclient" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.895166 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8551b13-7602-4c7c-bfb5-0bcf7e64954d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"d8551b13-7602-4c7c-bfb5-0bcf7e64954d\") " pod="openstack/openstackclient" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.895237 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d8551b13-7602-4c7c-bfb5-0bcf7e64954d-openstack-config\") pod \"openstackclient\" (UID: \"d8551b13-7602-4c7c-bfb5-0bcf7e64954d\") " pod="openstack/openstackclient" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.895325 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d8551b13-7602-4c7c-bfb5-0bcf7e64954d-openstack-config-secret\") pod \"openstackclient\" (UID: \"d8551b13-7602-4c7c-bfb5-0bcf7e64954d\") " pod="openstack/openstackclient" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.895375 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6gt8\" (UniqueName: \"kubernetes.io/projected/d8551b13-7602-4c7c-bfb5-0bcf7e64954d-kube-api-access-j6gt8\") pod \"openstackclient\" (UID: \"d8551b13-7602-4c7c-bfb5-0bcf7e64954d\") " pod="openstack/openstackclient" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.997438 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8551b13-7602-4c7c-bfb5-0bcf7e64954d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"d8551b13-7602-4c7c-bfb5-0bcf7e64954d\") " pod="openstack/openstackclient" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.997504 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d8551b13-7602-4c7c-bfb5-0bcf7e64954d-openstack-config\") pod \"openstackclient\" (UID: \"d8551b13-7602-4c7c-bfb5-0bcf7e64954d\") " pod="openstack/openstackclient" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.997562 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d8551b13-7602-4c7c-bfb5-0bcf7e64954d-openstack-config-secret\") pod \"openstackclient\" (UID: \"d8551b13-7602-4c7c-bfb5-0bcf7e64954d\") " pod="openstack/openstackclient" Nov 26 00:46:14 crc kubenswrapper[4766]: I1126 00:46:14.997595 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6gt8\" (UniqueName: \"kubernetes.io/projected/d8551b13-7602-4c7c-bfb5-0bcf7e64954d-kube-api-access-j6gt8\") pod \"openstackclient\" (UID: \"d8551b13-7602-4c7c-bfb5-0bcf7e64954d\") " pod="openstack/openstackclient" Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:14.998458 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d8551b13-7602-4c7c-bfb5-0bcf7e64954d-openstack-config\") pod \"openstackclient\" (UID: \"d8551b13-7602-4c7c-bfb5-0bcf7e64954d\") " pod="openstack/openstackclient" Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.001017 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d8551b13-7602-4c7c-bfb5-0bcf7e64954d-openstack-config-secret\") pod \"openstackclient\" (UID: \"d8551b13-7602-4c7c-bfb5-0bcf7e64954d\") " pod="openstack/openstackclient" Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.001238 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8551b13-7602-4c7c-bfb5-0bcf7e64954d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"d8551b13-7602-4c7c-bfb5-0bcf7e64954d\") " pod="openstack/openstackclient" Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.016919 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6gt8\" (UniqueName: \"kubernetes.io/projected/d8551b13-7602-4c7c-bfb5-0bcf7e64954d-kube-api-access-j6gt8\") pod \"openstackclient\" (UID: \"d8551b13-7602-4c7c-bfb5-0bcf7e64954d\") " pod="openstack/openstackclient" Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.179283 4766 generic.go:334] "Generic (PLEG): container finished" podID="645620df-fc9e-44ec-b035-a2b0b6a3b7d0" containerID="cccee7808c8f489ebb681b487a50eb11e480e8c26e414b30f31dbf4dfce60b22" exitCode=0 Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.179374 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.179359 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"645620df-fc9e-44ec-b035-a2b0b6a3b7d0","Type":"ContainerDied","Data":"cccee7808c8f489ebb681b487a50eb11e480e8c26e414b30f31dbf4dfce60b22"} Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.182377 4766 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="69c78e84-84eb-4f2f-a58b-1f6654142b26" podUID="d8551b13-7602-4c7c-bfb5-0bcf7e64954d" Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.191058 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.208208 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.301874 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/69c78e84-84eb-4f2f-a58b-1f6654142b26-openstack-config-secret\") pod \"69c78e84-84eb-4f2f-a58b-1f6654142b26\" (UID: \"69c78e84-84eb-4f2f-a58b-1f6654142b26\") " Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.302244 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/69c78e84-84eb-4f2f-a58b-1f6654142b26-openstack-config\") pod \"69c78e84-84eb-4f2f-a58b-1f6654142b26\" (UID: \"69c78e84-84eb-4f2f-a58b-1f6654142b26\") " Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.302281 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69c78e84-84eb-4f2f-a58b-1f6654142b26-combined-ca-bundle\") pod \"69c78e84-84eb-4f2f-a58b-1f6654142b26\" (UID: \"69c78e84-84eb-4f2f-a58b-1f6654142b26\") " Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.302731 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2sr6k\" (UniqueName: \"kubernetes.io/projected/69c78e84-84eb-4f2f-a58b-1f6654142b26-kube-api-access-2sr6k\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.302777 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69c78e84-84eb-4f2f-a58b-1f6654142b26-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "69c78e84-84eb-4f2f-a58b-1f6654142b26" (UID: "69c78e84-84eb-4f2f-a58b-1f6654142b26"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.305924 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69c78e84-84eb-4f2f-a58b-1f6654142b26-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "69c78e84-84eb-4f2f-a58b-1f6654142b26" (UID: "69c78e84-84eb-4f2f-a58b-1f6654142b26"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.306634 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69c78e84-84eb-4f2f-a58b-1f6654142b26-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69c78e84-84eb-4f2f-a58b-1f6654142b26" (UID: "69c78e84-84eb-4f2f-a58b-1f6654142b26"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.407088 4766 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/69c78e84-84eb-4f2f-a58b-1f6654142b26-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.407119 4766 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/69c78e84-84eb-4f2f-a58b-1f6654142b26-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.407132 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69c78e84-84eb-4f2f-a58b-1f6654142b26-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.597082 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.716986 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.839765 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69c78e84-84eb-4f2f-a58b-1f6654142b26" path="/var/lib/kubelet/pods/69c78e84-84eb-4f2f-a58b-1f6654142b26/volumes" Nov 26 00:46:15 crc kubenswrapper[4766]: I1126 00:46:15.841585 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf" path="/var/lib/kubelet/pods/de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf/volumes" Nov 26 00:46:16 crc kubenswrapper[4766]: I1126 00:46:16.189491 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 26 00:46:16 crc kubenswrapper[4766]: I1126 00:46:16.189566 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"d8551b13-7602-4c7c-bfb5-0bcf7e64954d","Type":"ContainerStarted","Data":"3ddf3136a83d29c94c85d420268c99701b58989a1f834d81d3f17b129ce5aa50"} Nov 26 00:46:16 crc kubenswrapper[4766]: I1126 00:46:16.197393 4766 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="69c78e84-84eb-4f2f-a58b-1f6654142b26" podUID="d8551b13-7602-4c7c-bfb5-0bcf7e64954d" Nov 26 00:46:17 crc kubenswrapper[4766]: I1126 00:46:17.871763 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.055391 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-combined-ca-bundle\") pod \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.055470 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-scripts\") pod \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.055571 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-config-data-custom\") pod \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.055720 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-etc-machine-id\") pod \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.055809 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcn86\" (UniqueName: \"kubernetes.io/projected/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-kube-api-access-gcn86\") pod \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.055832 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "645620df-fc9e-44ec-b035-a2b0b6a3b7d0" (UID: "645620df-fc9e-44ec-b035-a2b0b6a3b7d0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.055919 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-config-data\") pod \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\" (UID: \"645620df-fc9e-44ec-b035-a2b0b6a3b7d0\") " Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.056625 4766 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.066088 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-kube-api-access-gcn86" (OuterVolumeSpecName: "kube-api-access-gcn86") pod "645620df-fc9e-44ec-b035-a2b0b6a3b7d0" (UID: "645620df-fc9e-44ec-b035-a2b0b6a3b7d0"). InnerVolumeSpecName "kube-api-access-gcn86". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.076036 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "645620df-fc9e-44ec-b035-a2b0b6a3b7d0" (UID: "645620df-fc9e-44ec-b035-a2b0b6a3b7d0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.076175 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-scripts" (OuterVolumeSpecName: "scripts") pod "645620df-fc9e-44ec-b035-a2b0b6a3b7d0" (UID: "645620df-fc9e-44ec-b035-a2b0b6a3b7d0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.107895 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-848cf88cfc-6mpmd" podUID="de7bc7aa-0bf0-47d8-a95d-a12dd49ae3bf" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.182:5353: i/o timeout" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.140977 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "645620df-fc9e-44ec-b035-a2b0b6a3b7d0" (UID: "645620df-fc9e-44ec-b035-a2b0b6a3b7d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.160209 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.160411 4766 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.160517 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcn86\" (UniqueName: \"kubernetes.io/projected/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-kube-api-access-gcn86\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.160585 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.226593 4766 generic.go:334] "Generic (PLEG): container finished" podID="645620df-fc9e-44ec-b035-a2b0b6a3b7d0" containerID="f41589082a4ba0a4af8f39f96c8a863e223f64645064fa84e742af76a2734a0d" exitCode=0 Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.226636 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"645620df-fc9e-44ec-b035-a2b0b6a3b7d0","Type":"ContainerDied","Data":"f41589082a4ba0a4af8f39f96c8a863e223f64645064fa84e742af76a2734a0d"} Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.226691 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"645620df-fc9e-44ec-b035-a2b0b6a3b7d0","Type":"ContainerDied","Data":"5d11aec0f8bc21aebde7b77384683b9547928ccebcf394522fd2f3d83f0eb8bc"} Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.226714 4766 scope.go:117] "RemoveContainer" containerID="cccee7808c8f489ebb681b487a50eb11e480e8c26e414b30f31dbf4dfce60b22" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.227337 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.237573 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-config-data" (OuterVolumeSpecName: "config-data") pod "645620df-fc9e-44ec-b035-a2b0b6a3b7d0" (UID: "645620df-fc9e-44ec-b035-a2b0b6a3b7d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.256766 4766 scope.go:117] "RemoveContainer" containerID="f41589082a4ba0a4af8f39f96c8a863e223f64645064fa84e742af76a2734a0d" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.262706 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/645620df-fc9e-44ec-b035-a2b0b6a3b7d0-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.277848 4766 scope.go:117] "RemoveContainer" containerID="cccee7808c8f489ebb681b487a50eb11e480e8c26e414b30f31dbf4dfce60b22" Nov 26 00:46:18 crc kubenswrapper[4766]: E1126 00:46:18.278539 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cccee7808c8f489ebb681b487a50eb11e480e8c26e414b30f31dbf4dfce60b22\": container with ID starting with cccee7808c8f489ebb681b487a50eb11e480e8c26e414b30f31dbf4dfce60b22 not found: ID does not exist" containerID="cccee7808c8f489ebb681b487a50eb11e480e8c26e414b30f31dbf4dfce60b22" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.278578 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cccee7808c8f489ebb681b487a50eb11e480e8c26e414b30f31dbf4dfce60b22"} err="failed to get container status \"cccee7808c8f489ebb681b487a50eb11e480e8c26e414b30f31dbf4dfce60b22\": rpc error: code = NotFound desc = could not find container \"cccee7808c8f489ebb681b487a50eb11e480e8c26e414b30f31dbf4dfce60b22\": container with ID starting with cccee7808c8f489ebb681b487a50eb11e480e8c26e414b30f31dbf4dfce60b22 not found: ID does not exist" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.278602 4766 scope.go:117] "RemoveContainer" containerID="f41589082a4ba0a4af8f39f96c8a863e223f64645064fa84e742af76a2734a0d" Nov 26 00:46:18 crc kubenswrapper[4766]: E1126 00:46:18.279117 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f41589082a4ba0a4af8f39f96c8a863e223f64645064fa84e742af76a2734a0d\": container with ID starting with f41589082a4ba0a4af8f39f96c8a863e223f64645064fa84e742af76a2734a0d not found: ID does not exist" containerID="f41589082a4ba0a4af8f39f96c8a863e223f64645064fa84e742af76a2734a0d" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.279147 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f41589082a4ba0a4af8f39f96c8a863e223f64645064fa84e742af76a2734a0d"} err="failed to get container status \"f41589082a4ba0a4af8f39f96c8a863e223f64645064fa84e742af76a2734a0d\": rpc error: code = NotFound desc = could not find container \"f41589082a4ba0a4af8f39f96c8a863e223f64645064fa84e742af76a2734a0d\": container with ID starting with f41589082a4ba0a4af8f39f96c8a863e223f64645064fa84e742af76a2734a0d not found: ID does not exist" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.569101 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.579326 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.593405 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 26 00:46:18 crc kubenswrapper[4766]: E1126 00:46:18.595633 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="645620df-fc9e-44ec-b035-a2b0b6a3b7d0" containerName="probe" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.595671 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="645620df-fc9e-44ec-b035-a2b0b6a3b7d0" containerName="probe" Nov 26 00:46:18 crc kubenswrapper[4766]: E1126 00:46:18.595685 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="645620df-fc9e-44ec-b035-a2b0b6a3b7d0" containerName="cinder-scheduler" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.595690 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="645620df-fc9e-44ec-b035-a2b0b6a3b7d0" containerName="cinder-scheduler" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.595959 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="645620df-fc9e-44ec-b035-a2b0b6a3b7d0" containerName="cinder-scheduler" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.595984 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="645620df-fc9e-44ec-b035-a2b0b6a3b7d0" containerName="probe" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.597955 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.602299 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.606806 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.782874 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92d51b16-e82c-4fec-a095-53e3f0608fad-config-data\") pod \"cinder-scheduler-0\" (UID: \"92d51b16-e82c-4fec-a095-53e3f0608fad\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.782941 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/92d51b16-e82c-4fec-a095-53e3f0608fad-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"92d51b16-e82c-4fec-a095-53e3f0608fad\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.783001 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d51b16-e82c-4fec-a095-53e3f0608fad-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"92d51b16-e82c-4fec-a095-53e3f0608fad\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.783033 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvpbm\" (UniqueName: \"kubernetes.io/projected/92d51b16-e82c-4fec-a095-53e3f0608fad-kube-api-access-wvpbm\") pod \"cinder-scheduler-0\" (UID: \"92d51b16-e82c-4fec-a095-53e3f0608fad\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.783057 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92d51b16-e82c-4fec-a095-53e3f0608fad-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"92d51b16-e82c-4fec-a095-53e3f0608fad\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.783162 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92d51b16-e82c-4fec-a095-53e3f0608fad-scripts\") pod \"cinder-scheduler-0\" (UID: \"92d51b16-e82c-4fec-a095-53e3f0608fad\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.887651 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvpbm\" (UniqueName: \"kubernetes.io/projected/92d51b16-e82c-4fec-a095-53e3f0608fad-kube-api-access-wvpbm\") pod \"cinder-scheduler-0\" (UID: \"92d51b16-e82c-4fec-a095-53e3f0608fad\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.887715 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92d51b16-e82c-4fec-a095-53e3f0608fad-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"92d51b16-e82c-4fec-a095-53e3f0608fad\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.887763 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92d51b16-e82c-4fec-a095-53e3f0608fad-scripts\") pod \"cinder-scheduler-0\" (UID: \"92d51b16-e82c-4fec-a095-53e3f0608fad\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.887831 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92d51b16-e82c-4fec-a095-53e3f0608fad-config-data\") pod \"cinder-scheduler-0\" (UID: \"92d51b16-e82c-4fec-a095-53e3f0608fad\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.887870 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/92d51b16-e82c-4fec-a095-53e3f0608fad-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"92d51b16-e82c-4fec-a095-53e3f0608fad\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.887927 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d51b16-e82c-4fec-a095-53e3f0608fad-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"92d51b16-e82c-4fec-a095-53e3f0608fad\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.888114 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/92d51b16-e82c-4fec-a095-53e3f0608fad-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"92d51b16-e82c-4fec-a095-53e3f0608fad\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.892402 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d51b16-e82c-4fec-a095-53e3f0608fad-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"92d51b16-e82c-4fec-a095-53e3f0608fad\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.893695 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92d51b16-e82c-4fec-a095-53e3f0608fad-scripts\") pod \"cinder-scheduler-0\" (UID: \"92d51b16-e82c-4fec-a095-53e3f0608fad\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.895310 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92d51b16-e82c-4fec-a095-53e3f0608fad-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"92d51b16-e82c-4fec-a095-53e3f0608fad\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.905854 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92d51b16-e82c-4fec-a095-53e3f0608fad-config-data\") pod \"cinder-scheduler-0\" (UID: \"92d51b16-e82c-4fec-a095-53e3f0608fad\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.913354 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvpbm\" (UniqueName: \"kubernetes.io/projected/92d51b16-e82c-4fec-a095-53e3f0608fad-kube-api-access-wvpbm\") pod \"cinder-scheduler-0\" (UID: \"92d51b16-e82c-4fec-a095-53e3f0608fad\") " pod="openstack/cinder-scheduler-0" Nov 26 00:46:18 crc kubenswrapper[4766]: I1126 00:46:18.929245 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 26 00:46:19 crc kubenswrapper[4766]: I1126 00:46:19.540129 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 26 00:46:19 crc kubenswrapper[4766]: I1126 00:46:19.868292 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="645620df-fc9e-44ec-b035-a2b0b6a3b7d0" path="/var/lib/kubelet/pods/645620df-fc9e-44ec-b035-a2b0b6a3b7d0/volumes" Nov 26 00:46:20 crc kubenswrapper[4766]: I1126 00:46:20.267211 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"92d51b16-e82c-4fec-a095-53e3f0608fad","Type":"ContainerStarted","Data":"6319480dd8e77faa5a7c8def5cbbf2777da8240ba7ab72f06f98f6197a2a41df"} Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.283177 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"92d51b16-e82c-4fec-a095-53e3f0608fad","Type":"ContainerStarted","Data":"e7380e8c6b3864445be90ecfb79eaca0808daa5cb6b7d4ea661a526995160321"} Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.283765 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"92d51b16-e82c-4fec-a095-53e3f0608fad","Type":"ContainerStarted","Data":"285539157ed24960c6b1aeeb1479ed2a9c469673766559b1ed847fd7802e0068"} Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.305251 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.305164794 podStartE2EDuration="3.305164794s" podCreationTimestamp="2025-11-26 00:46:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:46:21.300829125 +0000 UTC m=+1362.149599555" watchObservedRunningTime="2025-11-26 00:46:21.305164794 +0000 UTC m=+1362.153935224" Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.743582 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-6dfcdbc5c7-znfmt"] Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.751676 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6dfcdbc5c7-znfmt" Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.757227 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.757378 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.759547 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-8fcdz" Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.765108 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-config-data-custom\") pod \"heat-engine-6dfcdbc5c7-znfmt\" (UID: \"13e19aaf-7cc3-40a6-852b-72b77b37cf1e\") " pod="openstack/heat-engine-6dfcdbc5c7-znfmt" Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.765173 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-config-data\") pod \"heat-engine-6dfcdbc5c7-znfmt\" (UID: \"13e19aaf-7cc3-40a6-852b-72b77b37cf1e\") " pod="openstack/heat-engine-6dfcdbc5c7-znfmt" Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.765247 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46chp\" (UniqueName: \"kubernetes.io/projected/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-kube-api-access-46chp\") pod \"heat-engine-6dfcdbc5c7-znfmt\" (UID: \"13e19aaf-7cc3-40a6-852b-72b77b37cf1e\") " pod="openstack/heat-engine-6dfcdbc5c7-znfmt" Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.765367 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-combined-ca-bundle\") pod \"heat-engine-6dfcdbc5c7-znfmt\" (UID: \"13e19aaf-7cc3-40a6-852b-72b77b37cf1e\") " pod="openstack/heat-engine-6dfcdbc5c7-znfmt" Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.774771 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6dfcdbc5c7-znfmt"] Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.867220 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46chp\" (UniqueName: \"kubernetes.io/projected/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-kube-api-access-46chp\") pod \"heat-engine-6dfcdbc5c7-znfmt\" (UID: \"13e19aaf-7cc3-40a6-852b-72b77b37cf1e\") " pod="openstack/heat-engine-6dfcdbc5c7-znfmt" Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.867417 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-combined-ca-bundle\") pod \"heat-engine-6dfcdbc5c7-znfmt\" (UID: \"13e19aaf-7cc3-40a6-852b-72b77b37cf1e\") " pod="openstack/heat-engine-6dfcdbc5c7-znfmt" Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.867528 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-config-data-custom\") pod \"heat-engine-6dfcdbc5c7-znfmt\" (UID: \"13e19aaf-7cc3-40a6-852b-72b77b37cf1e\") " pod="openstack/heat-engine-6dfcdbc5c7-znfmt" Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.867605 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-config-data\") pod \"heat-engine-6dfcdbc5c7-znfmt\" (UID: \"13e19aaf-7cc3-40a6-852b-72b77b37cf1e\") " pod="openstack/heat-engine-6dfcdbc5c7-znfmt" Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.883469 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-combined-ca-bundle\") pod \"heat-engine-6dfcdbc5c7-znfmt\" (UID: \"13e19aaf-7cc3-40a6-852b-72b77b37cf1e\") " pod="openstack/heat-engine-6dfcdbc5c7-znfmt" Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.883625 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-config-data-custom\") pod \"heat-engine-6dfcdbc5c7-znfmt\" (UID: \"13e19aaf-7cc3-40a6-852b-72b77b37cf1e\") " pod="openstack/heat-engine-6dfcdbc5c7-znfmt" Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.886407 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-config-data\") pod \"heat-engine-6dfcdbc5c7-znfmt\" (UID: \"13e19aaf-7cc3-40a6-852b-72b77b37cf1e\") " pod="openstack/heat-engine-6dfcdbc5c7-znfmt" Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.886478 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-b8c9d9569-lrlx6"] Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.899854 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.905571 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46chp\" (UniqueName: \"kubernetes.io/projected/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-kube-api-access-46chp\") pod \"heat-engine-6dfcdbc5c7-znfmt\" (UID: \"13e19aaf-7cc3-40a6-852b-72b77b37cf1e\") " pod="openstack/heat-engine-6dfcdbc5c7-znfmt" Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.906812 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.922149 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-b8c9d9569-lrlx6"] Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.958374 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-xtpgd"] Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.960110 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:21 crc kubenswrapper[4766]: I1126 00:46:21.984462 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-xtpgd"] Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.073300 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-856c5c6f86-fk66f"] Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.075576 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-config\") pod \"dnsmasq-dns-688b9f5b49-xtpgd\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.075636 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-ovsdbserver-sb\") pod \"dnsmasq-dns-688b9f5b49-xtpgd\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.075741 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-dns-swift-storage-0\") pod \"dnsmasq-dns-688b9f5b49-xtpgd\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.075766 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4442abe-9e23-4f4e-880c-89ca9d8ef785-config-data-custom\") pod \"heat-cfnapi-b8c9d9569-lrlx6\" (UID: \"d4442abe-9e23-4f4e-880c-89ca9d8ef785\") " pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.075789 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4442abe-9e23-4f4e-880c-89ca9d8ef785-combined-ca-bundle\") pod \"heat-cfnapi-b8c9d9569-lrlx6\" (UID: \"d4442abe-9e23-4f4e-880c-89ca9d8ef785\") " pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.075876 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-ovsdbserver-nb\") pod \"dnsmasq-dns-688b9f5b49-xtpgd\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.075937 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-dns-svc\") pod \"dnsmasq-dns-688b9f5b49-xtpgd\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.076101 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5m9s\" (UniqueName: \"kubernetes.io/projected/d4442abe-9e23-4f4e-880c-89ca9d8ef785-kube-api-access-v5m9s\") pod \"heat-cfnapi-b8c9d9569-lrlx6\" (UID: \"d4442abe-9e23-4f4e-880c-89ca9d8ef785\") " pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.076140 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnd8l\" (UniqueName: \"kubernetes.io/projected/22305d36-865a-4fe0-8337-92009c1c3c29-kube-api-access-wnd8l\") pod \"dnsmasq-dns-688b9f5b49-xtpgd\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.076229 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4442abe-9e23-4f4e-880c-89ca9d8ef785-config-data\") pod \"heat-cfnapi-b8c9d9569-lrlx6\" (UID: \"d4442abe-9e23-4f4e-880c-89ca9d8ef785\") " pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.082644 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-856c5c6f86-fk66f" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.083299 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6dfcdbc5c7-znfmt" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.086590 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.106734 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-856c5c6f86-fk66f"] Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.179352 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnd8l\" (UniqueName: \"kubernetes.io/projected/22305d36-865a-4fe0-8337-92009c1c3c29-kube-api-access-wnd8l\") pod \"dnsmasq-dns-688b9f5b49-xtpgd\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.179418 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4442abe-9e23-4f4e-880c-89ca9d8ef785-config-data\") pod \"heat-cfnapi-b8c9d9569-lrlx6\" (UID: \"d4442abe-9e23-4f4e-880c-89ca9d8ef785\") " pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.179449 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16337a9e-8651-492e-9f39-21515275e3cb-combined-ca-bundle\") pod \"heat-api-856c5c6f86-fk66f\" (UID: \"16337a9e-8651-492e-9f39-21515275e3cb\") " pod="openstack/heat-api-856c5c6f86-fk66f" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.179560 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-config\") pod \"dnsmasq-dns-688b9f5b49-xtpgd\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.179598 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-ovsdbserver-sb\") pod \"dnsmasq-dns-688b9f5b49-xtpgd\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.179880 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16337a9e-8651-492e-9f39-21515275e3cb-config-data\") pod \"heat-api-856c5c6f86-fk66f\" (UID: \"16337a9e-8651-492e-9f39-21515275e3cb\") " pod="openstack/heat-api-856c5c6f86-fk66f" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.179973 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16337a9e-8651-492e-9f39-21515275e3cb-config-data-custom\") pod \"heat-api-856c5c6f86-fk66f\" (UID: \"16337a9e-8651-492e-9f39-21515275e3cb\") " pod="openstack/heat-api-856c5c6f86-fk66f" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.180030 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmtqd\" (UniqueName: \"kubernetes.io/projected/16337a9e-8651-492e-9f39-21515275e3cb-kube-api-access-vmtqd\") pod \"heat-api-856c5c6f86-fk66f\" (UID: \"16337a9e-8651-492e-9f39-21515275e3cb\") " pod="openstack/heat-api-856c5c6f86-fk66f" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.180072 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-dns-swift-storage-0\") pod \"dnsmasq-dns-688b9f5b49-xtpgd\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.180100 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4442abe-9e23-4f4e-880c-89ca9d8ef785-config-data-custom\") pod \"heat-cfnapi-b8c9d9569-lrlx6\" (UID: \"d4442abe-9e23-4f4e-880c-89ca9d8ef785\") " pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.180134 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4442abe-9e23-4f4e-880c-89ca9d8ef785-combined-ca-bundle\") pod \"heat-cfnapi-b8c9d9569-lrlx6\" (UID: \"d4442abe-9e23-4f4e-880c-89ca9d8ef785\") " pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.180167 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-ovsdbserver-nb\") pod \"dnsmasq-dns-688b9f5b49-xtpgd\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.180269 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-dns-svc\") pod \"dnsmasq-dns-688b9f5b49-xtpgd\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.180317 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5m9s\" (UniqueName: \"kubernetes.io/projected/d4442abe-9e23-4f4e-880c-89ca9d8ef785-kube-api-access-v5m9s\") pod \"heat-cfnapi-b8c9d9569-lrlx6\" (UID: \"d4442abe-9e23-4f4e-880c-89ca9d8ef785\") " pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.181203 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-ovsdbserver-sb\") pod \"dnsmasq-dns-688b9f5b49-xtpgd\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.181754 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-config\") pod \"dnsmasq-dns-688b9f5b49-xtpgd\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.181791 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-dns-swift-storage-0\") pod \"dnsmasq-dns-688b9f5b49-xtpgd\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.182536 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-ovsdbserver-nb\") pod \"dnsmasq-dns-688b9f5b49-xtpgd\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.183602 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-dns-svc\") pod \"dnsmasq-dns-688b9f5b49-xtpgd\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.192447 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4442abe-9e23-4f4e-880c-89ca9d8ef785-combined-ca-bundle\") pod \"heat-cfnapi-b8c9d9569-lrlx6\" (UID: \"d4442abe-9e23-4f4e-880c-89ca9d8ef785\") " pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.192843 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4442abe-9e23-4f4e-880c-89ca9d8ef785-config-data\") pod \"heat-cfnapi-b8c9d9569-lrlx6\" (UID: \"d4442abe-9e23-4f4e-880c-89ca9d8ef785\") " pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.196815 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4442abe-9e23-4f4e-880c-89ca9d8ef785-config-data-custom\") pod \"heat-cfnapi-b8c9d9569-lrlx6\" (UID: \"d4442abe-9e23-4f4e-880c-89ca9d8ef785\") " pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.200092 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnd8l\" (UniqueName: \"kubernetes.io/projected/22305d36-865a-4fe0-8337-92009c1c3c29-kube-api-access-wnd8l\") pod \"dnsmasq-dns-688b9f5b49-xtpgd\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.210817 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5m9s\" (UniqueName: \"kubernetes.io/projected/d4442abe-9e23-4f4e-880c-89ca9d8ef785-kube-api-access-v5m9s\") pod \"heat-cfnapi-b8c9d9569-lrlx6\" (UID: \"d4442abe-9e23-4f4e-880c-89ca9d8ef785\") " pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.282282 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16337a9e-8651-492e-9f39-21515275e3cb-config-data\") pod \"heat-api-856c5c6f86-fk66f\" (UID: \"16337a9e-8651-492e-9f39-21515275e3cb\") " pod="openstack/heat-api-856c5c6f86-fk66f" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.282343 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16337a9e-8651-492e-9f39-21515275e3cb-config-data-custom\") pod \"heat-api-856c5c6f86-fk66f\" (UID: \"16337a9e-8651-492e-9f39-21515275e3cb\") " pod="openstack/heat-api-856c5c6f86-fk66f" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.282376 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmtqd\" (UniqueName: \"kubernetes.io/projected/16337a9e-8651-492e-9f39-21515275e3cb-kube-api-access-vmtqd\") pod \"heat-api-856c5c6f86-fk66f\" (UID: \"16337a9e-8651-492e-9f39-21515275e3cb\") " pod="openstack/heat-api-856c5c6f86-fk66f" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.282474 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16337a9e-8651-492e-9f39-21515275e3cb-combined-ca-bundle\") pod \"heat-api-856c5c6f86-fk66f\" (UID: \"16337a9e-8651-492e-9f39-21515275e3cb\") " pod="openstack/heat-api-856c5c6f86-fk66f" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.286581 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16337a9e-8651-492e-9f39-21515275e3cb-config-data-custom\") pod \"heat-api-856c5c6f86-fk66f\" (UID: \"16337a9e-8651-492e-9f39-21515275e3cb\") " pod="openstack/heat-api-856c5c6f86-fk66f" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.289492 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16337a9e-8651-492e-9f39-21515275e3cb-config-data\") pod \"heat-api-856c5c6f86-fk66f\" (UID: \"16337a9e-8651-492e-9f39-21515275e3cb\") " pod="openstack/heat-api-856c5c6f86-fk66f" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.299302 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16337a9e-8651-492e-9f39-21515275e3cb-combined-ca-bundle\") pod \"heat-api-856c5c6f86-fk66f\" (UID: \"16337a9e-8651-492e-9f39-21515275e3cb\") " pod="openstack/heat-api-856c5c6f86-fk66f" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.302276 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmtqd\" (UniqueName: \"kubernetes.io/projected/16337a9e-8651-492e-9f39-21515275e3cb-kube-api-access-vmtqd\") pod \"heat-api-856c5c6f86-fk66f\" (UID: \"16337a9e-8651-492e-9f39-21515275e3cb\") " pod="openstack/heat-api-856c5c6f86-fk66f" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.305183 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.317469 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.424324 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-856c5c6f86-fk66f" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.920909 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-74d678b987-q89lp"] Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.922554 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.926819 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.927185 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.927328 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.940724 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-74d678b987-q89lp"] Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.996936 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9n94\" (UniqueName: \"kubernetes.io/projected/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-kube-api-access-j9n94\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.997000 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-etc-swift\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.997021 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-internal-tls-certs\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.997054 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-log-httpd\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.997143 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-public-tls-certs\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.997172 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-run-httpd\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.997389 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-combined-ca-bundle\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:22 crc kubenswrapper[4766]: I1126 00:46:22.997413 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-config-data\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:23 crc kubenswrapper[4766]: I1126 00:46:23.099538 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-combined-ca-bundle\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:23 crc kubenswrapper[4766]: I1126 00:46:23.099581 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-config-data\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:23 crc kubenswrapper[4766]: I1126 00:46:23.099615 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9n94\" (UniqueName: \"kubernetes.io/projected/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-kube-api-access-j9n94\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:23 crc kubenswrapper[4766]: I1126 00:46:23.099642 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-etc-swift\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:23 crc kubenswrapper[4766]: I1126 00:46:23.099716 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-internal-tls-certs\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:23 crc kubenswrapper[4766]: I1126 00:46:23.099751 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-log-httpd\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:23 crc kubenswrapper[4766]: I1126 00:46:23.099801 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-public-tls-certs\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:23 crc kubenswrapper[4766]: I1126 00:46:23.099819 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-run-httpd\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:23 crc kubenswrapper[4766]: I1126 00:46:23.100308 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-run-httpd\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:23 crc kubenswrapper[4766]: I1126 00:46:23.101049 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-log-httpd\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:23 crc kubenswrapper[4766]: I1126 00:46:23.104740 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-etc-swift\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:23 crc kubenswrapper[4766]: I1126 00:46:23.105123 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-internal-tls-certs\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:23 crc kubenswrapper[4766]: I1126 00:46:23.105974 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-public-tls-certs\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:23 crc kubenswrapper[4766]: I1126 00:46:23.106114 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-combined-ca-bundle\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:23 crc kubenswrapper[4766]: I1126 00:46:23.111619 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-config-data\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:23 crc kubenswrapper[4766]: I1126 00:46:23.126208 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9n94\" (UniqueName: \"kubernetes.io/projected/3d4e6997-4bbc-4cb2-a0f2-abf3b940090c-kube-api-access-j9n94\") pod \"swift-proxy-74d678b987-q89lp\" (UID: \"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c\") " pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:23 crc kubenswrapper[4766]: I1126 00:46:23.270739 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:23 crc kubenswrapper[4766]: I1126 00:46:23.930463 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 26 00:46:25 crc kubenswrapper[4766]: I1126 00:46:25.039347 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:46:25 crc kubenswrapper[4766]: I1126 00:46:25.039856 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="85f22975-acba-4af1-8787-124a69cb68e1" containerName="ceilometer-central-agent" containerID="cri-o://8fb365e52e88bc33ca948883a43b83a045fad7d8bb2ec15511f497bc20abefe7" gracePeriod=30 Nov 26 00:46:25 crc kubenswrapper[4766]: I1126 00:46:25.040548 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="85f22975-acba-4af1-8787-124a69cb68e1" containerName="sg-core" containerID="cri-o://48d5602c1bf28f3d5d1828b4c020f553a9f7122a9bc64b06339dbed445c30f1a" gracePeriod=30 Nov 26 00:46:25 crc kubenswrapper[4766]: I1126 00:46:25.040770 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="85f22975-acba-4af1-8787-124a69cb68e1" containerName="proxy-httpd" containerID="cri-o://d24fe490bba51a44c3ebde3231849642624486ad1a4ea2d892d19770d1a77039" gracePeriod=30 Nov 26 00:46:25 crc kubenswrapper[4766]: I1126 00:46:25.040878 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="85f22975-acba-4af1-8787-124a69cb68e1" containerName="ceilometer-notification-agent" containerID="cri-o://2b52ab898b9b6da7a8b73b6a7b2491a7f5ae1f9e232a31b3bad2d13612f5f31d" gracePeriod=30 Nov 26 00:46:25 crc kubenswrapper[4766]: I1126 00:46:25.056002 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 26 00:46:25 crc kubenswrapper[4766]: I1126 00:46:25.385333 4766 generic.go:334] "Generic (PLEG): container finished" podID="85f22975-acba-4af1-8787-124a69cb68e1" containerID="d24fe490bba51a44c3ebde3231849642624486ad1a4ea2d892d19770d1a77039" exitCode=0 Nov 26 00:46:25 crc kubenswrapper[4766]: I1126 00:46:25.385640 4766 generic.go:334] "Generic (PLEG): container finished" podID="85f22975-acba-4af1-8787-124a69cb68e1" containerID="48d5602c1bf28f3d5d1828b4c020f553a9f7122a9bc64b06339dbed445c30f1a" exitCode=2 Nov 26 00:46:25 crc kubenswrapper[4766]: I1126 00:46:25.385607 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85f22975-acba-4af1-8787-124a69cb68e1","Type":"ContainerDied","Data":"d24fe490bba51a44c3ebde3231849642624486ad1a4ea2d892d19770d1a77039"} Nov 26 00:46:25 crc kubenswrapper[4766]: I1126 00:46:25.385693 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85f22975-acba-4af1-8787-124a69cb68e1","Type":"ContainerDied","Data":"48d5602c1bf28f3d5d1828b4c020f553a9f7122a9bc64b06339dbed445c30f1a"} Nov 26 00:46:26 crc kubenswrapper[4766]: I1126 00:46:26.398057 4766 generic.go:334] "Generic (PLEG): container finished" podID="85f22975-acba-4af1-8787-124a69cb68e1" containerID="8fb365e52e88bc33ca948883a43b83a045fad7d8bb2ec15511f497bc20abefe7" exitCode=0 Nov 26 00:46:26 crc kubenswrapper[4766]: I1126 00:46:26.398117 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85f22975-acba-4af1-8787-124a69cb68e1","Type":"ContainerDied","Data":"8fb365e52e88bc33ca948883a43b83a045fad7d8bb2ec15511f497bc20abefe7"} Nov 26 00:46:27 crc kubenswrapper[4766]: I1126 00:46:27.923262 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-668d6b6d67-r2rrd"] Nov 26 00:46:27 crc kubenswrapper[4766]: I1126 00:46:27.937547 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-668d6b6d67-r2rrd" Nov 26 00:46:27 crc kubenswrapper[4766]: I1126 00:46:27.943974 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5669dfd6f8-4wb69"] Nov 26 00:46:27 crc kubenswrapper[4766]: I1126 00:46:27.945359 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5669dfd6f8-4wb69" Nov 26 00:46:27 crc kubenswrapper[4766]: I1126 00:46:27.961723 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-6f6bc96c7-x82w9"] Nov 26 00:46:27 crc kubenswrapper[4766]: I1126 00:46:27.963368 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" Nov 26 00:46:27 crc kubenswrapper[4766]: I1126 00:46:27.972020 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-668d6b6d67-r2rrd"] Nov 26 00:46:27 crc kubenswrapper[4766]: I1126 00:46:27.989270 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6f6bc96c7-x82w9"] Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.001953 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5669dfd6f8-4wb69"] Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.039764 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/930724bb-b084-4ae4-8936-ea4ee7b80978-config-data\") pod \"heat-api-5669dfd6f8-4wb69\" (UID: \"930724bb-b084-4ae4-8936-ea4ee7b80978\") " pod="openstack/heat-api-5669dfd6f8-4wb69" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.039859 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-config-data-custom\") pod \"heat-cfnapi-6f6bc96c7-x82w9\" (UID: \"d173ebef-8b7b-46a1-b3b1-d8cd931dff50\") " pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.039954 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8bk7\" (UniqueName: \"kubernetes.io/projected/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-kube-api-access-d8bk7\") pod \"heat-cfnapi-6f6bc96c7-x82w9\" (UID: \"d173ebef-8b7b-46a1-b3b1-d8cd931dff50\") " pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.040005 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-config-data\") pod \"heat-cfnapi-6f6bc96c7-x82w9\" (UID: \"d173ebef-8b7b-46a1-b3b1-d8cd931dff50\") " pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.040062 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njkxg\" (UniqueName: \"kubernetes.io/projected/c0d62d37-c9d8-46f3-8fcd-802578c573fb-kube-api-access-njkxg\") pod \"heat-engine-668d6b6d67-r2rrd\" (UID: \"c0d62d37-c9d8-46f3-8fcd-802578c573fb\") " pod="openstack/heat-engine-668d6b6d67-r2rrd" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.040109 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bm5xl\" (UniqueName: \"kubernetes.io/projected/930724bb-b084-4ae4-8936-ea4ee7b80978-kube-api-access-bm5xl\") pod \"heat-api-5669dfd6f8-4wb69\" (UID: \"930724bb-b084-4ae4-8936-ea4ee7b80978\") " pod="openstack/heat-api-5669dfd6f8-4wb69" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.040135 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/930724bb-b084-4ae4-8936-ea4ee7b80978-config-data-custom\") pod \"heat-api-5669dfd6f8-4wb69\" (UID: \"930724bb-b084-4ae4-8936-ea4ee7b80978\") " pod="openstack/heat-api-5669dfd6f8-4wb69" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.040204 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0d62d37-c9d8-46f3-8fcd-802578c573fb-config-data-custom\") pod \"heat-engine-668d6b6d67-r2rrd\" (UID: \"c0d62d37-c9d8-46f3-8fcd-802578c573fb\") " pod="openstack/heat-engine-668d6b6d67-r2rrd" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.040256 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0d62d37-c9d8-46f3-8fcd-802578c573fb-combined-ca-bundle\") pod \"heat-engine-668d6b6d67-r2rrd\" (UID: \"c0d62d37-c9d8-46f3-8fcd-802578c573fb\") " pod="openstack/heat-engine-668d6b6d67-r2rrd" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.040275 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0d62d37-c9d8-46f3-8fcd-802578c573fb-config-data\") pod \"heat-engine-668d6b6d67-r2rrd\" (UID: \"c0d62d37-c9d8-46f3-8fcd-802578c573fb\") " pod="openstack/heat-engine-668d6b6d67-r2rrd" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.040296 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/930724bb-b084-4ae4-8936-ea4ee7b80978-combined-ca-bundle\") pod \"heat-api-5669dfd6f8-4wb69\" (UID: \"930724bb-b084-4ae4-8936-ea4ee7b80978\") " pod="openstack/heat-api-5669dfd6f8-4wb69" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.040367 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-combined-ca-bundle\") pod \"heat-cfnapi-6f6bc96c7-x82w9\" (UID: \"d173ebef-8b7b-46a1-b3b1-d8cd931dff50\") " pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.142154 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8bk7\" (UniqueName: \"kubernetes.io/projected/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-kube-api-access-d8bk7\") pod \"heat-cfnapi-6f6bc96c7-x82w9\" (UID: \"d173ebef-8b7b-46a1-b3b1-d8cd931dff50\") " pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.142200 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-config-data\") pod \"heat-cfnapi-6f6bc96c7-x82w9\" (UID: \"d173ebef-8b7b-46a1-b3b1-d8cd931dff50\") " pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.142251 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njkxg\" (UniqueName: \"kubernetes.io/projected/c0d62d37-c9d8-46f3-8fcd-802578c573fb-kube-api-access-njkxg\") pod \"heat-engine-668d6b6d67-r2rrd\" (UID: \"c0d62d37-c9d8-46f3-8fcd-802578c573fb\") " pod="openstack/heat-engine-668d6b6d67-r2rrd" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.142273 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bm5xl\" (UniqueName: \"kubernetes.io/projected/930724bb-b084-4ae4-8936-ea4ee7b80978-kube-api-access-bm5xl\") pod \"heat-api-5669dfd6f8-4wb69\" (UID: \"930724bb-b084-4ae4-8936-ea4ee7b80978\") " pod="openstack/heat-api-5669dfd6f8-4wb69" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.142297 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/930724bb-b084-4ae4-8936-ea4ee7b80978-config-data-custom\") pod \"heat-api-5669dfd6f8-4wb69\" (UID: \"930724bb-b084-4ae4-8936-ea4ee7b80978\") " pod="openstack/heat-api-5669dfd6f8-4wb69" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.142332 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0d62d37-c9d8-46f3-8fcd-802578c573fb-config-data-custom\") pod \"heat-engine-668d6b6d67-r2rrd\" (UID: \"c0d62d37-c9d8-46f3-8fcd-802578c573fb\") " pod="openstack/heat-engine-668d6b6d67-r2rrd" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.142353 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0d62d37-c9d8-46f3-8fcd-802578c573fb-combined-ca-bundle\") pod \"heat-engine-668d6b6d67-r2rrd\" (UID: \"c0d62d37-c9d8-46f3-8fcd-802578c573fb\") " pod="openstack/heat-engine-668d6b6d67-r2rrd" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.142369 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0d62d37-c9d8-46f3-8fcd-802578c573fb-config-data\") pod \"heat-engine-668d6b6d67-r2rrd\" (UID: \"c0d62d37-c9d8-46f3-8fcd-802578c573fb\") " pod="openstack/heat-engine-668d6b6d67-r2rrd" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.142385 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/930724bb-b084-4ae4-8936-ea4ee7b80978-combined-ca-bundle\") pod \"heat-api-5669dfd6f8-4wb69\" (UID: \"930724bb-b084-4ae4-8936-ea4ee7b80978\") " pod="openstack/heat-api-5669dfd6f8-4wb69" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.142421 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-combined-ca-bundle\") pod \"heat-cfnapi-6f6bc96c7-x82w9\" (UID: \"d173ebef-8b7b-46a1-b3b1-d8cd931dff50\") " pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.142488 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/930724bb-b084-4ae4-8936-ea4ee7b80978-config-data\") pod \"heat-api-5669dfd6f8-4wb69\" (UID: \"930724bb-b084-4ae4-8936-ea4ee7b80978\") " pod="openstack/heat-api-5669dfd6f8-4wb69" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.142524 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-config-data-custom\") pod \"heat-cfnapi-6f6bc96c7-x82w9\" (UID: \"d173ebef-8b7b-46a1-b3b1-d8cd931dff50\") " pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.150932 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-config-data\") pod \"heat-cfnapi-6f6bc96c7-x82w9\" (UID: \"d173ebef-8b7b-46a1-b3b1-d8cd931dff50\") " pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.154770 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0d62d37-c9d8-46f3-8fcd-802578c573fb-combined-ca-bundle\") pod \"heat-engine-668d6b6d67-r2rrd\" (UID: \"c0d62d37-c9d8-46f3-8fcd-802578c573fb\") " pod="openstack/heat-engine-668d6b6d67-r2rrd" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.157117 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-combined-ca-bundle\") pod \"heat-cfnapi-6f6bc96c7-x82w9\" (UID: \"d173ebef-8b7b-46a1-b3b1-d8cd931dff50\") " pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.157133 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0d62d37-c9d8-46f3-8fcd-802578c573fb-config-data-custom\") pod \"heat-engine-668d6b6d67-r2rrd\" (UID: \"c0d62d37-c9d8-46f3-8fcd-802578c573fb\") " pod="openstack/heat-engine-668d6b6d67-r2rrd" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.160061 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/930724bb-b084-4ae4-8936-ea4ee7b80978-config-data\") pod \"heat-api-5669dfd6f8-4wb69\" (UID: \"930724bb-b084-4ae4-8936-ea4ee7b80978\") " pod="openstack/heat-api-5669dfd6f8-4wb69" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.172555 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-config-data-custom\") pod \"heat-cfnapi-6f6bc96c7-x82w9\" (UID: \"d173ebef-8b7b-46a1-b3b1-d8cd931dff50\") " pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.172883 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njkxg\" (UniqueName: \"kubernetes.io/projected/c0d62d37-c9d8-46f3-8fcd-802578c573fb-kube-api-access-njkxg\") pod \"heat-engine-668d6b6d67-r2rrd\" (UID: \"c0d62d37-c9d8-46f3-8fcd-802578c573fb\") " pod="openstack/heat-engine-668d6b6d67-r2rrd" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.175596 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0d62d37-c9d8-46f3-8fcd-802578c573fb-config-data\") pod \"heat-engine-668d6b6d67-r2rrd\" (UID: \"c0d62d37-c9d8-46f3-8fcd-802578c573fb\") " pod="openstack/heat-engine-668d6b6d67-r2rrd" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.176706 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/930724bb-b084-4ae4-8936-ea4ee7b80978-config-data-custom\") pod \"heat-api-5669dfd6f8-4wb69\" (UID: \"930724bb-b084-4ae4-8936-ea4ee7b80978\") " pod="openstack/heat-api-5669dfd6f8-4wb69" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.178323 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8bk7\" (UniqueName: \"kubernetes.io/projected/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-kube-api-access-d8bk7\") pod \"heat-cfnapi-6f6bc96c7-x82w9\" (UID: \"d173ebef-8b7b-46a1-b3b1-d8cd931dff50\") " pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.192611 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/930724bb-b084-4ae4-8936-ea4ee7b80978-combined-ca-bundle\") pod \"heat-api-5669dfd6f8-4wb69\" (UID: \"930724bb-b084-4ae4-8936-ea4ee7b80978\") " pod="openstack/heat-api-5669dfd6f8-4wb69" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.197573 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bm5xl\" (UniqueName: \"kubernetes.io/projected/930724bb-b084-4ae4-8936-ea4ee7b80978-kube-api-access-bm5xl\") pod \"heat-api-5669dfd6f8-4wb69\" (UID: \"930724bb-b084-4ae4-8936-ea4ee7b80978\") " pod="openstack/heat-api-5669dfd6f8-4wb69" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.275661 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-668d6b6d67-r2rrd" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.290358 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5669dfd6f8-4wb69" Nov 26 00:46:28 crc kubenswrapper[4766]: I1126 00:46:28.306956 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" Nov 26 00:46:29 crc kubenswrapper[4766]: I1126 00:46:29.204119 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 26 00:46:29 crc kubenswrapper[4766]: I1126 00:46:29.326109 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="85f22975-acba-4af1-8787-124a69cb68e1" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.187:3000/\": dial tcp 10.217.0.187:3000: connect: connection refused" Nov 26 00:46:30 crc kubenswrapper[4766]: E1126 00:46:30.360158 4766 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85f22975_acba_4af1_8787_124a69cb68e1.slice/crio-conmon-2b52ab898b9b6da7a8b73b6a7b2491a7f5ae1f9e232a31b3bad2d13612f5f31d.scope\": RecentStats: unable to find data in memory cache]" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.455982 4766 generic.go:334] "Generic (PLEG): container finished" podID="85f22975-acba-4af1-8787-124a69cb68e1" containerID="2b52ab898b9b6da7a8b73b6a7b2491a7f5ae1f9e232a31b3bad2d13612f5f31d" exitCode=0 Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.456049 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85f22975-acba-4af1-8787-124a69cb68e1","Type":"ContainerDied","Data":"2b52ab898b9b6da7a8b73b6a7b2491a7f5ae1f9e232a31b3bad2d13612f5f31d"} Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.711868 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-856c5c6f86-fk66f"] Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.731554 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-b8c9d9569-lrlx6"] Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.745500 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-6d74f96c45-v8sqm"] Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.747602 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.750802 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.761348 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6d74f96c45-v8sqm"] Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.772539 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.785166 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-864c6d46fc-tchw8"] Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.786519 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.800476 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.800707 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.808822 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-864c6d46fc-tchw8"] Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.809045 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-public-tls-certs\") pod \"heat-cfnapi-864c6d46fc-tchw8\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.809133 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-config-data\") pod \"heat-api-6d74f96c45-v8sqm\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.809163 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-internal-tls-certs\") pod \"heat-api-6d74f96c45-v8sqm\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.809187 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-config-data-custom\") pod \"heat-api-6d74f96c45-v8sqm\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.809221 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-public-tls-certs\") pod \"heat-api-6d74f96c45-v8sqm\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.809391 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-combined-ca-bundle\") pod \"heat-api-6d74f96c45-v8sqm\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.809442 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7smgn\" (UniqueName: \"kubernetes.io/projected/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-kube-api-access-7smgn\") pod \"heat-api-6d74f96c45-v8sqm\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.809515 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-config-data\") pod \"heat-cfnapi-864c6d46fc-tchw8\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.809553 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckrt5\" (UniqueName: \"kubernetes.io/projected/69584201-dc28-478a-a45c-ede768957725-kube-api-access-ckrt5\") pod \"heat-cfnapi-864c6d46fc-tchw8\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.809601 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-internal-tls-certs\") pod \"heat-cfnapi-864c6d46fc-tchw8\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.809635 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-combined-ca-bundle\") pod \"heat-cfnapi-864c6d46fc-tchw8\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.809789 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-config-data-custom\") pod \"heat-cfnapi-864c6d46fc-tchw8\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.912083 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-config-data-custom\") pod \"heat-cfnapi-864c6d46fc-tchw8\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.912165 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-public-tls-certs\") pod \"heat-cfnapi-864c6d46fc-tchw8\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.912263 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-config-data\") pod \"heat-api-6d74f96c45-v8sqm\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.912290 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-internal-tls-certs\") pod \"heat-api-6d74f96c45-v8sqm\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.912327 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-config-data-custom\") pod \"heat-api-6d74f96c45-v8sqm\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.912360 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-public-tls-certs\") pod \"heat-api-6d74f96c45-v8sqm\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.912417 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-combined-ca-bundle\") pod \"heat-api-6d74f96c45-v8sqm\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.912466 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7smgn\" (UniqueName: \"kubernetes.io/projected/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-kube-api-access-7smgn\") pod \"heat-api-6d74f96c45-v8sqm\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.912532 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-config-data\") pod \"heat-cfnapi-864c6d46fc-tchw8\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.912566 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckrt5\" (UniqueName: \"kubernetes.io/projected/69584201-dc28-478a-a45c-ede768957725-kube-api-access-ckrt5\") pod \"heat-cfnapi-864c6d46fc-tchw8\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.912608 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-internal-tls-certs\") pod \"heat-cfnapi-864c6d46fc-tchw8\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.912641 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-combined-ca-bundle\") pod \"heat-cfnapi-864c6d46fc-tchw8\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.920410 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-config-data\") pod \"heat-cfnapi-864c6d46fc-tchw8\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.921516 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-internal-tls-certs\") pod \"heat-cfnapi-864c6d46fc-tchw8\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.921704 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-config-data\") pod \"heat-api-6d74f96c45-v8sqm\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.921963 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-config-data-custom\") pod \"heat-api-6d74f96c45-v8sqm\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.923136 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-combined-ca-bundle\") pod \"heat-cfnapi-864c6d46fc-tchw8\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.924061 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-public-tls-certs\") pod \"heat-api-6d74f96c45-v8sqm\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.924435 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-combined-ca-bundle\") pod \"heat-api-6d74f96c45-v8sqm\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.926334 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-public-tls-certs\") pod \"heat-cfnapi-864c6d46fc-tchw8\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.935521 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-config-data-custom\") pod \"heat-cfnapi-864c6d46fc-tchw8\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.935616 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7smgn\" (UniqueName: \"kubernetes.io/projected/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-kube-api-access-7smgn\") pod \"heat-api-6d74f96c45-v8sqm\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.936725 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckrt5\" (UniqueName: \"kubernetes.io/projected/69584201-dc28-478a-a45c-ede768957725-kube-api-access-ckrt5\") pod \"heat-cfnapi-864c6d46fc-tchw8\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:30 crc kubenswrapper[4766]: I1126 00:46:30.948961 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-internal-tls-certs\") pod \"heat-api-6d74f96c45-v8sqm\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.075908 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.114273 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.670033 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.739450 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85f22975-acba-4af1-8787-124a69cb68e1-run-httpd\") pod \"85f22975-acba-4af1-8787-124a69cb68e1\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.739801 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-sg-core-conf-yaml\") pod \"85f22975-acba-4af1-8787-124a69cb68e1\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.739829 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-scripts\") pod \"85f22975-acba-4af1-8787-124a69cb68e1\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.739885 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85f22975-acba-4af1-8787-124a69cb68e1-log-httpd\") pod \"85f22975-acba-4af1-8787-124a69cb68e1\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.739949 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlf77\" (UniqueName: \"kubernetes.io/projected/85f22975-acba-4af1-8787-124a69cb68e1-kube-api-access-hlf77\") pod \"85f22975-acba-4af1-8787-124a69cb68e1\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.740057 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-config-data\") pod \"85f22975-acba-4af1-8787-124a69cb68e1\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.740089 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-combined-ca-bundle\") pod \"85f22975-acba-4af1-8787-124a69cb68e1\" (UID: \"85f22975-acba-4af1-8787-124a69cb68e1\") " Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.742822 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85f22975-acba-4af1-8787-124a69cb68e1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "85f22975-acba-4af1-8787-124a69cb68e1" (UID: "85f22975-acba-4af1-8787-124a69cb68e1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.746603 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85f22975-acba-4af1-8787-124a69cb68e1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "85f22975-acba-4af1-8787-124a69cb68e1" (UID: "85f22975-acba-4af1-8787-124a69cb68e1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.754695 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-scripts" (OuterVolumeSpecName: "scripts") pod "85f22975-acba-4af1-8787-124a69cb68e1" (UID: "85f22975-acba-4af1-8787-124a69cb68e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.770429 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85f22975-acba-4af1-8787-124a69cb68e1-kube-api-access-hlf77" (OuterVolumeSpecName: "kube-api-access-hlf77") pod "85f22975-acba-4af1-8787-124a69cb68e1" (UID: "85f22975-acba-4af1-8787-124a69cb68e1"). InnerVolumeSpecName "kube-api-access-hlf77". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.809372 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "85f22975-acba-4af1-8787-124a69cb68e1" (UID: "85f22975-acba-4af1-8787-124a69cb68e1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.843104 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlf77\" (UniqueName: \"kubernetes.io/projected/85f22975-acba-4af1-8787-124a69cb68e1-kube-api-access-hlf77\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.843135 4766 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85f22975-acba-4af1-8787-124a69cb68e1-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.843147 4766 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.843157 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.843166 4766 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85f22975-acba-4af1-8787-124a69cb68e1-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.922059 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85f22975-acba-4af1-8787-124a69cb68e1" (UID: "85f22975-acba-4af1-8787-124a69cb68e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.926050 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-xtpgd"] Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.932965 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-config-data" (OuterVolumeSpecName: "config-data") pod "85f22975-acba-4af1-8787-124a69cb68e1" (UID: "85f22975-acba-4af1-8787-124a69cb68e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.946402 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:31 crc kubenswrapper[4766]: I1126 00:46:31.946687 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85f22975-acba-4af1-8787-124a69cb68e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.496331 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"d8551b13-7602-4c7c-bfb5-0bcf7e64954d","Type":"ContainerStarted","Data":"0d8cd9d1e22451e17419a91be75b05c560437085371d327ae640f7facd438799"} Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.499814 4766 generic.go:334] "Generic (PLEG): container finished" podID="22305d36-865a-4fe0-8337-92009c1c3c29" containerID="a615525881a275d3591fca34860f3c0ccc52cf8df614d018dcb30b4f1d99e563" exitCode=0 Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.499966 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" event={"ID":"22305d36-865a-4fe0-8337-92009c1c3c29","Type":"ContainerDied","Data":"a615525881a275d3591fca34860f3c0ccc52cf8df614d018dcb30b4f1d99e563"} Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.500027 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" event={"ID":"22305d36-865a-4fe0-8337-92009c1c3c29","Type":"ContainerStarted","Data":"ae8f81a11891320b7c03174056ad671c03e5aeefa79dbfe2cb5db6346bf64469"} Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.505992 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85f22975-acba-4af1-8787-124a69cb68e1","Type":"ContainerDied","Data":"ad698c6d5cf8113a583acc6ceeddde765e77c38680960fd925d00c490af7d4c9"} Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.506046 4766 scope.go:117] "RemoveContainer" containerID="d24fe490bba51a44c3ebde3231849642624486ad1a4ea2d892d19770d1a77039" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.506206 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.517553 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.011656577 podStartE2EDuration="18.517536853s" podCreationTimestamp="2025-11-26 00:46:14 +0000 UTC" firstStartedPulling="2025-11-26 00:46:15.72133648 +0000 UTC m=+1356.570106910" lastFinishedPulling="2025-11-26 00:46:31.227216756 +0000 UTC m=+1372.075987186" observedRunningTime="2025-11-26 00:46:32.516168719 +0000 UTC m=+1373.364939149" watchObservedRunningTime="2025-11-26 00:46:32.517536853 +0000 UTC m=+1373.366307273" Nov 26 00:46:32 crc kubenswrapper[4766]: W1126 00:46:32.584719 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4442abe_9e23_4f4e_880c_89ca9d8ef785.slice/crio-21ccd534a3c8245de0c1ae495219c0024a5d7281a517c44b7030c1563c7f24d3 WatchSource:0}: Error finding container 21ccd534a3c8245de0c1ae495219c0024a5d7281a517c44b7030c1563c7f24d3: Status 404 returned error can't find the container with id 21ccd534a3c8245de0c1ae495219c0024a5d7281a517c44b7030c1563c7f24d3 Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.624403 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6f6bc96c7-x82w9"] Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.650354 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-b8c9d9569-lrlx6"] Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.667158 4766 scope.go:117] "RemoveContainer" containerID="48d5602c1bf28f3d5d1828b4c020f553a9f7122a9bc64b06339dbed445c30f1a" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.700614 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-668d6b6d67-r2rrd"] Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.737756 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-856c5c6f86-fk66f"] Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.753587 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5669dfd6f8-4wb69"] Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.780067 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.797001 4766 scope.go:117] "RemoveContainer" containerID="2b52ab898b9b6da7a8b73b6a7b2491a7f5ae1f9e232a31b3bad2d13612f5f31d" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.836060 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.851523 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:46:32 crc kubenswrapper[4766]: E1126 00:46:32.852269 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85f22975-acba-4af1-8787-124a69cb68e1" containerName="ceilometer-central-agent" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.852287 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="85f22975-acba-4af1-8787-124a69cb68e1" containerName="ceilometer-central-agent" Nov 26 00:46:32 crc kubenswrapper[4766]: E1126 00:46:32.852303 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85f22975-acba-4af1-8787-124a69cb68e1" containerName="proxy-httpd" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.852309 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="85f22975-acba-4af1-8787-124a69cb68e1" containerName="proxy-httpd" Nov 26 00:46:32 crc kubenswrapper[4766]: E1126 00:46:32.852323 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85f22975-acba-4af1-8787-124a69cb68e1" containerName="ceilometer-notification-agent" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.852329 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="85f22975-acba-4af1-8787-124a69cb68e1" containerName="ceilometer-notification-agent" Nov 26 00:46:32 crc kubenswrapper[4766]: E1126 00:46:32.852351 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85f22975-acba-4af1-8787-124a69cb68e1" containerName="sg-core" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.852357 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="85f22975-acba-4af1-8787-124a69cb68e1" containerName="sg-core" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.852561 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="85f22975-acba-4af1-8787-124a69cb68e1" containerName="ceilometer-central-agent" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.852583 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="85f22975-acba-4af1-8787-124a69cb68e1" containerName="proxy-httpd" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.852596 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="85f22975-acba-4af1-8787-124a69cb68e1" containerName="sg-core" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.852604 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="85f22975-acba-4af1-8787-124a69cb68e1" containerName="ceilometer-notification-agent" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.855002 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.861312 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.861569 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.870916 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-74d678b987-q89lp"] Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.873791 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " pod="openstack/ceilometer-0" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.873841 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-scripts\") pod \"ceilometer-0\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " pod="openstack/ceilometer-0" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.873859 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae9e2d8d-6ce6-490a-8463-a078626b551f-log-httpd\") pod \"ceilometer-0\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " pod="openstack/ceilometer-0" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.873877 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " pod="openstack/ceilometer-0" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.873897 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae9e2d8d-6ce6-490a-8463-a078626b551f-run-httpd\") pod \"ceilometer-0\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " pod="openstack/ceilometer-0" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.873936 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27ft2\" (UniqueName: \"kubernetes.io/projected/ae9e2d8d-6ce6-490a-8463-a078626b551f-kube-api-access-27ft2\") pod \"ceilometer-0\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " pod="openstack/ceilometer-0" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.873979 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-config-data\") pod \"ceilometer-0\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " pod="openstack/ceilometer-0" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.887672 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.887959 4766 scope.go:117] "RemoveContainer" containerID="8fb365e52e88bc33ca948883a43b83a045fad7d8bb2ec15511f497bc20abefe7" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.899661 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6dfcdbc5c7-znfmt"] Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.975934 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " pod="openstack/ceilometer-0" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.975989 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-scripts\") pod \"ceilometer-0\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " pod="openstack/ceilometer-0" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.976009 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae9e2d8d-6ce6-490a-8463-a078626b551f-log-httpd\") pod \"ceilometer-0\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " pod="openstack/ceilometer-0" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.976024 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " pod="openstack/ceilometer-0" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.976047 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae9e2d8d-6ce6-490a-8463-a078626b551f-run-httpd\") pod \"ceilometer-0\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " pod="openstack/ceilometer-0" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.976096 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27ft2\" (UniqueName: \"kubernetes.io/projected/ae9e2d8d-6ce6-490a-8463-a078626b551f-kube-api-access-27ft2\") pod \"ceilometer-0\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " pod="openstack/ceilometer-0" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.976139 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-config-data\") pod \"ceilometer-0\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " pod="openstack/ceilometer-0" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.976505 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-864c6d46fc-tchw8"] Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.977434 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae9e2d8d-6ce6-490a-8463-a078626b551f-log-httpd\") pod \"ceilometer-0\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " pod="openstack/ceilometer-0" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.979909 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae9e2d8d-6ce6-490a-8463-a078626b551f-run-httpd\") pod \"ceilometer-0\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " pod="openstack/ceilometer-0" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.981400 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " pod="openstack/ceilometer-0" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.981578 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-config-data\") pod \"ceilometer-0\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " pod="openstack/ceilometer-0" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.987266 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-scripts\") pod \"ceilometer-0\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " pod="openstack/ceilometer-0" Nov 26 00:46:32 crc kubenswrapper[4766]: I1126 00:46:32.994130 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27ft2\" (UniqueName: \"kubernetes.io/projected/ae9e2d8d-6ce6-490a-8463-a078626b551f-kube-api-access-27ft2\") pod \"ceilometer-0\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " pod="openstack/ceilometer-0" Nov 26 00:46:33 crc kubenswrapper[4766]: W1126 00:46:33.012566 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69584201_dc28_478a_a45c_ede768957725.slice/crio-6e2942c9d25e30c668f40d5f83f7675185d231dedbbd07e031b2f9a4cf612f8c WatchSource:0}: Error finding container 6e2942c9d25e30c668f40d5f83f7675185d231dedbbd07e031b2f9a4cf612f8c: Status 404 returned error can't find the container with id 6e2942c9d25e30c668f40d5f83f7675185d231dedbbd07e031b2f9a4cf612f8c Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.013620 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " pod="openstack/ceilometer-0" Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.019128 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6d74f96c45-v8sqm"] Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.195600 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.551295 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" event={"ID":"d173ebef-8b7b-46a1-b3b1-d8cd931dff50","Type":"ContainerStarted","Data":"dbb8ce7c262109558e2316ab31cec4d1bc1d8d32d473d05609d14b3cce6b4662"} Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.554254 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-864c6d46fc-tchw8" event={"ID":"69584201-dc28-478a-a45c-ede768957725","Type":"ContainerStarted","Data":"6e2942c9d25e30c668f40d5f83f7675185d231dedbbd07e031b2f9a4cf612f8c"} Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.567969 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-668d6b6d67-r2rrd" event={"ID":"c0d62d37-c9d8-46f3-8fcd-802578c573fb","Type":"ContainerStarted","Data":"c5c61611bbcbcf3f57f04c4c93929a9c73ad6b0ae1d455e0599b25ed4f3f2da3"} Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.568050 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-668d6b6d67-r2rrd" event={"ID":"c0d62d37-c9d8-46f3-8fcd-802578c573fb","Type":"ContainerStarted","Data":"5d26fcc7a8368e29a1bb2c4edb3e42a99f9d7886bff196cd60e5372b1ead65b0"} Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.568079 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-668d6b6d67-r2rrd" Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.590090 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5669dfd6f8-4wb69" event={"ID":"930724bb-b084-4ae4-8936-ea4ee7b80978","Type":"ContainerStarted","Data":"bb9302cf6e621dc4dda4542b7166e9fca1f584bda054f5e7ed21e34babb64c36"} Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.591037 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-668d6b6d67-r2rrd" podStartSLOduration=6.590631256 podStartE2EDuration="6.590631256s" podCreationTimestamp="2025-11-26 00:46:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:46:33.585681482 +0000 UTC m=+1374.434451912" watchObservedRunningTime="2025-11-26 00:46:33.590631256 +0000 UTC m=+1374.439401686" Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.613773 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" event={"ID":"d4442abe-9e23-4f4e-880c-89ca9d8ef785","Type":"ContainerStarted","Data":"21ccd534a3c8245de0c1ae495219c0024a5d7281a517c44b7030c1563c7f24d3"} Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.621713 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-74d678b987-q89lp" event={"ID":"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c","Type":"ContainerStarted","Data":"a9bbbed4ab7184198fc3e25a03f7896d2db79c1b1f31fab8b6bc2ccc487c5bdb"} Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.621756 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-74d678b987-q89lp" event={"ID":"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c","Type":"ContainerStarted","Data":"e1a220dd49f6f74e404a0d8c8a43c2fc4eb3561135c1dcb0d7db70e42fb1906a"} Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.622991 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.623051 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.626889 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" event={"ID":"22305d36-865a-4fe0-8337-92009c1c3c29","Type":"ContainerStarted","Data":"a24b82cf93de432af5c5aaac20f2f469f4bb029ec14c1b455b23cf107a1d967d"} Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.627182 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.631297 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6dfcdbc5c7-znfmt" event={"ID":"13e19aaf-7cc3-40a6-852b-72b77b37cf1e","Type":"ContainerStarted","Data":"10340d467e512bd68e6e6b88a6746642d2c9504b21804f3be63d9c75ddd06bd8"} Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.631347 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6dfcdbc5c7-znfmt" event={"ID":"13e19aaf-7cc3-40a6-852b-72b77b37cf1e","Type":"ContainerStarted","Data":"0e23105208a328d7118d59c228dbc87fa4ee01cb7e3a50b48fa25c206c1c9c52"} Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.631561 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-6dfcdbc5c7-znfmt" Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.654842 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-856c5c6f86-fk66f" event={"ID":"16337a9e-8651-492e-9f39-21515275e3cb","Type":"ContainerStarted","Data":"30c3c0b146e53a755a19931c1c4a4daedb4f44e55542080a7073ec8139c7ccd5"} Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.665477 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6d74f96c45-v8sqm" event={"ID":"d37231fe-97e9-4f01-b914-e0bd6ebfcf21","Type":"ContainerStarted","Data":"228523eac023b6c65a0a7320b9fd6ff6f49765792458a5e4d9a99994ad531fc4"} Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.666599 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-74d678b987-q89lp" podStartSLOduration=11.666587 podStartE2EDuration="11.666587s" podCreationTimestamp="2025-11-26 00:46:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:46:33.655208835 +0000 UTC m=+1374.503979265" watchObservedRunningTime="2025-11-26 00:46:33.666587 +0000 UTC m=+1374.515357430" Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.682055 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" podStartSLOduration=12.682039187 podStartE2EDuration="12.682039187s" podCreationTimestamp="2025-11-26 00:46:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:46:33.676491528 +0000 UTC m=+1374.525261958" watchObservedRunningTime="2025-11-26 00:46:33.682039187 +0000 UTC m=+1374.530809617" Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.712495 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-6dfcdbc5c7-znfmt" podStartSLOduration=12.71247231 podStartE2EDuration="12.71247231s" podCreationTimestamp="2025-11-26 00:46:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:46:33.708847279 +0000 UTC m=+1374.557617709" watchObservedRunningTime="2025-11-26 00:46:33.71247231 +0000 UTC m=+1374.561242740" Nov 26 00:46:33 crc kubenswrapper[4766]: W1126 00:46:33.780324 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae9e2d8d_6ce6_490a_8463_a078626b551f.slice/crio-47739c076815090533bc3fe60cf248c55e509bb6013b7a28d69fb1e9c273b18f WatchSource:0}: Error finding container 47739c076815090533bc3fe60cf248c55e509bb6013b7a28d69fb1e9c273b18f: Status 404 returned error can't find the container with id 47739c076815090533bc3fe60cf248c55e509bb6013b7a28d69fb1e9c273b18f Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.788807 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:46:33 crc kubenswrapper[4766]: I1126 00:46:33.848789 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85f22975-acba-4af1-8787-124a69cb68e1" path="/var/lib/kubelet/pods/85f22975-acba-4af1-8787-124a69cb68e1/volumes" Nov 26 00:46:34 crc kubenswrapper[4766]: I1126 00:46:34.690741 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae9e2d8d-6ce6-490a-8463-a078626b551f","Type":"ContainerStarted","Data":"47739c076815090533bc3fe60cf248c55e509bb6013b7a28d69fb1e9c273b18f"} Nov 26 00:46:34 crc kubenswrapper[4766]: I1126 00:46:34.696999 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-74d678b987-q89lp" event={"ID":"3d4e6997-4bbc-4cb2-a0f2-abf3b940090c","Type":"ContainerStarted","Data":"dad2247dc27d1c6e038a23b1e97967654f3b8a5401dd6f9aaedeb0d68eee5a57"} Nov 26 00:46:35 crc kubenswrapper[4766]: I1126 00:46:35.711397 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae9e2d8d-6ce6-490a-8463-a078626b551f","Type":"ContainerStarted","Data":"c55670df27b91fd6586d9d88a8f46c9f9ff1651c0f04ee65aa18cbb449d8df61"} Nov 26 00:46:35 crc kubenswrapper[4766]: I1126 00:46:35.824045 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.746015 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" event={"ID":"d4442abe-9e23-4f4e-880c-89ca9d8ef785","Type":"ContainerStarted","Data":"9242e3ede9b06520f5b153cd3544ca4062ce8e5818ecaf3d453337c2fddf0419"} Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.747327 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" podUID="d4442abe-9e23-4f4e-880c-89ca9d8ef785" containerName="heat-cfnapi" containerID="cri-o://9242e3ede9b06520f5b153cd3544ca4062ce8e5818ecaf3d453337c2fddf0419" gracePeriod=60 Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.747786 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.751841 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" event={"ID":"d173ebef-8b7b-46a1-b3b1-d8cd931dff50","Type":"ContainerStarted","Data":"6870abe3616f0bffb71336cff27e044286d42110b5b85c6daaf1d11810dbb9fb"} Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.752801 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.763805 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-856c5c6f86-fk66f" event={"ID":"16337a9e-8651-492e-9f39-21515275e3cb","Type":"ContainerStarted","Data":"0392ad20ec296c2eddec755e7b42f4a4b721787556d71d58364e5cb5612bf3e0"} Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.764167 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-856c5c6f86-fk66f" Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.764439 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-856c5c6f86-fk66f" podUID="16337a9e-8651-492e-9f39-21515275e3cb" containerName="heat-api" containerID="cri-o://0392ad20ec296c2eddec755e7b42f4a4b721787556d71d58364e5cb5612bf3e0" gracePeriod=60 Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.782386 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-864c6d46fc-tchw8" event={"ID":"69584201-dc28-478a-a45c-ede768957725","Type":"ContainerStarted","Data":"97f6d0894f2afe83918ab8a0506b5257dab87c66af95e6b9a225f13af601974a"} Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.782772 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.787215 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6d74f96c45-v8sqm" event={"ID":"d37231fe-97e9-4f01-b914-e0bd6ebfcf21","Type":"ContainerStarted","Data":"49046a5ad31c38fa307b5b32dc33c292711774a7a49cba0bbb076428eacad0e9"} Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.799200 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.839685 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0f76ca58-391f-43c9-810b-f80f94213a31","Type":"ContainerDied","Data":"d1c39cb41e9ec9c525837a97a31e77e47f4ff563f15ae600a44c0e8b141327d1"} Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.803628 4766 generic.go:334] "Generic (PLEG): container finished" podID="0f76ca58-391f-43c9-810b-f80f94213a31" containerID="d1c39cb41e9ec9c525837a97a31e77e47f4ff563f15ae600a44c0e8b141327d1" exitCode=137 Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.863989 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae9e2d8d-6ce6-490a-8463-a078626b551f","Type":"ContainerStarted","Data":"11e1c320e18b1ba9311a6168bb1564ba38aaaf3a89e6ed61fa800644fa78f7b1"} Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.866018 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" podStartSLOduration=12.699295691 podStartE2EDuration="15.865996738s" podCreationTimestamp="2025-11-26 00:46:21 +0000 UTC" firstStartedPulling="2025-11-26 00:46:32.590625696 +0000 UTC m=+1373.439396126" lastFinishedPulling="2025-11-26 00:46:35.757326743 +0000 UTC m=+1376.606097173" observedRunningTime="2025-11-26 00:46:36.787021858 +0000 UTC m=+1377.635792298" watchObservedRunningTime="2025-11-26 00:46:36.865996738 +0000 UTC m=+1377.714767168" Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.868947 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5669dfd6f8-4wb69" event={"ID":"930724bb-b084-4ae4-8936-ea4ee7b80978","Type":"ContainerStarted","Data":"de8d52ac669ab0babed559fa3fbfb2066c6ee8f12ae83a771a2cc4b1377fa3f2"} Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.871086 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5669dfd6f8-4wb69" Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.880740 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-864c6d46fc-tchw8" podStartSLOduration=4.156255426 podStartE2EDuration="6.880717577s" podCreationTimestamp="2025-11-26 00:46:30 +0000 UTC" firstStartedPulling="2025-11-26 00:46:33.033102168 +0000 UTC m=+1373.881872598" lastFinishedPulling="2025-11-26 00:46:35.757564319 +0000 UTC m=+1376.606334749" observedRunningTime="2025-11-26 00:46:36.818281051 +0000 UTC m=+1377.667051481" watchObservedRunningTime="2025-11-26 00:46:36.880717577 +0000 UTC m=+1377.729488007" Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.909038 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" podStartSLOduration=6.740403581 podStartE2EDuration="9.909010516s" podCreationTimestamp="2025-11-26 00:46:27 +0000 UTC" firstStartedPulling="2025-11-26 00:46:32.591034686 +0000 UTC m=+1373.439805116" lastFinishedPulling="2025-11-26 00:46:35.759641621 +0000 UTC m=+1376.608412051" observedRunningTime="2025-11-26 00:46:36.844415446 +0000 UTC m=+1377.693185876" watchObservedRunningTime="2025-11-26 00:46:36.909010516 +0000 UTC m=+1377.757780966" Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.929623 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-856c5c6f86-fk66f" podStartSLOduration=11.803057511 podStartE2EDuration="14.929578041s" podCreationTimestamp="2025-11-26 00:46:22 +0000 UTC" firstStartedPulling="2025-11-26 00:46:32.636216779 +0000 UTC m=+1373.484987209" lastFinishedPulling="2025-11-26 00:46:35.762737319 +0000 UTC m=+1376.611507739" observedRunningTime="2025-11-26 00:46:36.871724311 +0000 UTC m=+1377.720494751" watchObservedRunningTime="2025-11-26 00:46:36.929578041 +0000 UTC m=+1377.778348471" Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.946706 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-5669dfd6f8-4wb69" podStartSLOduration=6.862505161 podStartE2EDuration="9.94668159s" podCreationTimestamp="2025-11-26 00:46:27 +0000 UTC" firstStartedPulling="2025-11-26 00:46:32.673143864 +0000 UTC m=+1373.521914294" lastFinishedPulling="2025-11-26 00:46:35.757320293 +0000 UTC m=+1376.606090723" observedRunningTime="2025-11-26 00:46:36.912229197 +0000 UTC m=+1377.760999627" watchObservedRunningTime="2025-11-26 00:46:36.94668159 +0000 UTC m=+1377.795452020" Nov 26 00:46:36 crc kubenswrapper[4766]: I1126 00:46:36.953559 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-6d74f96c45-v8sqm" podStartSLOduration=4.235038191 podStartE2EDuration="6.953542442s" podCreationTimestamp="2025-11-26 00:46:30 +0000 UTC" firstStartedPulling="2025-11-26 00:46:33.043002737 +0000 UTC m=+1373.891773167" lastFinishedPulling="2025-11-26 00:46:35.761506988 +0000 UTC m=+1376.610277418" observedRunningTime="2025-11-26 00:46:36.940133896 +0000 UTC m=+1377.788904326" watchObservedRunningTime="2025-11-26 00:46:36.953542442 +0000 UTC m=+1377.802312872" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.295150 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.321932 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.390348 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-config-data-custom\") pod \"0f76ca58-391f-43c9-810b-f80f94213a31\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.390488 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-config-data\") pod \"0f76ca58-391f-43c9-810b-f80f94213a31\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.390557 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f76ca58-391f-43c9-810b-f80f94213a31-logs\") pod \"0f76ca58-391f-43c9-810b-f80f94213a31\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.390577 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0f76ca58-391f-43c9-810b-f80f94213a31-etc-machine-id\") pod \"0f76ca58-391f-43c9-810b-f80f94213a31\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.390604 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rf9g\" (UniqueName: \"kubernetes.io/projected/0f76ca58-391f-43c9-810b-f80f94213a31-kube-api-access-7rf9g\") pod \"0f76ca58-391f-43c9-810b-f80f94213a31\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.390622 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-scripts\") pod \"0f76ca58-391f-43c9-810b-f80f94213a31\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.390700 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-combined-ca-bundle\") pod \"0f76ca58-391f-43c9-810b-f80f94213a31\" (UID: \"0f76ca58-391f-43c9-810b-f80f94213a31\") " Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.391604 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f76ca58-391f-43c9-810b-f80f94213a31-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0f76ca58-391f-43c9-810b-f80f94213a31" (UID: "0f76ca58-391f-43c9-810b-f80f94213a31"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.394989 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f76ca58-391f-43c9-810b-f80f94213a31-logs" (OuterVolumeSpecName: "logs") pod "0f76ca58-391f-43c9-810b-f80f94213a31" (UID: "0f76ca58-391f-43c9-810b-f80f94213a31"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.397843 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f76ca58-391f-43c9-810b-f80f94213a31-kube-api-access-7rf9g" (OuterVolumeSpecName: "kube-api-access-7rf9g") pod "0f76ca58-391f-43c9-810b-f80f94213a31" (UID: "0f76ca58-391f-43c9-810b-f80f94213a31"). InnerVolumeSpecName "kube-api-access-7rf9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.414821 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-scripts" (OuterVolumeSpecName: "scripts") pod "0f76ca58-391f-43c9-810b-f80f94213a31" (UID: "0f76ca58-391f-43c9-810b-f80f94213a31"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.418805 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0f76ca58-391f-43c9-810b-f80f94213a31" (UID: "0f76ca58-391f-43c9-810b-f80f94213a31"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.419629 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-dpnb6"] Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.420036 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" podUID="7cb97caa-8678-4b20-96e7-0ddf811bcc38" containerName="dnsmasq-dns" containerID="cri-o://cff0ffc937f2588a8daa1295e1eaefe470b14765cc13965c55dba86981dfc9fc" gracePeriod=10 Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.498760 4766 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f76ca58-391f-43c9-810b-f80f94213a31-logs\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.498790 4766 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0f76ca58-391f-43c9-810b-f80f94213a31-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.498800 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rf9g\" (UniqueName: \"kubernetes.io/projected/0f76ca58-391f-43c9-810b-f80f94213a31-kube-api-access-7rf9g\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.498808 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.498817 4766 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.561635 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f76ca58-391f-43c9-810b-f80f94213a31" (UID: "0f76ca58-391f-43c9-810b-f80f94213a31"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.601029 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.623802 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-config-data" (OuterVolumeSpecName: "config-data") pod "0f76ca58-391f-43c9-810b-f80f94213a31" (UID: "0f76ca58-391f-43c9-810b-f80f94213a31"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.703180 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f76ca58-391f-43c9-810b-f80f94213a31-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.890440 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0f76ca58-391f-43c9-810b-f80f94213a31","Type":"ContainerDied","Data":"b04115395f9da31610eb16d10f8701131c2b93ad59ab125138b8538f04740754"} Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.890798 4766 scope.go:117] "RemoveContainer" containerID="d1c39cb41e9ec9c525837a97a31e77e47f4ff563f15ae600a44c0e8b141327d1" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.890962 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.919938 4766 generic.go:334] "Generic (PLEG): container finished" podID="7cb97caa-8678-4b20-96e7-0ddf811bcc38" containerID="cff0ffc937f2588a8daa1295e1eaefe470b14765cc13965c55dba86981dfc9fc" exitCode=0 Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.920040 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" event={"ID":"7cb97caa-8678-4b20-96e7-0ddf811bcc38","Type":"ContainerDied","Data":"cff0ffc937f2588a8daa1295e1eaefe470b14765cc13965c55dba86981dfc9fc"} Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.946113 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.946154 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae9e2d8d-6ce6-490a-8463-a078626b551f","Type":"ContainerStarted","Data":"3dc581af474a7f996d19f0dd09526f5df82d41a31b74f3887285df3d473ba827"} Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.950624 4766 scope.go:117] "RemoveContainer" containerID="76a3fa02501f1c62ee260c108e730ca16307f8c74245cc6f3b2f7a91e1e162c7" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.964574 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.969417 4766 generic.go:334] "Generic (PLEG): container finished" podID="930724bb-b084-4ae4-8936-ea4ee7b80978" containerID="de8d52ac669ab0babed559fa3fbfb2066c6ee8f12ae83a771a2cc4b1377fa3f2" exitCode=1 Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.969535 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5669dfd6f8-4wb69" event={"ID":"930724bb-b084-4ae4-8936-ea4ee7b80978","Type":"ContainerDied","Data":"de8d52ac669ab0babed559fa3fbfb2066c6ee8f12ae83a771a2cc4b1377fa3f2"} Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.970237 4766 scope.go:117] "RemoveContainer" containerID="de8d52ac669ab0babed559fa3fbfb2066c6ee8f12ae83a771a2cc4b1377fa3f2" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.990764 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 26 00:46:37 crc kubenswrapper[4766]: E1126 00:46:37.991325 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f76ca58-391f-43c9-810b-f80f94213a31" containerName="cinder-api" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.991342 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f76ca58-391f-43c9-810b-f80f94213a31" containerName="cinder-api" Nov 26 00:46:37 crc kubenswrapper[4766]: E1126 00:46:37.991373 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f76ca58-391f-43c9-810b-f80f94213a31" containerName="cinder-api-log" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.991381 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f76ca58-391f-43c9-810b-f80f94213a31" containerName="cinder-api-log" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.991625 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f76ca58-391f-43c9-810b-f80f94213a31" containerName="cinder-api-log" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.991643 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f76ca58-391f-43c9-810b-f80f94213a31" containerName="cinder-api" Nov 26 00:46:37 crc kubenswrapper[4766]: I1126 00:46:37.992999 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.001412 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.001768 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.001892 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.007575 4766 generic.go:334] "Generic (PLEG): container finished" podID="d173ebef-8b7b-46a1-b3b1-d8cd931dff50" containerID="6870abe3616f0bffb71336cff27e044286d42110b5b85c6daaf1d11810dbb9fb" exitCode=1 Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.008954 4766 scope.go:117] "RemoveContainer" containerID="6870abe3616f0bffb71336cff27e044286d42110b5b85c6daaf1d11810dbb9fb" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.009707 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" event={"ID":"d173ebef-8b7b-46a1-b3b1-d8cd931dff50","Type":"ContainerDied","Data":"6870abe3616f0bffb71336cff27e044286d42110b5b85c6daaf1d11810dbb9fb"} Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.010122 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.119665 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.119812 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.119988 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-logs\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.120029 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.120080 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.120141 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-config-data\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.120174 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-828pl\" (UniqueName: \"kubernetes.io/projected/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-kube-api-access-828pl\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.120252 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-config-data-custom\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.123044 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-scripts\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.225960 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.226087 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-logs\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.226121 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.226160 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.226199 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-config-data\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.226230 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-828pl\" (UniqueName: \"kubernetes.io/projected/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-kube-api-access-828pl\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.226278 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-config-data-custom\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.226308 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-scripts\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.226341 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.235452 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.237020 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.237132 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-logs\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.240954 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-config-data\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.252511 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.255158 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-scripts\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.255343 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-config-data-custom\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.255561 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.290819 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-5669dfd6f8-4wb69" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.301810 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-828pl\" (UniqueName: \"kubernetes.io/projected/d4e1503f-0e08-4e3a-a578-fcbd3b51989a-kube-api-access-828pl\") pod \"cinder-api-0\" (UID: \"d4e1503f-0e08-4e3a-a578-fcbd3b51989a\") " pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.311488 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.335346 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.486454 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.489944 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-74d678b987-q89lp" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.506520 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.635362 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-dns-svc\") pod \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.635624 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-ovsdbserver-sb\") pod \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.635671 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-config\") pod \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.635749 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-ovsdbserver-nb\") pod \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.635789 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fwcc\" (UniqueName: \"kubernetes.io/projected/7cb97caa-8678-4b20-96e7-0ddf811bcc38-kube-api-access-5fwcc\") pod \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.635840 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-dns-swift-storage-0\") pod \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\" (UID: \"7cb97caa-8678-4b20-96e7-0ddf811bcc38\") " Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.669698 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cb97caa-8678-4b20-96e7-0ddf811bcc38-kube-api-access-5fwcc" (OuterVolumeSpecName: "kube-api-access-5fwcc") pod "7cb97caa-8678-4b20-96e7-0ddf811bcc38" (UID: "7cb97caa-8678-4b20-96e7-0ddf811bcc38"). InnerVolumeSpecName "kube-api-access-5fwcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.716595 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-config" (OuterVolumeSpecName: "config") pod "7cb97caa-8678-4b20-96e7-0ddf811bcc38" (UID: "7cb97caa-8678-4b20-96e7-0ddf811bcc38"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.738085 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.738114 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fwcc\" (UniqueName: \"kubernetes.io/projected/7cb97caa-8678-4b20-96e7-0ddf811bcc38-kube-api-access-5fwcc\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.745799 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7cb97caa-8678-4b20-96e7-0ddf811bcc38" (UID: "7cb97caa-8678-4b20-96e7-0ddf811bcc38"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.754677 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7cb97caa-8678-4b20-96e7-0ddf811bcc38" (UID: "7cb97caa-8678-4b20-96e7-0ddf811bcc38"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.764159 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7cb97caa-8678-4b20-96e7-0ddf811bcc38" (UID: "7cb97caa-8678-4b20-96e7-0ddf811bcc38"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.791405 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7cb97caa-8678-4b20-96e7-0ddf811bcc38" (UID: "7cb97caa-8678-4b20-96e7-0ddf811bcc38"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.843760 4766 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.843799 4766 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.843813 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:38 crc kubenswrapper[4766]: I1126 00:46:38.843825 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cb97caa-8678-4b20-96e7-0ddf811bcc38-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:39 crc kubenswrapper[4766]: I1126 00:46:39.024359 4766 generic.go:334] "Generic (PLEG): container finished" podID="d173ebef-8b7b-46a1-b3b1-d8cd931dff50" containerID="bf97e78b50358aaf0360b54f2d3c39b4dbde8c92662f49da6e2f0f9acf4484a0" exitCode=1 Nov 26 00:46:39 crc kubenswrapper[4766]: I1126 00:46:39.024535 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" event={"ID":"d173ebef-8b7b-46a1-b3b1-d8cd931dff50","Type":"ContainerDied","Data":"bf97e78b50358aaf0360b54f2d3c39b4dbde8c92662f49da6e2f0f9acf4484a0"} Nov 26 00:46:39 crc kubenswrapper[4766]: I1126 00:46:39.024794 4766 scope.go:117] "RemoveContainer" containerID="6870abe3616f0bffb71336cff27e044286d42110b5b85c6daaf1d11810dbb9fb" Nov 26 00:46:39 crc kubenswrapper[4766]: I1126 00:46:39.025219 4766 scope.go:117] "RemoveContainer" containerID="bf97e78b50358aaf0360b54f2d3c39b4dbde8c92662f49da6e2f0f9acf4484a0" Nov 26 00:46:39 crc kubenswrapper[4766]: E1126 00:46:39.025444 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-6f6bc96c7-x82w9_openstack(d173ebef-8b7b-46a1-b3b1-d8cd931dff50)\"" pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" podUID="d173ebef-8b7b-46a1-b3b1-d8cd931dff50" Nov 26 00:46:39 crc kubenswrapper[4766]: I1126 00:46:39.043942 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" Nov 26 00:46:39 crc kubenswrapper[4766]: I1126 00:46:39.043937 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" event={"ID":"7cb97caa-8678-4b20-96e7-0ddf811bcc38","Type":"ContainerDied","Data":"9298e63d560510a6bd5d7f9cb76b9fb9d4403eb49c2c756321f02de08ed029ef"} Nov 26 00:46:39 crc kubenswrapper[4766]: I1126 00:46:39.079012 4766 generic.go:334] "Generic (PLEG): container finished" podID="930724bb-b084-4ae4-8936-ea4ee7b80978" containerID="e74e130e406c81eebd0655f2dbb802bd0f3fd1694944061c781712f8d9f17215" exitCode=1 Nov 26 00:46:39 crc kubenswrapper[4766]: I1126 00:46:39.082516 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5669dfd6f8-4wb69" event={"ID":"930724bb-b084-4ae4-8936-ea4ee7b80978","Type":"ContainerDied","Data":"e74e130e406c81eebd0655f2dbb802bd0f3fd1694944061c781712f8d9f17215"} Nov 26 00:46:39 crc kubenswrapper[4766]: I1126 00:46:39.082532 4766 scope.go:117] "RemoveContainer" containerID="e74e130e406c81eebd0655f2dbb802bd0f3fd1694944061c781712f8d9f17215" Nov 26 00:46:39 crc kubenswrapper[4766]: E1126 00:46:39.099425 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-5669dfd6f8-4wb69_openstack(930724bb-b084-4ae4-8936-ea4ee7b80978)\"" pod="openstack/heat-api-5669dfd6f8-4wb69" podUID="930724bb-b084-4ae4-8936-ea4ee7b80978" Nov 26 00:46:39 crc kubenswrapper[4766]: I1126 00:46:39.145259 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 26 00:46:39 crc kubenswrapper[4766]: I1126 00:46:39.203426 4766 scope.go:117] "RemoveContainer" containerID="cff0ffc937f2588a8daa1295e1eaefe470b14765cc13965c55dba86981dfc9fc" Nov 26 00:46:39 crc kubenswrapper[4766]: I1126 00:46:39.230707 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-dpnb6"] Nov 26 00:46:39 crc kubenswrapper[4766]: I1126 00:46:39.237791 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-dpnb6"] Nov 26 00:46:39 crc kubenswrapper[4766]: I1126 00:46:39.258794 4766 scope.go:117] "RemoveContainer" containerID="9a6723f51bd3ee62f15560e4742b57b6c094e8ef945f848ce06b0cdab422ef9f" Nov 26 00:46:39 crc kubenswrapper[4766]: I1126 00:46:39.287140 4766 scope.go:117] "RemoveContainer" containerID="de8d52ac669ab0babed559fa3fbfb2066c6ee8f12ae83a771a2cc4b1377fa3f2" Nov 26 00:46:39 crc kubenswrapper[4766]: I1126 00:46:39.845182 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f76ca58-391f-43c9-810b-f80f94213a31" path="/var/lib/kubelet/pods/0f76ca58-391f-43c9-810b-f80f94213a31/volumes" Nov 26 00:46:39 crc kubenswrapper[4766]: I1126 00:46:39.846213 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cb97caa-8678-4b20-96e7-0ddf811bcc38" path="/var/lib/kubelet/pods/7cb97caa-8678-4b20-96e7-0ddf811bcc38/volumes" Nov 26 00:46:40 crc kubenswrapper[4766]: I1126 00:46:40.095924 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae9e2d8d-6ce6-490a-8463-a078626b551f","Type":"ContainerStarted","Data":"84145d2dc6ed5f8fb4ea3ecc5fc4fadaa23dea0252b4b50ff8d1361c0a888d4f"} Nov 26 00:46:40 crc kubenswrapper[4766]: I1126 00:46:40.096075 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae9e2d8d-6ce6-490a-8463-a078626b551f" containerName="ceilometer-central-agent" containerID="cri-o://c55670df27b91fd6586d9d88a8f46c9f9ff1651c0f04ee65aa18cbb449d8df61" gracePeriod=30 Nov 26 00:46:40 crc kubenswrapper[4766]: I1126 00:46:40.096636 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 26 00:46:40 crc kubenswrapper[4766]: I1126 00:46:40.096920 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae9e2d8d-6ce6-490a-8463-a078626b551f" containerName="proxy-httpd" containerID="cri-o://84145d2dc6ed5f8fb4ea3ecc5fc4fadaa23dea0252b4b50ff8d1361c0a888d4f" gracePeriod=30 Nov 26 00:46:40 crc kubenswrapper[4766]: I1126 00:46:40.096965 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae9e2d8d-6ce6-490a-8463-a078626b551f" containerName="sg-core" containerID="cri-o://3dc581af474a7f996d19f0dd09526f5df82d41a31b74f3887285df3d473ba827" gracePeriod=30 Nov 26 00:46:40 crc kubenswrapper[4766]: I1126 00:46:40.096996 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae9e2d8d-6ce6-490a-8463-a078626b551f" containerName="ceilometer-notification-agent" containerID="cri-o://11e1c320e18b1ba9311a6168bb1564ba38aaaf3a89e6ed61fa800644fa78f7b1" gracePeriod=30 Nov 26 00:46:40 crc kubenswrapper[4766]: I1126 00:46:40.100534 4766 scope.go:117] "RemoveContainer" containerID="e74e130e406c81eebd0655f2dbb802bd0f3fd1694944061c781712f8d9f17215" Nov 26 00:46:40 crc kubenswrapper[4766]: E1126 00:46:40.100748 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-5669dfd6f8-4wb69_openstack(930724bb-b084-4ae4-8936-ea4ee7b80978)\"" pod="openstack/heat-api-5669dfd6f8-4wb69" podUID="930724bb-b084-4ae4-8936-ea4ee7b80978" Nov 26 00:46:40 crc kubenswrapper[4766]: I1126 00:46:40.103770 4766 scope.go:117] "RemoveContainer" containerID="bf97e78b50358aaf0360b54f2d3c39b4dbde8c92662f49da6e2f0f9acf4484a0" Nov 26 00:46:40 crc kubenswrapper[4766]: E1126 00:46:40.104065 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-6f6bc96c7-x82w9_openstack(d173ebef-8b7b-46a1-b3b1-d8cd931dff50)\"" pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" podUID="d173ebef-8b7b-46a1-b3b1-d8cd931dff50" Nov 26 00:46:40 crc kubenswrapper[4766]: I1126 00:46:40.105824 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d4e1503f-0e08-4e3a-a578-fcbd3b51989a","Type":"ContainerStarted","Data":"21c66060e605a0d443525083ce564d4f4a58b1b806a864a3bfae6c82879e1a57"} Nov 26 00:46:40 crc kubenswrapper[4766]: I1126 00:46:40.105885 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d4e1503f-0e08-4e3a-a578-fcbd3b51989a","Type":"ContainerStarted","Data":"7aa81e978dbfbd7a76613bb0646540ba8e9a4a2c100c1cc6a31e7c599a8649c6"} Nov 26 00:46:40 crc kubenswrapper[4766]: I1126 00:46:40.125494 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.8749228430000002 podStartE2EDuration="8.125332648s" podCreationTimestamp="2025-11-26 00:46:32 +0000 UTC" firstStartedPulling="2025-11-26 00:46:33.782962587 +0000 UTC m=+1374.631733017" lastFinishedPulling="2025-11-26 00:46:39.033372392 +0000 UTC m=+1379.882142822" observedRunningTime="2025-11-26 00:46:40.118144177 +0000 UTC m=+1380.966914617" watchObservedRunningTime="2025-11-26 00:46:40.125332648 +0000 UTC m=+1380.974103078" Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.115190 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d4e1503f-0e08-4e3a-a578-fcbd3b51989a","Type":"ContainerStarted","Data":"52d5c3f3cae431474324e348c62b13b6dfd802d2806f75f477697c367bcbe818"} Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.115789 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.119546 4766 generic.go:334] "Generic (PLEG): container finished" podID="ae9e2d8d-6ce6-490a-8463-a078626b551f" containerID="84145d2dc6ed5f8fb4ea3ecc5fc4fadaa23dea0252b4b50ff8d1361c0a888d4f" exitCode=0 Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.119579 4766 generic.go:334] "Generic (PLEG): container finished" podID="ae9e2d8d-6ce6-490a-8463-a078626b551f" containerID="3dc581af474a7f996d19f0dd09526f5df82d41a31b74f3887285df3d473ba827" exitCode=2 Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.119587 4766 generic.go:334] "Generic (PLEG): container finished" podID="ae9e2d8d-6ce6-490a-8463-a078626b551f" containerID="11e1c320e18b1ba9311a6168bb1564ba38aaaf3a89e6ed61fa800644fa78f7b1" exitCode=0 Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.119608 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae9e2d8d-6ce6-490a-8463-a078626b551f","Type":"ContainerDied","Data":"84145d2dc6ed5f8fb4ea3ecc5fc4fadaa23dea0252b4b50ff8d1361c0a888d4f"} Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.119633 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae9e2d8d-6ce6-490a-8463-a078626b551f","Type":"ContainerDied","Data":"3dc581af474a7f996d19f0dd09526f5df82d41a31b74f3887285df3d473ba827"} Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.119642 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae9e2d8d-6ce6-490a-8463-a078626b551f","Type":"ContainerDied","Data":"11e1c320e18b1ba9311a6168bb1564ba38aaaf3a89e6ed61fa800644fa78f7b1"} Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.151087 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.151067953 podStartE2EDuration="4.151067953s" podCreationTimestamp="2025-11-26 00:46:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:46:41.139994715 +0000 UTC m=+1381.988765145" watchObservedRunningTime="2025-11-26 00:46:41.151067953 +0000 UTC m=+1381.999838383" Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.697891 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 26 00:46:41 crc kubenswrapper[4766]: E1126 00:46:41.700377 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cb97caa-8678-4b20-96e7-0ddf811bcc38" containerName="dnsmasq-dns" Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.700474 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cb97caa-8678-4b20-96e7-0ddf811bcc38" containerName="dnsmasq-dns" Nov 26 00:46:41 crc kubenswrapper[4766]: E1126 00:46:41.700540 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cb97caa-8678-4b20-96e7-0ddf811bcc38" containerName="init" Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.700593 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cb97caa-8678-4b20-96e7-0ddf811bcc38" containerName="init" Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.700906 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cb97caa-8678-4b20-96e7-0ddf811bcc38" containerName="dnsmasq-dns" Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.701742 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.708744 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.708744 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.720304 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.815435 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/61b7dc74-d057-4254-a56d-dafecc2515c5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"61b7dc74-d057-4254-a56d-dafecc2515c5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.815701 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/61b7dc74-d057-4254-a56d-dafecc2515c5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"61b7dc74-d057-4254-a56d-dafecc2515c5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.917839 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/61b7dc74-d057-4254-a56d-dafecc2515c5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"61b7dc74-d057-4254-a56d-dafecc2515c5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.917951 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/61b7dc74-d057-4254-a56d-dafecc2515c5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"61b7dc74-d057-4254-a56d-dafecc2515c5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.918460 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/61b7dc74-d057-4254-a56d-dafecc2515c5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"61b7dc74-d057-4254-a56d-dafecc2515c5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 00:46:41 crc kubenswrapper[4766]: I1126 00:46:41.945517 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/61b7dc74-d057-4254-a56d-dafecc2515c5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"61b7dc74-d057-4254-a56d-dafecc2515c5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 00:46:42 crc kubenswrapper[4766]: I1126 00:46:42.024464 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 00:46:42 crc kubenswrapper[4766]: I1126 00:46:42.592155 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 26 00:46:42 crc kubenswrapper[4766]: I1126 00:46:42.708255 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6578955fd5-dpnb6" podUID="7cb97caa-8678-4b20-96e7-0ddf811bcc38" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.189:5353: i/o timeout" Nov 26 00:46:42 crc kubenswrapper[4766]: I1126 00:46:42.870385 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:46:42 crc kubenswrapper[4766]: I1126 00:46:42.874751 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:46:42 crc kubenswrapper[4766]: I1126 00:46:42.952803 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-6f6bc96c7-x82w9"] Nov 26 00:46:42 crc kubenswrapper[4766]: I1126 00:46:42.991633 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5669dfd6f8-4wb69"] Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.171671 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"61b7dc74-d057-4254-a56d-dafecc2515c5","Type":"ContainerStarted","Data":"003747c3d811e95d8e5956a6b4965626bc75f2024780f554d589502235dbfb7e"} Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.171720 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"61b7dc74-d057-4254-a56d-dafecc2515c5","Type":"ContainerStarted","Data":"56b61ef83a8dbbd6a64f18e53add9f783ef31a8375b28e64f2dafae81e272532"} Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.193630 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=2.193606848 podStartE2EDuration="2.193606848s" podCreationTimestamp="2025-11-26 00:46:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:46:43.190764136 +0000 UTC m=+1384.039534566" watchObservedRunningTime="2025-11-26 00:46:43.193606848 +0000 UTC m=+1384.042377278" Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.293264 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5669dfd6f8-4wb69" Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.309775 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.727554 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.744020 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5669dfd6f8-4wb69" Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.879418 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/930724bb-b084-4ae4-8936-ea4ee7b80978-config-data-custom\") pod \"930724bb-b084-4ae4-8936-ea4ee7b80978\" (UID: \"930724bb-b084-4ae4-8936-ea4ee7b80978\") " Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.880170 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/930724bb-b084-4ae4-8936-ea4ee7b80978-combined-ca-bundle\") pod \"930724bb-b084-4ae4-8936-ea4ee7b80978\" (UID: \"930724bb-b084-4ae4-8936-ea4ee7b80978\") " Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.880409 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bm5xl\" (UniqueName: \"kubernetes.io/projected/930724bb-b084-4ae4-8936-ea4ee7b80978-kube-api-access-bm5xl\") pod \"930724bb-b084-4ae4-8936-ea4ee7b80978\" (UID: \"930724bb-b084-4ae4-8936-ea4ee7b80978\") " Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.880437 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-combined-ca-bundle\") pod \"d173ebef-8b7b-46a1-b3b1-d8cd931dff50\" (UID: \"d173ebef-8b7b-46a1-b3b1-d8cd931dff50\") " Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.880477 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-config-data\") pod \"d173ebef-8b7b-46a1-b3b1-d8cd931dff50\" (UID: \"d173ebef-8b7b-46a1-b3b1-d8cd931dff50\") " Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.880538 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/930724bb-b084-4ae4-8936-ea4ee7b80978-config-data\") pod \"930724bb-b084-4ae4-8936-ea4ee7b80978\" (UID: \"930724bb-b084-4ae4-8936-ea4ee7b80978\") " Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.880569 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-config-data-custom\") pod \"d173ebef-8b7b-46a1-b3b1-d8cd931dff50\" (UID: \"d173ebef-8b7b-46a1-b3b1-d8cd931dff50\") " Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.880838 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8bk7\" (UniqueName: \"kubernetes.io/projected/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-kube-api-access-d8bk7\") pod \"d173ebef-8b7b-46a1-b3b1-d8cd931dff50\" (UID: \"d173ebef-8b7b-46a1-b3b1-d8cd931dff50\") " Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.908887 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/930724bb-b084-4ae4-8936-ea4ee7b80978-kube-api-access-bm5xl" (OuterVolumeSpecName: "kube-api-access-bm5xl") pod "930724bb-b084-4ae4-8936-ea4ee7b80978" (UID: "930724bb-b084-4ae4-8936-ea4ee7b80978"). InnerVolumeSpecName "kube-api-access-bm5xl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.908980 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d173ebef-8b7b-46a1-b3b1-d8cd931dff50" (UID: "d173ebef-8b7b-46a1-b3b1-d8cd931dff50"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.910901 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/930724bb-b084-4ae4-8936-ea4ee7b80978-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "930724bb-b084-4ae4-8936-ea4ee7b80978" (UID: "930724bb-b084-4ae4-8936-ea4ee7b80978"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.912066 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-kube-api-access-d8bk7" (OuterVolumeSpecName: "kube-api-access-d8bk7") pod "d173ebef-8b7b-46a1-b3b1-d8cd931dff50" (UID: "d173ebef-8b7b-46a1-b3b1-d8cd931dff50"). InnerVolumeSpecName "kube-api-access-d8bk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.928803 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d173ebef-8b7b-46a1-b3b1-d8cd931dff50" (UID: "d173ebef-8b7b-46a1-b3b1-d8cd931dff50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.944794 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/930724bb-b084-4ae4-8936-ea4ee7b80978-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "930724bb-b084-4ae4-8936-ea4ee7b80978" (UID: "930724bb-b084-4ae4-8936-ea4ee7b80978"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.956475 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-config-data" (OuterVolumeSpecName: "config-data") pod "d173ebef-8b7b-46a1-b3b1-d8cd931dff50" (UID: "d173ebef-8b7b-46a1-b3b1-d8cd931dff50"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.964337 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/930724bb-b084-4ae4-8936-ea4ee7b80978-config-data" (OuterVolumeSpecName: "config-data") pod "930724bb-b084-4ae4-8936-ea4ee7b80978" (UID: "930724bb-b084-4ae4-8936-ea4ee7b80978"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.988307 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8bk7\" (UniqueName: \"kubernetes.io/projected/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-kube-api-access-d8bk7\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.988343 4766 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/930724bb-b084-4ae4-8936-ea4ee7b80978-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.988352 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/930724bb-b084-4ae4-8936-ea4ee7b80978-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.988365 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bm5xl\" (UniqueName: \"kubernetes.io/projected/930724bb-b084-4ae4-8936-ea4ee7b80978-kube-api-access-bm5xl\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.988374 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.988397 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.988406 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/930724bb-b084-4ae4-8936-ea4ee7b80978-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:43 crc kubenswrapper[4766]: I1126 00:46:43.988414 4766 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d173ebef-8b7b-46a1-b3b1-d8cd931dff50-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:44 crc kubenswrapper[4766]: I1126 00:46:44.182973 4766 generic.go:334] "Generic (PLEG): container finished" podID="61b7dc74-d057-4254-a56d-dafecc2515c5" containerID="003747c3d811e95d8e5956a6b4965626bc75f2024780f554d589502235dbfb7e" exitCode=0 Nov 26 00:46:44 crc kubenswrapper[4766]: I1126 00:46:44.183030 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"61b7dc74-d057-4254-a56d-dafecc2515c5","Type":"ContainerDied","Data":"003747c3d811e95d8e5956a6b4965626bc75f2024780f554d589502235dbfb7e"} Nov 26 00:46:44 crc kubenswrapper[4766]: I1126 00:46:44.187038 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5669dfd6f8-4wb69" event={"ID":"930724bb-b084-4ae4-8936-ea4ee7b80978","Type":"ContainerDied","Data":"bb9302cf6e621dc4dda4542b7166e9fca1f584bda054f5e7ed21e34babb64c36"} Nov 26 00:46:44 crc kubenswrapper[4766]: I1126 00:46:44.187085 4766 scope.go:117] "RemoveContainer" containerID="e74e130e406c81eebd0655f2dbb802bd0f3fd1694944061c781712f8d9f17215" Nov 26 00:46:44 crc kubenswrapper[4766]: I1126 00:46:44.187175 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5669dfd6f8-4wb69" Nov 26 00:46:44 crc kubenswrapper[4766]: I1126 00:46:44.192087 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" event={"ID":"d173ebef-8b7b-46a1-b3b1-d8cd931dff50","Type":"ContainerDied","Data":"dbb8ce7c262109558e2316ab31cec4d1bc1d8d32d473d05609d14b3cce6b4662"} Nov 26 00:46:44 crc kubenswrapper[4766]: I1126 00:46:44.192175 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6f6bc96c7-x82w9" Nov 26 00:46:44 crc kubenswrapper[4766]: I1126 00:46:44.225240 4766 scope.go:117] "RemoveContainer" containerID="bf97e78b50358aaf0360b54f2d3c39b4dbde8c92662f49da6e2f0f9acf4484a0" Nov 26 00:46:44 crc kubenswrapper[4766]: I1126 00:46:44.285039 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" Nov 26 00:46:44 crc kubenswrapper[4766]: I1126 00:46:44.306278 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-6f6bc96c7-x82w9"] Nov 26 00:46:44 crc kubenswrapper[4766]: I1126 00:46:44.317995 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-6f6bc96c7-x82w9"] Nov 26 00:46:44 crc kubenswrapper[4766]: I1126 00:46:44.335781 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5669dfd6f8-4wb69"] Nov 26 00:46:44 crc kubenswrapper[4766]: I1126 00:46:44.339550 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-856c5c6f86-fk66f" Nov 26 00:46:44 crc kubenswrapper[4766]: I1126 00:46:44.345909 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-5669dfd6f8-4wb69"] Nov 26 00:46:45 crc kubenswrapper[4766]: I1126 00:46:45.666482 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 00:46:45 crc kubenswrapper[4766]: I1126 00:46:45.723421 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/61b7dc74-d057-4254-a56d-dafecc2515c5-kube-api-access\") pod \"61b7dc74-d057-4254-a56d-dafecc2515c5\" (UID: \"61b7dc74-d057-4254-a56d-dafecc2515c5\") " Nov 26 00:46:45 crc kubenswrapper[4766]: I1126 00:46:45.723767 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/61b7dc74-d057-4254-a56d-dafecc2515c5-kubelet-dir\") pod \"61b7dc74-d057-4254-a56d-dafecc2515c5\" (UID: \"61b7dc74-d057-4254-a56d-dafecc2515c5\") " Nov 26 00:46:45 crc kubenswrapper[4766]: I1126 00:46:45.723864 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61b7dc74-d057-4254-a56d-dafecc2515c5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "61b7dc74-d057-4254-a56d-dafecc2515c5" (UID: "61b7dc74-d057-4254-a56d-dafecc2515c5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:46:45 crc kubenswrapper[4766]: I1126 00:46:45.724487 4766 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/61b7dc74-d057-4254-a56d-dafecc2515c5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:45 crc kubenswrapper[4766]: I1126 00:46:45.729564 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61b7dc74-d057-4254-a56d-dafecc2515c5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "61b7dc74-d057-4254-a56d-dafecc2515c5" (UID: "61b7dc74-d057-4254-a56d-dafecc2515c5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:46:45 crc kubenswrapper[4766]: I1126 00:46:45.825890 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/61b7dc74-d057-4254-a56d-dafecc2515c5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:45 crc kubenswrapper[4766]: I1126 00:46:45.837601 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="930724bb-b084-4ae4-8936-ea4ee7b80978" path="/var/lib/kubelet/pods/930724bb-b084-4ae4-8936-ea4ee7b80978/volumes" Nov 26 00:46:45 crc kubenswrapper[4766]: I1126 00:46:45.838170 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d173ebef-8b7b-46a1-b3b1-d8cd931dff50" path="/var/lib/kubelet/pods/d173ebef-8b7b-46a1-b3b1-d8cd931dff50/volumes" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.213196 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"61b7dc74-d057-4254-a56d-dafecc2515c5","Type":"ContainerDied","Data":"56b61ef83a8dbbd6a64f18e53add9f783ef31a8375b28e64f2dafae81e272532"} Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.213230 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.213241 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56b61ef83a8dbbd6a64f18e53add9f783ef31a8375b28e64f2dafae81e272532" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.694675 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 26 00:46:46 crc kubenswrapper[4766]: E1126 00:46:46.696002 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="930724bb-b084-4ae4-8936-ea4ee7b80978" containerName="heat-api" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.696099 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="930724bb-b084-4ae4-8936-ea4ee7b80978" containerName="heat-api" Nov 26 00:46:46 crc kubenswrapper[4766]: E1126 00:46:46.696162 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d173ebef-8b7b-46a1-b3b1-d8cd931dff50" containerName="heat-cfnapi" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.696223 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d173ebef-8b7b-46a1-b3b1-d8cd931dff50" containerName="heat-cfnapi" Nov 26 00:46:46 crc kubenswrapper[4766]: E1126 00:46:46.696287 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61b7dc74-d057-4254-a56d-dafecc2515c5" containerName="pruner" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.696343 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="61b7dc74-d057-4254-a56d-dafecc2515c5" containerName="pruner" Nov 26 00:46:46 crc kubenswrapper[4766]: E1126 00:46:46.696418 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d173ebef-8b7b-46a1-b3b1-d8cd931dff50" containerName="heat-cfnapi" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.696486 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d173ebef-8b7b-46a1-b3b1-d8cd931dff50" containerName="heat-cfnapi" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.696759 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="61b7dc74-d057-4254-a56d-dafecc2515c5" containerName="pruner" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.696840 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="d173ebef-8b7b-46a1-b3b1-d8cd931dff50" containerName="heat-cfnapi" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.696910 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="d173ebef-8b7b-46a1-b3b1-d8cd931dff50" containerName="heat-cfnapi" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.696983 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="930724bb-b084-4ae4-8936-ea4ee7b80978" containerName="heat-api" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.697937 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.700993 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.705046 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.707879 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.742718 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2bba0b1b-4848-4452-a535-c0cf5d13b30b-kube-api-access\") pod \"installer-9-crc\" (UID: \"2bba0b1b-4848-4452-a535-c0cf5d13b30b\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.742801 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2bba0b1b-4848-4452-a535-c0cf5d13b30b-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2bba0b1b-4848-4452-a535-c0cf5d13b30b\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.742831 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2bba0b1b-4848-4452-a535-c0cf5d13b30b-var-lock\") pod \"installer-9-crc\" (UID: \"2bba0b1b-4848-4452-a535-c0cf5d13b30b\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.844915 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2bba0b1b-4848-4452-a535-c0cf5d13b30b-kube-api-access\") pod \"installer-9-crc\" (UID: \"2bba0b1b-4848-4452-a535-c0cf5d13b30b\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.845498 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2bba0b1b-4848-4452-a535-c0cf5d13b30b-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2bba0b1b-4848-4452-a535-c0cf5d13b30b\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.845613 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2bba0b1b-4848-4452-a535-c0cf5d13b30b-var-lock\") pod \"installer-9-crc\" (UID: \"2bba0b1b-4848-4452-a535-c0cf5d13b30b\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.845788 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2bba0b1b-4848-4452-a535-c0cf5d13b30b-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2bba0b1b-4848-4452-a535-c0cf5d13b30b\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.846674 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2bba0b1b-4848-4452-a535-c0cf5d13b30b-var-lock\") pod \"installer-9-crc\" (UID: \"2bba0b1b-4848-4452-a535-c0cf5d13b30b\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 00:46:46 crc kubenswrapper[4766]: I1126 00:46:46.866358 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2bba0b1b-4848-4452-a535-c0cf5d13b30b-kube-api-access\") pod \"installer-9-crc\" (UID: \"2bba0b1b-4848-4452-a535-c0cf5d13b30b\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.028709 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.059956 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-k2tlc"] Nov 26 00:46:47 crc kubenswrapper[4766]: E1126 00:46:47.060452 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="930724bb-b084-4ae4-8936-ea4ee7b80978" containerName="heat-api" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.060472 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="930724bb-b084-4ae4-8936-ea4ee7b80978" containerName="heat-api" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.060697 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="930724bb-b084-4ae4-8936-ea4ee7b80978" containerName="heat-api" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.061391 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-k2tlc" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.062009 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.073690 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-k2tlc"] Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.197717 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtnln\" (UniqueName: \"kubernetes.io/projected/619650be-59d2-47c8-815d-3337c156c83f-kube-api-access-mtnln\") pod \"nova-api-db-create-k2tlc\" (UID: \"619650be-59d2-47c8-815d-3337c156c83f\") " pod="openstack/nova-api-db-create-k2tlc" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.197913 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/619650be-59d2-47c8-815d-3337c156c83f-operator-scripts\") pod \"nova-api-db-create-k2tlc\" (UID: \"619650be-59d2-47c8-815d-3337c156c83f\") " pod="openstack/nova-api-db-create-k2tlc" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.234231 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-t4vp6"] Nov 26 00:46:47 crc kubenswrapper[4766]: E1126 00:46:47.234707 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae9e2d8d-6ce6-490a-8463-a078626b551f" containerName="ceilometer-central-agent" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.234720 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae9e2d8d-6ce6-490a-8463-a078626b551f" containerName="ceilometer-central-agent" Nov 26 00:46:47 crc kubenswrapper[4766]: E1126 00:46:47.234739 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae9e2d8d-6ce6-490a-8463-a078626b551f" containerName="proxy-httpd" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.234746 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae9e2d8d-6ce6-490a-8463-a078626b551f" containerName="proxy-httpd" Nov 26 00:46:47 crc kubenswrapper[4766]: E1126 00:46:47.234758 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae9e2d8d-6ce6-490a-8463-a078626b551f" containerName="sg-core" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.234765 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae9e2d8d-6ce6-490a-8463-a078626b551f" containerName="sg-core" Nov 26 00:46:47 crc kubenswrapper[4766]: E1126 00:46:47.234774 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae9e2d8d-6ce6-490a-8463-a078626b551f" containerName="ceilometer-notification-agent" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.234780 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae9e2d8d-6ce6-490a-8463-a078626b551f" containerName="ceilometer-notification-agent" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.234992 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae9e2d8d-6ce6-490a-8463-a078626b551f" containerName="proxy-httpd" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.235005 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae9e2d8d-6ce6-490a-8463-a078626b551f" containerName="ceilometer-notification-agent" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.235014 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae9e2d8d-6ce6-490a-8463-a078626b551f" containerName="sg-core" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.235033 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae9e2d8d-6ce6-490a-8463-a078626b551f" containerName="ceilometer-central-agent" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.242728 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-t4vp6" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.261340 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-t4vp6"] Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.291263 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-88b3-account-create-update-zvx6c"] Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.293343 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-88b3-account-create-update-zvx6c" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.295843 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.299201 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27ft2\" (UniqueName: \"kubernetes.io/projected/ae9e2d8d-6ce6-490a-8463-a078626b551f-kube-api-access-27ft2\") pod \"ae9e2d8d-6ce6-490a-8463-a078626b551f\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.299744 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae9e2d8d-6ce6-490a-8463-a078626b551f-run-httpd\") pod \"ae9e2d8d-6ce6-490a-8463-a078626b551f\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.299952 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-config-data\") pod \"ae9e2d8d-6ce6-490a-8463-a078626b551f\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.300054 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-combined-ca-bundle\") pod \"ae9e2d8d-6ce6-490a-8463-a078626b551f\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.300170 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-sg-core-conf-yaml\") pod \"ae9e2d8d-6ce6-490a-8463-a078626b551f\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.300278 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae9e2d8d-6ce6-490a-8463-a078626b551f-log-httpd\") pod \"ae9e2d8d-6ce6-490a-8463-a078626b551f\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.300421 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-scripts\") pod \"ae9e2d8d-6ce6-490a-8463-a078626b551f\" (UID: \"ae9e2d8d-6ce6-490a-8463-a078626b551f\") " Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.301295 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fzmx\" (UniqueName: \"kubernetes.io/projected/2046b547-611f-4e98-ba95-a34cd0d45f50-kube-api-access-9fzmx\") pod \"nova-cell0-db-create-t4vp6\" (UID: \"2046b547-611f-4e98-ba95-a34cd0d45f50\") " pod="openstack/nova-cell0-db-create-t4vp6" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.301398 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/751e5ae1-be9d-48e8-8cc1-e08012854203-operator-scripts\") pod \"nova-api-88b3-account-create-update-zvx6c\" (UID: \"751e5ae1-be9d-48e8-8cc1-e08012854203\") " pod="openstack/nova-api-88b3-account-create-update-zvx6c" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.301530 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ml5d\" (UniqueName: \"kubernetes.io/projected/751e5ae1-be9d-48e8-8cc1-e08012854203-kube-api-access-6ml5d\") pod \"nova-api-88b3-account-create-update-zvx6c\" (UID: \"751e5ae1-be9d-48e8-8cc1-e08012854203\") " pod="openstack/nova-api-88b3-account-create-update-zvx6c" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.301671 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtnln\" (UniqueName: \"kubernetes.io/projected/619650be-59d2-47c8-815d-3337c156c83f-kube-api-access-mtnln\") pod \"nova-api-db-create-k2tlc\" (UID: \"619650be-59d2-47c8-815d-3337c156c83f\") " pod="openstack/nova-api-db-create-k2tlc" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.301841 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2046b547-611f-4e98-ba95-a34cd0d45f50-operator-scripts\") pod \"nova-cell0-db-create-t4vp6\" (UID: \"2046b547-611f-4e98-ba95-a34cd0d45f50\") " pod="openstack/nova-cell0-db-create-t4vp6" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.301957 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/619650be-59d2-47c8-815d-3337c156c83f-operator-scripts\") pod \"nova-api-db-create-k2tlc\" (UID: \"619650be-59d2-47c8-815d-3337c156c83f\") " pod="openstack/nova-api-db-create-k2tlc" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.302742 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/619650be-59d2-47c8-815d-3337c156c83f-operator-scripts\") pod \"nova-api-db-create-k2tlc\" (UID: \"619650be-59d2-47c8-815d-3337c156c83f\") " pod="openstack/nova-api-db-create-k2tlc" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.303135 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae9e2d8d-6ce6-490a-8463-a078626b551f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ae9e2d8d-6ce6-490a-8463-a078626b551f" (UID: "ae9e2d8d-6ce6-490a-8463-a078626b551f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.304697 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae9e2d8d-6ce6-490a-8463-a078626b551f-kube-api-access-27ft2" (OuterVolumeSpecName: "kube-api-access-27ft2") pod "ae9e2d8d-6ce6-490a-8463-a078626b551f" (UID: "ae9e2d8d-6ce6-490a-8463-a078626b551f"). InnerVolumeSpecName "kube-api-access-27ft2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.304926 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae9e2d8d-6ce6-490a-8463-a078626b551f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ae9e2d8d-6ce6-490a-8463-a078626b551f" (UID: "ae9e2d8d-6ce6-490a-8463-a078626b551f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.343854 4766 generic.go:334] "Generic (PLEG): container finished" podID="ae9e2d8d-6ce6-490a-8463-a078626b551f" containerID="c55670df27b91fd6586d9d88a8f46c9f9ff1651c0f04ee65aa18cbb449d8df61" exitCode=0 Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.343905 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae9e2d8d-6ce6-490a-8463-a078626b551f","Type":"ContainerDied","Data":"c55670df27b91fd6586d9d88a8f46c9f9ff1651c0f04ee65aa18cbb449d8df61"} Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.343940 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae9e2d8d-6ce6-490a-8463-a078626b551f","Type":"ContainerDied","Data":"47739c076815090533bc3fe60cf248c55e509bb6013b7a28d69fb1e9c273b18f"} Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.343962 4766 scope.go:117] "RemoveContainer" containerID="84145d2dc6ed5f8fb4ea3ecc5fc4fadaa23dea0252b4b50ff8d1361c0a888d4f" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.344175 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.353357 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-scripts" (OuterVolumeSpecName: "scripts") pod "ae9e2d8d-6ce6-490a-8463-a078626b551f" (UID: "ae9e2d8d-6ce6-490a-8463-a078626b551f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.358140 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtnln\" (UniqueName: \"kubernetes.io/projected/619650be-59d2-47c8-815d-3337c156c83f-kube-api-access-mtnln\") pod \"nova-api-db-create-k2tlc\" (UID: \"619650be-59d2-47c8-815d-3337c156c83f\") " pod="openstack/nova-api-db-create-k2tlc" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.367778 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ae9e2d8d-6ce6-490a-8463-a078626b551f" (UID: "ae9e2d8d-6ce6-490a-8463-a078626b551f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.379784 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-88b3-account-create-update-zvx6c"] Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.382329 4766 scope.go:117] "RemoveContainer" containerID="3dc581af474a7f996d19f0dd09526f5df82d41a31b74f3887285df3d473ba827" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.404097 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2046b547-611f-4e98-ba95-a34cd0d45f50-operator-scripts\") pod \"nova-cell0-db-create-t4vp6\" (UID: \"2046b547-611f-4e98-ba95-a34cd0d45f50\") " pod="openstack/nova-cell0-db-create-t4vp6" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.404195 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fzmx\" (UniqueName: \"kubernetes.io/projected/2046b547-611f-4e98-ba95-a34cd0d45f50-kube-api-access-9fzmx\") pod \"nova-cell0-db-create-t4vp6\" (UID: \"2046b547-611f-4e98-ba95-a34cd0d45f50\") " pod="openstack/nova-cell0-db-create-t4vp6" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.404237 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/751e5ae1-be9d-48e8-8cc1-e08012854203-operator-scripts\") pod \"nova-api-88b3-account-create-update-zvx6c\" (UID: \"751e5ae1-be9d-48e8-8cc1-e08012854203\") " pod="openstack/nova-api-88b3-account-create-update-zvx6c" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.404261 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ml5d\" (UniqueName: \"kubernetes.io/projected/751e5ae1-be9d-48e8-8cc1-e08012854203-kube-api-access-6ml5d\") pod \"nova-api-88b3-account-create-update-zvx6c\" (UID: \"751e5ae1-be9d-48e8-8cc1-e08012854203\") " pod="openstack/nova-api-88b3-account-create-update-zvx6c" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.404405 4766 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.404418 4766 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae9e2d8d-6ce6-490a-8463-a078626b551f-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.404426 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.404436 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27ft2\" (UniqueName: \"kubernetes.io/projected/ae9e2d8d-6ce6-490a-8463-a078626b551f-kube-api-access-27ft2\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.404445 4766 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae9e2d8d-6ce6-490a-8463-a078626b551f-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.405676 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2046b547-611f-4e98-ba95-a34cd0d45f50-operator-scripts\") pod \"nova-cell0-db-create-t4vp6\" (UID: \"2046b547-611f-4e98-ba95-a34cd0d45f50\") " pod="openstack/nova-cell0-db-create-t4vp6" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.408250 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/751e5ae1-be9d-48e8-8cc1-e08012854203-operator-scripts\") pod \"nova-api-88b3-account-create-update-zvx6c\" (UID: \"751e5ae1-be9d-48e8-8cc1-e08012854203\") " pod="openstack/nova-api-88b3-account-create-update-zvx6c" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.418063 4766 scope.go:117] "RemoveContainer" containerID="11e1c320e18b1ba9311a6168bb1564ba38aaaf3a89e6ed61fa800644fa78f7b1" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.418721 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-jgd8f"] Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.420526 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jgd8f" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.431007 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ml5d\" (UniqueName: \"kubernetes.io/projected/751e5ae1-be9d-48e8-8cc1-e08012854203-kube-api-access-6ml5d\") pod \"nova-api-88b3-account-create-update-zvx6c\" (UID: \"751e5ae1-be9d-48e8-8cc1-e08012854203\") " pod="openstack/nova-api-88b3-account-create-update-zvx6c" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.436190 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-jgd8f"] Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.443865 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fzmx\" (UniqueName: \"kubernetes.io/projected/2046b547-611f-4e98-ba95-a34cd0d45f50-kube-api-access-9fzmx\") pod \"nova-cell0-db-create-t4vp6\" (UID: \"2046b547-611f-4e98-ba95-a34cd0d45f50\") " pod="openstack/nova-cell0-db-create-t4vp6" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.459627 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-k2tlc" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.484126 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-f00c-account-create-update-7jpjm"] Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.484664 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae9e2d8d-6ce6-490a-8463-a078626b551f" (UID: "ae9e2d8d-6ce6-490a-8463-a078626b551f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.486978 4766 scope.go:117] "RemoveContainer" containerID="c55670df27b91fd6586d9d88a8f46c9f9ff1651c0f04ee65aa18cbb449d8df61" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.490988 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f00c-account-create-update-7jpjm"] Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.491070 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f00c-account-create-update-7jpjm" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.494178 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.508048 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.533932 4766 scope.go:117] "RemoveContainer" containerID="84145d2dc6ed5f8fb4ea3ecc5fc4fadaa23dea0252b4b50ff8d1361c0a888d4f" Nov 26 00:46:47 crc kubenswrapper[4766]: E1126 00:46:47.542603 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84145d2dc6ed5f8fb4ea3ecc5fc4fadaa23dea0252b4b50ff8d1361c0a888d4f\": container with ID starting with 84145d2dc6ed5f8fb4ea3ecc5fc4fadaa23dea0252b4b50ff8d1361c0a888d4f not found: ID does not exist" containerID="84145d2dc6ed5f8fb4ea3ecc5fc4fadaa23dea0252b4b50ff8d1361c0a888d4f" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.542664 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84145d2dc6ed5f8fb4ea3ecc5fc4fadaa23dea0252b4b50ff8d1361c0a888d4f"} err="failed to get container status \"84145d2dc6ed5f8fb4ea3ecc5fc4fadaa23dea0252b4b50ff8d1361c0a888d4f\": rpc error: code = NotFound desc = could not find container \"84145d2dc6ed5f8fb4ea3ecc5fc4fadaa23dea0252b4b50ff8d1361c0a888d4f\": container with ID starting with 84145d2dc6ed5f8fb4ea3ecc5fc4fadaa23dea0252b4b50ff8d1361c0a888d4f not found: ID does not exist" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.542689 4766 scope.go:117] "RemoveContainer" containerID="3dc581af474a7f996d19f0dd09526f5df82d41a31b74f3887285df3d473ba827" Nov 26 00:46:47 crc kubenswrapper[4766]: E1126 00:46:47.549921 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dc581af474a7f996d19f0dd09526f5df82d41a31b74f3887285df3d473ba827\": container with ID starting with 3dc581af474a7f996d19f0dd09526f5df82d41a31b74f3887285df3d473ba827 not found: ID does not exist" containerID="3dc581af474a7f996d19f0dd09526f5df82d41a31b74f3887285df3d473ba827" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.549975 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dc581af474a7f996d19f0dd09526f5df82d41a31b74f3887285df3d473ba827"} err="failed to get container status \"3dc581af474a7f996d19f0dd09526f5df82d41a31b74f3887285df3d473ba827\": rpc error: code = NotFound desc = could not find container \"3dc581af474a7f996d19f0dd09526f5df82d41a31b74f3887285df3d473ba827\": container with ID starting with 3dc581af474a7f996d19f0dd09526f5df82d41a31b74f3887285df3d473ba827 not found: ID does not exist" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.550005 4766 scope.go:117] "RemoveContainer" containerID="11e1c320e18b1ba9311a6168bb1564ba38aaaf3a89e6ed61fa800644fa78f7b1" Nov 26 00:46:47 crc kubenswrapper[4766]: E1126 00:46:47.553767 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11e1c320e18b1ba9311a6168bb1564ba38aaaf3a89e6ed61fa800644fa78f7b1\": container with ID starting with 11e1c320e18b1ba9311a6168bb1564ba38aaaf3a89e6ed61fa800644fa78f7b1 not found: ID does not exist" containerID="11e1c320e18b1ba9311a6168bb1564ba38aaaf3a89e6ed61fa800644fa78f7b1" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.553814 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11e1c320e18b1ba9311a6168bb1564ba38aaaf3a89e6ed61fa800644fa78f7b1"} err="failed to get container status \"11e1c320e18b1ba9311a6168bb1564ba38aaaf3a89e6ed61fa800644fa78f7b1\": rpc error: code = NotFound desc = could not find container \"11e1c320e18b1ba9311a6168bb1564ba38aaaf3a89e6ed61fa800644fa78f7b1\": container with ID starting with 11e1c320e18b1ba9311a6168bb1564ba38aaaf3a89e6ed61fa800644fa78f7b1 not found: ID does not exist" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.553837 4766 scope.go:117] "RemoveContainer" containerID="c55670df27b91fd6586d9d88a8f46c9f9ff1651c0f04ee65aa18cbb449d8df61" Nov 26 00:46:47 crc kubenswrapper[4766]: E1126 00:46:47.554189 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c55670df27b91fd6586d9d88a8f46c9f9ff1651c0f04ee65aa18cbb449d8df61\": container with ID starting with c55670df27b91fd6586d9d88a8f46c9f9ff1651c0f04ee65aa18cbb449d8df61 not found: ID does not exist" containerID="c55670df27b91fd6586d9d88a8f46c9f9ff1651c0f04ee65aa18cbb449d8df61" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.554242 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c55670df27b91fd6586d9d88a8f46c9f9ff1651c0f04ee65aa18cbb449d8df61"} err="failed to get container status \"c55670df27b91fd6586d9d88a8f46c9f9ff1651c0f04ee65aa18cbb449d8df61\": rpc error: code = NotFound desc = could not find container \"c55670df27b91fd6586d9d88a8f46c9f9ff1651c0f04ee65aa18cbb449d8df61\": container with ID starting with c55670df27b91fd6586d9d88a8f46c9f9ff1651c0f04ee65aa18cbb449d8df61 not found: ID does not exist" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.556750 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-config-data" (OuterVolumeSpecName: "config-data") pod "ae9e2d8d-6ce6-490a-8463-a078626b551f" (UID: "ae9e2d8d-6ce6-490a-8463-a078626b551f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.582175 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-t4vp6" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.610202 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4794066a-0e54-4387-ae01-078abc184684-operator-scripts\") pod \"nova-cell0-f00c-account-create-update-7jpjm\" (UID: \"4794066a-0e54-4387-ae01-078abc184684\") " pod="openstack/nova-cell0-f00c-account-create-update-7jpjm" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.610509 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7vj9\" (UniqueName: \"kubernetes.io/projected/4794066a-0e54-4387-ae01-078abc184684-kube-api-access-x7vj9\") pod \"nova-cell0-f00c-account-create-update-7jpjm\" (UID: \"4794066a-0e54-4387-ae01-078abc184684\") " pod="openstack/nova-cell0-f00c-account-create-update-7jpjm" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.610530 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr976\" (UniqueName: \"kubernetes.io/projected/b20f822d-71d5-4cb8-823b-8c261af60109-kube-api-access-gr976\") pod \"nova-cell1-db-create-jgd8f\" (UID: \"b20f822d-71d5-4cb8-823b-8c261af60109\") " pod="openstack/nova-cell1-db-create-jgd8f" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.610562 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b20f822d-71d5-4cb8-823b-8c261af60109-operator-scripts\") pod \"nova-cell1-db-create-jgd8f\" (UID: \"b20f822d-71d5-4cb8-823b-8c261af60109\") " pod="openstack/nova-cell1-db-create-jgd8f" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.610734 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae9e2d8d-6ce6-490a-8463-a078626b551f-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.643104 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-88b3-account-create-update-zvx6c" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.662773 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.681746 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-d77a-account-create-update-8gxtj"] Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.683551 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d77a-account-create-update-8gxtj" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.689325 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.711539 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-d77a-account-create-update-8gxtj"] Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.713104 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr9qz\" (UniqueName: \"kubernetes.io/projected/2d4b645a-ddab-4a29-8ffa-77e7873591c8-kube-api-access-kr9qz\") pod \"nova-cell1-d77a-account-create-update-8gxtj\" (UID: \"2d4b645a-ddab-4a29-8ffa-77e7873591c8\") " pod="openstack/nova-cell1-d77a-account-create-update-8gxtj" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.713215 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4794066a-0e54-4387-ae01-078abc184684-operator-scripts\") pod \"nova-cell0-f00c-account-create-update-7jpjm\" (UID: \"4794066a-0e54-4387-ae01-078abc184684\") " pod="openstack/nova-cell0-f00c-account-create-update-7jpjm" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.713245 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7vj9\" (UniqueName: \"kubernetes.io/projected/4794066a-0e54-4387-ae01-078abc184684-kube-api-access-x7vj9\") pod \"nova-cell0-f00c-account-create-update-7jpjm\" (UID: \"4794066a-0e54-4387-ae01-078abc184684\") " pod="openstack/nova-cell0-f00c-account-create-update-7jpjm" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.713264 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr976\" (UniqueName: \"kubernetes.io/projected/b20f822d-71d5-4cb8-823b-8c261af60109-kube-api-access-gr976\") pod \"nova-cell1-db-create-jgd8f\" (UID: \"b20f822d-71d5-4cb8-823b-8c261af60109\") " pod="openstack/nova-cell1-db-create-jgd8f" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.713296 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b20f822d-71d5-4cb8-823b-8c261af60109-operator-scripts\") pod \"nova-cell1-db-create-jgd8f\" (UID: \"b20f822d-71d5-4cb8-823b-8c261af60109\") " pod="openstack/nova-cell1-db-create-jgd8f" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.713348 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d4b645a-ddab-4a29-8ffa-77e7873591c8-operator-scripts\") pod \"nova-cell1-d77a-account-create-update-8gxtj\" (UID: \"2d4b645a-ddab-4a29-8ffa-77e7873591c8\") " pod="openstack/nova-cell1-d77a-account-create-update-8gxtj" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.714024 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4794066a-0e54-4387-ae01-078abc184684-operator-scripts\") pod \"nova-cell0-f00c-account-create-update-7jpjm\" (UID: \"4794066a-0e54-4387-ae01-078abc184684\") " pod="openstack/nova-cell0-f00c-account-create-update-7jpjm" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.714901 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b20f822d-71d5-4cb8-823b-8c261af60109-operator-scripts\") pod \"nova-cell1-db-create-jgd8f\" (UID: \"b20f822d-71d5-4cb8-823b-8c261af60109\") " pod="openstack/nova-cell1-db-create-jgd8f" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.768273 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7vj9\" (UniqueName: \"kubernetes.io/projected/4794066a-0e54-4387-ae01-078abc184684-kube-api-access-x7vj9\") pod \"nova-cell0-f00c-account-create-update-7jpjm\" (UID: \"4794066a-0e54-4387-ae01-078abc184684\") " pod="openstack/nova-cell0-f00c-account-create-update-7jpjm" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.784437 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr976\" (UniqueName: \"kubernetes.io/projected/b20f822d-71d5-4cb8-823b-8c261af60109-kube-api-access-gr976\") pod \"nova-cell1-db-create-jgd8f\" (UID: \"b20f822d-71d5-4cb8-823b-8c261af60109\") " pod="openstack/nova-cell1-db-create-jgd8f" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.825348 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f00c-account-create-update-7jpjm" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.826643 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d4b645a-ddab-4a29-8ffa-77e7873591c8-operator-scripts\") pod \"nova-cell1-d77a-account-create-update-8gxtj\" (UID: \"2d4b645a-ddab-4a29-8ffa-77e7873591c8\") " pod="openstack/nova-cell1-d77a-account-create-update-8gxtj" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.826779 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kr9qz\" (UniqueName: \"kubernetes.io/projected/2d4b645a-ddab-4a29-8ffa-77e7873591c8-kube-api-access-kr9qz\") pod \"nova-cell1-d77a-account-create-update-8gxtj\" (UID: \"2d4b645a-ddab-4a29-8ffa-77e7873591c8\") " pod="openstack/nova-cell1-d77a-account-create-update-8gxtj" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.827813 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d4b645a-ddab-4a29-8ffa-77e7873591c8-operator-scripts\") pod \"nova-cell1-d77a-account-create-update-8gxtj\" (UID: \"2d4b645a-ddab-4a29-8ffa-77e7873591c8\") " pod="openstack/nova-cell1-d77a-account-create-update-8gxtj" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.859589 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kr9qz\" (UniqueName: \"kubernetes.io/projected/2d4b645a-ddab-4a29-8ffa-77e7873591c8-kube-api-access-kr9qz\") pod \"nova-cell1-d77a-account-create-update-8gxtj\" (UID: \"2d4b645a-ddab-4a29-8ffa-77e7873591c8\") " pod="openstack/nova-cell1-d77a-account-create-update-8gxtj" Nov 26 00:46:47 crc kubenswrapper[4766]: I1126 00:46:47.949934 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d77a-account-create-update-8gxtj" Nov 26 00:46:48 crc kubenswrapper[4766]: I1126 00:46:48.063539 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jgd8f" Nov 26 00:46:48 crc kubenswrapper[4766]: I1126 00:46:48.300733 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-k2tlc"] Nov 26 00:46:48 crc kubenswrapper[4766]: I1126 00:46:48.338871 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-t4vp6"] Nov 26 00:46:48 crc kubenswrapper[4766]: I1126 00:46:48.384726 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-668d6b6d67-r2rrd" Nov 26 00:46:48 crc kubenswrapper[4766]: I1126 00:46:48.385505 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2bba0b1b-4848-4452-a535-c0cf5d13b30b","Type":"ContainerStarted","Data":"439388a00f68c195095ae7ba1fdbfc797d7bcd3ad14451123a18decdb7c2c6e0"} Nov 26 00:46:48 crc kubenswrapper[4766]: I1126 00:46:48.429302 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-t4vp6" event={"ID":"2046b547-611f-4e98-ba95-a34cd0d45f50","Type":"ContainerStarted","Data":"06153b388c9dc8b1f01ad3f8d215dcadbf1d93c2cfc7bf914d724dbc2edc47ad"} Nov 26 00:46:48 crc kubenswrapper[4766]: I1126 00:46:48.440551 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-6dfcdbc5c7-znfmt"] Nov 26 00:46:48 crc kubenswrapper[4766]: I1126 00:46:48.440882 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-6dfcdbc5c7-znfmt" podUID="13e19aaf-7cc3-40a6-852b-72b77b37cf1e" containerName="heat-engine" containerID="cri-o://10340d467e512bd68e6e6b88a6746642d2c9504b21804f3be63d9c75ddd06bd8" gracePeriod=60 Nov 26 00:46:48 crc kubenswrapper[4766]: E1126 00:46:48.448548 4766 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="10340d467e512bd68e6e6b88a6746642d2c9504b21804f3be63d9c75ddd06bd8" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 26 00:46:48 crc kubenswrapper[4766]: E1126 00:46:48.452371 4766 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="10340d467e512bd68e6e6b88a6746642d2c9504b21804f3be63d9c75ddd06bd8" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 26 00:46:48 crc kubenswrapper[4766]: E1126 00:46:48.468147 4766 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="10340d467e512bd68e6e6b88a6746642d2c9504b21804f3be63d9c75ddd06bd8" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 26 00:46:48 crc kubenswrapper[4766]: E1126 00:46:48.468212 4766 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-6dfcdbc5c7-znfmt" podUID="13e19aaf-7cc3-40a6-852b-72b77b37cf1e" containerName="heat-engine" Nov 26 00:46:48 crc kubenswrapper[4766]: I1126 00:46:48.624515 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-88b3-account-create-update-zvx6c"] Nov 26 00:46:48 crc kubenswrapper[4766]: I1126 00:46:48.811406 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-d77a-account-create-update-8gxtj"] Nov 26 00:46:49 crc kubenswrapper[4766]: I1126 00:46:49.056895 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f00c-account-create-update-7jpjm"] Nov 26 00:46:49 crc kubenswrapper[4766]: I1126 00:46:49.098864 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-jgd8f"] Nov 26 00:46:49 crc kubenswrapper[4766]: I1126 00:46:49.445820 4766 generic.go:334] "Generic (PLEG): container finished" podID="2046b547-611f-4e98-ba95-a34cd0d45f50" containerID="ad711cfd11b280b1818b457932f68682de01db5dad05a749c800cdb8c87f64f3" exitCode=0 Nov 26 00:46:49 crc kubenswrapper[4766]: I1126 00:46:49.445882 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-t4vp6" event={"ID":"2046b547-611f-4e98-ba95-a34cd0d45f50","Type":"ContainerDied","Data":"ad711cfd11b280b1818b457932f68682de01db5dad05a749c800cdb8c87f64f3"} Nov 26 00:46:49 crc kubenswrapper[4766]: I1126 00:46:49.455291 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d77a-account-create-update-8gxtj" event={"ID":"2d4b645a-ddab-4a29-8ffa-77e7873591c8","Type":"ContainerStarted","Data":"97606350c9b9dcf26aa878a45371aa3ff32e85934db2d23fad1153f8d0243913"} Nov 26 00:46:49 crc kubenswrapper[4766]: I1126 00:46:49.455343 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d77a-account-create-update-8gxtj" event={"ID":"2d4b645a-ddab-4a29-8ffa-77e7873591c8","Type":"ContainerStarted","Data":"2e835821f2b9e0af0eff67b2a4bd3fbfe5aeaaa837856fa9ef6361a333c02ce5"} Nov 26 00:46:49 crc kubenswrapper[4766]: I1126 00:46:49.475791 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jgd8f" event={"ID":"b20f822d-71d5-4cb8-823b-8c261af60109","Type":"ContainerStarted","Data":"7ef0ad34acd79fd20333d11bb7ff0b1dcfccd0cb0517a49063c065d8eaf6eeb0"} Nov 26 00:46:49 crc kubenswrapper[4766]: I1126 00:46:49.475930 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jgd8f" event={"ID":"b20f822d-71d5-4cb8-823b-8c261af60109","Type":"ContainerStarted","Data":"f59936cc2d3cb22f275a36a6594fcd36b63dce5654bf06d3947fe579a90f7311"} Nov 26 00:46:49 crc kubenswrapper[4766]: I1126 00:46:49.485191 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-d77a-account-create-update-8gxtj" podStartSLOduration=2.485168464 podStartE2EDuration="2.485168464s" podCreationTimestamp="2025-11-26 00:46:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:46:49.477046081 +0000 UTC m=+1390.325816511" watchObservedRunningTime="2025-11-26 00:46:49.485168464 +0000 UTC m=+1390.333938894" Nov 26 00:46:49 crc kubenswrapper[4766]: I1126 00:46:49.496724 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f00c-account-create-update-7jpjm" event={"ID":"4794066a-0e54-4387-ae01-078abc184684","Type":"ContainerStarted","Data":"4cd953a4be032f4df75da118bdc15379473d4b9254e10720f6d0045cb08ac8a2"} Nov 26 00:46:49 crc kubenswrapper[4766]: I1126 00:46:49.496776 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f00c-account-create-update-7jpjm" event={"ID":"4794066a-0e54-4387-ae01-078abc184684","Type":"ContainerStarted","Data":"90f154c1367b84398af5309a7403114c25c1f7297ed96a88078f9e55a521ca28"} Nov 26 00:46:49 crc kubenswrapper[4766]: I1126 00:46:49.503604 4766 generic.go:334] "Generic (PLEG): container finished" podID="619650be-59d2-47c8-815d-3337c156c83f" containerID="5c730b2e7bbdd15b5a9c9c7517ec31cbe511dffb42429118fd3355ce409a52af" exitCode=0 Nov 26 00:46:49 crc kubenswrapper[4766]: I1126 00:46:49.503733 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-k2tlc" event={"ID":"619650be-59d2-47c8-815d-3337c156c83f","Type":"ContainerDied","Data":"5c730b2e7bbdd15b5a9c9c7517ec31cbe511dffb42429118fd3355ce409a52af"} Nov 26 00:46:49 crc kubenswrapper[4766]: I1126 00:46:49.503770 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-k2tlc" event={"ID":"619650be-59d2-47c8-815d-3337c156c83f","Type":"ContainerStarted","Data":"e9f346fea6d355678d03fff6368a75e1f94b5a2a6af90540fdf3b654bf6860d0"} Nov 26 00:46:49 crc kubenswrapper[4766]: I1126 00:46:49.511963 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2bba0b1b-4848-4452-a535-c0cf5d13b30b","Type":"ContainerStarted","Data":"bf1673580fa8e1a522a2bf6426b11db0235234df253170289ee2b6d16def7445"} Nov 26 00:46:49 crc kubenswrapper[4766]: I1126 00:46:49.518062 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-jgd8f" podStartSLOduration=2.518039968 podStartE2EDuration="2.518039968s" podCreationTimestamp="2025-11-26 00:46:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:46:49.495071233 +0000 UTC m=+1390.343841663" watchObservedRunningTime="2025-11-26 00:46:49.518039968 +0000 UTC m=+1390.366810398" Nov 26 00:46:49 crc kubenswrapper[4766]: I1126 00:46:49.523340 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-88b3-account-create-update-zvx6c" event={"ID":"751e5ae1-be9d-48e8-8cc1-e08012854203","Type":"ContainerStarted","Data":"fd8749b745d14b29e866e787691471be81cb43466bb8932ef270ac001dc988eb"} Nov 26 00:46:49 crc kubenswrapper[4766]: I1126 00:46:49.523386 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-88b3-account-create-update-zvx6c" event={"ID":"751e5ae1-be9d-48e8-8cc1-e08012854203","Type":"ContainerStarted","Data":"f24719beaba831a8878bee83380f994cc05aa12b8c4140d7985aefe9eb23b644"} Nov 26 00:46:49 crc kubenswrapper[4766]: I1126 00:46:49.529060 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-f00c-account-create-update-7jpjm" podStartSLOduration=2.529040374 podStartE2EDuration="2.529040374s" podCreationTimestamp="2025-11-26 00:46:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:46:49.515524745 +0000 UTC m=+1390.364295175" watchObservedRunningTime="2025-11-26 00:46:49.529040374 +0000 UTC m=+1390.377810804" Nov 26 00:46:49 crc kubenswrapper[4766]: I1126 00:46:49.564929 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=3.564909203 podStartE2EDuration="3.564909203s" podCreationTimestamp="2025-11-26 00:46:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:46:49.553394625 +0000 UTC m=+1390.402165055" watchObservedRunningTime="2025-11-26 00:46:49.564909203 +0000 UTC m=+1390.413679643" Nov 26 00:46:50 crc kubenswrapper[4766]: I1126 00:46:50.536291 4766 generic.go:334] "Generic (PLEG): container finished" podID="2d4b645a-ddab-4a29-8ffa-77e7873591c8" containerID="97606350c9b9dcf26aa878a45371aa3ff32e85934db2d23fad1153f8d0243913" exitCode=0 Nov 26 00:46:50 crc kubenswrapper[4766]: I1126 00:46:50.536496 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d77a-account-create-update-8gxtj" event={"ID":"2d4b645a-ddab-4a29-8ffa-77e7873591c8","Type":"ContainerDied","Data":"97606350c9b9dcf26aa878a45371aa3ff32e85934db2d23fad1153f8d0243913"} Nov 26 00:46:50 crc kubenswrapper[4766]: I1126 00:46:50.539846 4766 generic.go:334] "Generic (PLEG): container finished" podID="b20f822d-71d5-4cb8-823b-8c261af60109" containerID="7ef0ad34acd79fd20333d11bb7ff0b1dcfccd0cb0517a49063c065d8eaf6eeb0" exitCode=0 Nov 26 00:46:50 crc kubenswrapper[4766]: I1126 00:46:50.539929 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jgd8f" event={"ID":"b20f822d-71d5-4cb8-823b-8c261af60109","Type":"ContainerDied","Data":"7ef0ad34acd79fd20333d11bb7ff0b1dcfccd0cb0517a49063c065d8eaf6eeb0"} Nov 26 00:46:50 crc kubenswrapper[4766]: I1126 00:46:50.542271 4766 generic.go:334] "Generic (PLEG): container finished" podID="4794066a-0e54-4387-ae01-078abc184684" containerID="4cd953a4be032f4df75da118bdc15379473d4b9254e10720f6d0045cb08ac8a2" exitCode=0 Nov 26 00:46:50 crc kubenswrapper[4766]: I1126 00:46:50.542327 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f00c-account-create-update-7jpjm" event={"ID":"4794066a-0e54-4387-ae01-078abc184684","Type":"ContainerDied","Data":"4cd953a4be032f4df75da118bdc15379473d4b9254e10720f6d0045cb08ac8a2"} Nov 26 00:46:50 crc kubenswrapper[4766]: I1126 00:46:50.543723 4766 generic.go:334] "Generic (PLEG): container finished" podID="751e5ae1-be9d-48e8-8cc1-e08012854203" containerID="fd8749b745d14b29e866e787691471be81cb43466bb8932ef270ac001dc988eb" exitCode=0 Nov 26 00:46:50 crc kubenswrapper[4766]: I1126 00:46:50.543908 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-88b3-account-create-update-zvx6c" event={"ID":"751e5ae1-be9d-48e8-8cc1-e08012854203","Type":"ContainerDied","Data":"fd8749b745d14b29e866e787691471be81cb43466bb8932ef270ac001dc988eb"} Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.131964 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.166774 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-88b3-account-create-update-zvx6c" Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.326477 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ml5d\" (UniqueName: \"kubernetes.io/projected/751e5ae1-be9d-48e8-8cc1-e08012854203-kube-api-access-6ml5d\") pod \"751e5ae1-be9d-48e8-8cc1-e08012854203\" (UID: \"751e5ae1-be9d-48e8-8cc1-e08012854203\") " Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.326575 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/751e5ae1-be9d-48e8-8cc1-e08012854203-operator-scripts\") pod \"751e5ae1-be9d-48e8-8cc1-e08012854203\" (UID: \"751e5ae1-be9d-48e8-8cc1-e08012854203\") " Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.327922 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/751e5ae1-be9d-48e8-8cc1-e08012854203-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "751e5ae1-be9d-48e8-8cc1-e08012854203" (UID: "751e5ae1-be9d-48e8-8cc1-e08012854203"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.339644 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/751e5ae1-be9d-48e8-8cc1-e08012854203-kube-api-access-6ml5d" (OuterVolumeSpecName: "kube-api-access-6ml5d") pod "751e5ae1-be9d-48e8-8cc1-e08012854203" (UID: "751e5ae1-be9d-48e8-8cc1-e08012854203"). InnerVolumeSpecName "kube-api-access-6ml5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.425575 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-t4vp6" Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.431162 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ml5d\" (UniqueName: \"kubernetes.io/projected/751e5ae1-be9d-48e8-8cc1-e08012854203-kube-api-access-6ml5d\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.431198 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/751e5ae1-be9d-48e8-8cc1-e08012854203-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.432862 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-k2tlc" Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.538638 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtnln\" (UniqueName: \"kubernetes.io/projected/619650be-59d2-47c8-815d-3337c156c83f-kube-api-access-mtnln\") pod \"619650be-59d2-47c8-815d-3337c156c83f\" (UID: \"619650be-59d2-47c8-815d-3337c156c83f\") " Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.540130 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2046b547-611f-4e98-ba95-a34cd0d45f50-operator-scripts\") pod \"2046b547-611f-4e98-ba95-a34cd0d45f50\" (UID: \"2046b547-611f-4e98-ba95-a34cd0d45f50\") " Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.540158 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fzmx\" (UniqueName: \"kubernetes.io/projected/2046b547-611f-4e98-ba95-a34cd0d45f50-kube-api-access-9fzmx\") pod \"2046b547-611f-4e98-ba95-a34cd0d45f50\" (UID: \"2046b547-611f-4e98-ba95-a34cd0d45f50\") " Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.540206 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/619650be-59d2-47c8-815d-3337c156c83f-operator-scripts\") pod \"619650be-59d2-47c8-815d-3337c156c83f\" (UID: \"619650be-59d2-47c8-815d-3337c156c83f\") " Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.541356 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2046b547-611f-4e98-ba95-a34cd0d45f50-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2046b547-611f-4e98-ba95-a34cd0d45f50" (UID: "2046b547-611f-4e98-ba95-a34cd0d45f50"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.541426 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/619650be-59d2-47c8-815d-3337c156c83f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "619650be-59d2-47c8-815d-3337c156c83f" (UID: "619650be-59d2-47c8-815d-3337c156c83f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.548430 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/619650be-59d2-47c8-815d-3337c156c83f-kube-api-access-mtnln" (OuterVolumeSpecName: "kube-api-access-mtnln") pod "619650be-59d2-47c8-815d-3337c156c83f" (UID: "619650be-59d2-47c8-815d-3337c156c83f"). InnerVolumeSpecName "kube-api-access-mtnln". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.607062 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2046b547-611f-4e98-ba95-a34cd0d45f50-kube-api-access-9fzmx" (OuterVolumeSpecName: "kube-api-access-9fzmx") pod "2046b547-611f-4e98-ba95-a34cd0d45f50" (UID: "2046b547-611f-4e98-ba95-a34cd0d45f50"). InnerVolumeSpecName "kube-api-access-9fzmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.607172 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-k2tlc" event={"ID":"619650be-59d2-47c8-815d-3337c156c83f","Type":"ContainerDied","Data":"e9f346fea6d355678d03fff6368a75e1f94b5a2a6af90540fdf3b654bf6860d0"} Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.607797 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9f346fea6d355678d03fff6368a75e1f94b5a2a6af90540fdf3b654bf6860d0" Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.607260 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-k2tlc" Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.623643 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-88b3-account-create-update-zvx6c" event={"ID":"751e5ae1-be9d-48e8-8cc1-e08012854203","Type":"ContainerDied","Data":"f24719beaba831a8878bee83380f994cc05aa12b8c4140d7985aefe9eb23b644"} Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.623708 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f24719beaba831a8878bee83380f994cc05aa12b8c4140d7985aefe9eb23b644" Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.623805 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-88b3-account-create-update-zvx6c" Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.634990 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-t4vp6" Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.641181 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-t4vp6" event={"ID":"2046b547-611f-4e98-ba95-a34cd0d45f50","Type":"ContainerDied","Data":"06153b388c9dc8b1f01ad3f8d215dcadbf1d93c2cfc7bf914d724dbc2edc47ad"} Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.641221 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06153b388c9dc8b1f01ad3f8d215dcadbf1d93c2cfc7bf914d724dbc2edc47ad" Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.643048 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2046b547-611f-4e98-ba95-a34cd0d45f50-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.643068 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fzmx\" (UniqueName: \"kubernetes.io/projected/2046b547-611f-4e98-ba95-a34cd0d45f50-kube-api-access-9fzmx\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.643078 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/619650be-59d2-47c8-815d-3337c156c83f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:51 crc kubenswrapper[4766]: I1126 00:46:51.643097 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtnln\" (UniqueName: \"kubernetes.io/projected/619650be-59d2-47c8-815d-3337c156c83f-kube-api-access-mtnln\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:52 crc kubenswrapper[4766]: E1126 00:46:52.092240 4766 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="10340d467e512bd68e6e6b88a6746642d2c9504b21804f3be63d9c75ddd06bd8" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 26 00:46:52 crc kubenswrapper[4766]: E1126 00:46:52.096806 4766 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="10340d467e512bd68e6e6b88a6746642d2c9504b21804f3be63d9c75ddd06bd8" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 26 00:46:52 crc kubenswrapper[4766]: E1126 00:46:52.100772 4766 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="10340d467e512bd68e6e6b88a6746642d2c9504b21804f3be63d9c75ddd06bd8" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 26 00:46:52 crc kubenswrapper[4766]: E1126 00:46:52.100843 4766 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-6dfcdbc5c7-znfmt" podUID="13e19aaf-7cc3-40a6-852b-72b77b37cf1e" containerName="heat-engine" Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.447232 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jgd8f" Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.459545 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d77a-account-create-update-8gxtj" Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.478895 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f00c-account-create-update-7jpjm" Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.579597 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b20f822d-71d5-4cb8-823b-8c261af60109-operator-scripts\") pod \"b20f822d-71d5-4cb8-823b-8c261af60109\" (UID: \"b20f822d-71d5-4cb8-823b-8c261af60109\") " Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.579683 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kr9qz\" (UniqueName: \"kubernetes.io/projected/2d4b645a-ddab-4a29-8ffa-77e7873591c8-kube-api-access-kr9qz\") pod \"2d4b645a-ddab-4a29-8ffa-77e7873591c8\" (UID: \"2d4b645a-ddab-4a29-8ffa-77e7873591c8\") " Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.579729 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4794066a-0e54-4387-ae01-078abc184684-operator-scripts\") pod \"4794066a-0e54-4387-ae01-078abc184684\" (UID: \"4794066a-0e54-4387-ae01-078abc184684\") " Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.579854 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d4b645a-ddab-4a29-8ffa-77e7873591c8-operator-scripts\") pod \"2d4b645a-ddab-4a29-8ffa-77e7873591c8\" (UID: \"2d4b645a-ddab-4a29-8ffa-77e7873591c8\") " Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.579914 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7vj9\" (UniqueName: \"kubernetes.io/projected/4794066a-0e54-4387-ae01-078abc184684-kube-api-access-x7vj9\") pod \"4794066a-0e54-4387-ae01-078abc184684\" (UID: \"4794066a-0e54-4387-ae01-078abc184684\") " Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.579978 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gr976\" (UniqueName: \"kubernetes.io/projected/b20f822d-71d5-4cb8-823b-8c261af60109-kube-api-access-gr976\") pod \"b20f822d-71d5-4cb8-823b-8c261af60109\" (UID: \"b20f822d-71d5-4cb8-823b-8c261af60109\") " Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.580642 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b20f822d-71d5-4cb8-823b-8c261af60109-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b20f822d-71d5-4cb8-823b-8c261af60109" (UID: "b20f822d-71d5-4cb8-823b-8c261af60109"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.580686 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4794066a-0e54-4387-ae01-078abc184684-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4794066a-0e54-4387-ae01-078abc184684" (UID: "4794066a-0e54-4387-ae01-078abc184684"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.581068 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d4b645a-ddab-4a29-8ffa-77e7873591c8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2d4b645a-ddab-4a29-8ffa-77e7873591c8" (UID: "2d4b645a-ddab-4a29-8ffa-77e7873591c8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.581412 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d4b645a-ddab-4a29-8ffa-77e7873591c8-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.581424 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b20f822d-71d5-4cb8-823b-8c261af60109-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.581431 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4794066a-0e54-4387-ae01-078abc184684-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.587681 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b20f822d-71d5-4cb8-823b-8c261af60109-kube-api-access-gr976" (OuterVolumeSpecName: "kube-api-access-gr976") pod "b20f822d-71d5-4cb8-823b-8c261af60109" (UID: "b20f822d-71d5-4cb8-823b-8c261af60109"). InnerVolumeSpecName "kube-api-access-gr976". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.588445 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d4b645a-ddab-4a29-8ffa-77e7873591c8-kube-api-access-kr9qz" (OuterVolumeSpecName: "kube-api-access-kr9qz") pod "2d4b645a-ddab-4a29-8ffa-77e7873591c8" (UID: "2d4b645a-ddab-4a29-8ffa-77e7873591c8"). InnerVolumeSpecName "kube-api-access-kr9qz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.598397 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4794066a-0e54-4387-ae01-078abc184684-kube-api-access-x7vj9" (OuterVolumeSpecName: "kube-api-access-x7vj9") pod "4794066a-0e54-4387-ae01-078abc184684" (UID: "4794066a-0e54-4387-ae01-078abc184684"). InnerVolumeSpecName "kube-api-access-x7vj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.651746 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d77a-account-create-update-8gxtj" event={"ID":"2d4b645a-ddab-4a29-8ffa-77e7873591c8","Type":"ContainerDied","Data":"2e835821f2b9e0af0eff67b2a4bd3fbfe5aeaaa837856fa9ef6361a333c02ce5"} Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.651824 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e835821f2b9e0af0eff67b2a4bd3fbfe5aeaaa837856fa9ef6361a333c02ce5" Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.652064 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d77a-account-create-update-8gxtj" Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.657861 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jgd8f" event={"ID":"b20f822d-71d5-4cb8-823b-8c261af60109","Type":"ContainerDied","Data":"f59936cc2d3cb22f275a36a6594fcd36b63dce5654bf06d3947fe579a90f7311"} Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.657902 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f59936cc2d3cb22f275a36a6594fcd36b63dce5654bf06d3947fe579a90f7311" Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.658048 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jgd8f" Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.665423 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f00c-account-create-update-7jpjm" event={"ID":"4794066a-0e54-4387-ae01-078abc184684","Type":"ContainerDied","Data":"90f154c1367b84398af5309a7403114c25c1f7297ed96a88078f9e55a521ca28"} Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.665462 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90f154c1367b84398af5309a7403114c25c1f7297ed96a88078f9e55a521ca28" Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.665514 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f00c-account-create-update-7jpjm" Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.683103 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7vj9\" (UniqueName: \"kubernetes.io/projected/4794066a-0e54-4387-ae01-078abc184684-kube-api-access-x7vj9\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.683339 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gr976\" (UniqueName: \"kubernetes.io/projected/b20f822d-71d5-4cb8-823b-8c261af60109-kube-api-access-gr976\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:52 crc kubenswrapper[4766]: I1126 00:46:52.683432 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kr9qz\" (UniqueName: \"kubernetes.io/projected/2d4b645a-ddab-4a29-8ffa-77e7873591c8-kube-api-access-kr9qz\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:56 crc kubenswrapper[4766]: I1126 00:46:56.639379 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6dfcdbc5c7-znfmt" Nov 26 00:46:56 crc kubenswrapper[4766]: I1126 00:46:56.670959 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46chp\" (UniqueName: \"kubernetes.io/projected/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-kube-api-access-46chp\") pod \"13e19aaf-7cc3-40a6-852b-72b77b37cf1e\" (UID: \"13e19aaf-7cc3-40a6-852b-72b77b37cf1e\") " Nov 26 00:46:56 crc kubenswrapper[4766]: I1126 00:46:56.671180 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-combined-ca-bundle\") pod \"13e19aaf-7cc3-40a6-852b-72b77b37cf1e\" (UID: \"13e19aaf-7cc3-40a6-852b-72b77b37cf1e\") " Nov 26 00:46:56 crc kubenswrapper[4766]: I1126 00:46:56.671262 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-config-data-custom\") pod \"13e19aaf-7cc3-40a6-852b-72b77b37cf1e\" (UID: \"13e19aaf-7cc3-40a6-852b-72b77b37cf1e\") " Nov 26 00:46:56 crc kubenswrapper[4766]: I1126 00:46:56.671302 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-config-data\") pod \"13e19aaf-7cc3-40a6-852b-72b77b37cf1e\" (UID: \"13e19aaf-7cc3-40a6-852b-72b77b37cf1e\") " Nov 26 00:46:56 crc kubenswrapper[4766]: I1126 00:46:56.677973 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "13e19aaf-7cc3-40a6-852b-72b77b37cf1e" (UID: "13e19aaf-7cc3-40a6-852b-72b77b37cf1e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:56 crc kubenswrapper[4766]: I1126 00:46:56.684202 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-kube-api-access-46chp" (OuterVolumeSpecName: "kube-api-access-46chp") pod "13e19aaf-7cc3-40a6-852b-72b77b37cf1e" (UID: "13e19aaf-7cc3-40a6-852b-72b77b37cf1e"). InnerVolumeSpecName "kube-api-access-46chp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:46:56 crc kubenswrapper[4766]: I1126 00:46:56.710632 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13e19aaf-7cc3-40a6-852b-72b77b37cf1e" (UID: "13e19aaf-7cc3-40a6-852b-72b77b37cf1e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:56 crc kubenswrapper[4766]: I1126 00:46:56.713771 4766 generic.go:334] "Generic (PLEG): container finished" podID="13e19aaf-7cc3-40a6-852b-72b77b37cf1e" containerID="10340d467e512bd68e6e6b88a6746642d2c9504b21804f3be63d9c75ddd06bd8" exitCode=0 Nov 26 00:46:56 crc kubenswrapper[4766]: I1126 00:46:56.713812 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6dfcdbc5c7-znfmt" Nov 26 00:46:56 crc kubenswrapper[4766]: I1126 00:46:56.713839 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6dfcdbc5c7-znfmt" event={"ID":"13e19aaf-7cc3-40a6-852b-72b77b37cf1e","Type":"ContainerDied","Data":"10340d467e512bd68e6e6b88a6746642d2c9504b21804f3be63d9c75ddd06bd8"} Nov 26 00:46:56 crc kubenswrapper[4766]: I1126 00:46:56.713866 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6dfcdbc5c7-znfmt" event={"ID":"13e19aaf-7cc3-40a6-852b-72b77b37cf1e","Type":"ContainerDied","Data":"0e23105208a328d7118d59c228dbc87fa4ee01cb7e3a50b48fa25c206c1c9c52"} Nov 26 00:46:56 crc kubenswrapper[4766]: I1126 00:46:56.713896 4766 scope.go:117] "RemoveContainer" containerID="10340d467e512bd68e6e6b88a6746642d2c9504b21804f3be63d9c75ddd06bd8" Nov 26 00:46:56 crc kubenswrapper[4766]: I1126 00:46:56.768140 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-config-data" (OuterVolumeSpecName: "config-data") pod "13e19aaf-7cc3-40a6-852b-72b77b37cf1e" (UID: "13e19aaf-7cc3-40a6-852b-72b77b37cf1e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:46:56 crc kubenswrapper[4766]: I1126 00:46:56.773953 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:56 crc kubenswrapper[4766]: I1126 00:46:56.773994 4766 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:56 crc kubenswrapper[4766]: I1126 00:46:56.774008 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:56 crc kubenswrapper[4766]: I1126 00:46:56.774021 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46chp\" (UniqueName: \"kubernetes.io/projected/13e19aaf-7cc3-40a6-852b-72b77b37cf1e-kube-api-access-46chp\") on node \"crc\" DevicePath \"\"" Nov 26 00:46:56 crc kubenswrapper[4766]: I1126 00:46:56.802299 4766 scope.go:117] "RemoveContainer" containerID="10340d467e512bd68e6e6b88a6746642d2c9504b21804f3be63d9c75ddd06bd8" Nov 26 00:46:56 crc kubenswrapper[4766]: E1126 00:46:56.804998 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10340d467e512bd68e6e6b88a6746642d2c9504b21804f3be63d9c75ddd06bd8\": container with ID starting with 10340d467e512bd68e6e6b88a6746642d2c9504b21804f3be63d9c75ddd06bd8 not found: ID does not exist" containerID="10340d467e512bd68e6e6b88a6746642d2c9504b21804f3be63d9c75ddd06bd8" Nov 26 00:46:56 crc kubenswrapper[4766]: I1126 00:46:56.805048 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10340d467e512bd68e6e6b88a6746642d2c9504b21804f3be63d9c75ddd06bd8"} err="failed to get container status \"10340d467e512bd68e6e6b88a6746642d2c9504b21804f3be63d9c75ddd06bd8\": rpc error: code = NotFound desc = could not find container \"10340d467e512bd68e6e6b88a6746642d2c9504b21804f3be63d9c75ddd06bd8\": container with ID starting with 10340d467e512bd68e6e6b88a6746642d2c9504b21804f3be63d9c75ddd06bd8 not found: ID does not exist" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.051239 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-6dfcdbc5c7-znfmt"] Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.060971 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-6dfcdbc5c7-znfmt"] Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.688789 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2t966"] Nov 26 00:46:57 crc kubenswrapper[4766]: E1126 00:46:57.689230 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2046b547-611f-4e98-ba95-a34cd0d45f50" containerName="mariadb-database-create" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.689242 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="2046b547-611f-4e98-ba95-a34cd0d45f50" containerName="mariadb-database-create" Nov 26 00:46:57 crc kubenswrapper[4766]: E1126 00:46:57.689258 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d4b645a-ddab-4a29-8ffa-77e7873591c8" containerName="mariadb-account-create-update" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.689264 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d4b645a-ddab-4a29-8ffa-77e7873591c8" containerName="mariadb-account-create-update" Nov 26 00:46:57 crc kubenswrapper[4766]: E1126 00:46:57.689278 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13e19aaf-7cc3-40a6-852b-72b77b37cf1e" containerName="heat-engine" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.689285 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="13e19aaf-7cc3-40a6-852b-72b77b37cf1e" containerName="heat-engine" Nov 26 00:46:57 crc kubenswrapper[4766]: E1126 00:46:57.689302 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4794066a-0e54-4387-ae01-078abc184684" containerName="mariadb-account-create-update" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.689308 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="4794066a-0e54-4387-ae01-078abc184684" containerName="mariadb-account-create-update" Nov 26 00:46:57 crc kubenswrapper[4766]: E1126 00:46:57.689318 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="751e5ae1-be9d-48e8-8cc1-e08012854203" containerName="mariadb-account-create-update" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.689324 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="751e5ae1-be9d-48e8-8cc1-e08012854203" containerName="mariadb-account-create-update" Nov 26 00:46:57 crc kubenswrapper[4766]: E1126 00:46:57.689340 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="619650be-59d2-47c8-815d-3337c156c83f" containerName="mariadb-database-create" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.689346 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="619650be-59d2-47c8-815d-3337c156c83f" containerName="mariadb-database-create" Nov 26 00:46:57 crc kubenswrapper[4766]: E1126 00:46:57.689359 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b20f822d-71d5-4cb8-823b-8c261af60109" containerName="mariadb-database-create" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.689365 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="b20f822d-71d5-4cb8-823b-8c261af60109" containerName="mariadb-database-create" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.689538 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="4794066a-0e54-4387-ae01-078abc184684" containerName="mariadb-account-create-update" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.689550 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d4b645a-ddab-4a29-8ffa-77e7873591c8" containerName="mariadb-account-create-update" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.689565 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="619650be-59d2-47c8-815d-3337c156c83f" containerName="mariadb-database-create" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.689575 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="b20f822d-71d5-4cb8-823b-8c261af60109" containerName="mariadb-database-create" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.689591 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="13e19aaf-7cc3-40a6-852b-72b77b37cf1e" containerName="heat-engine" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.689607 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="2046b547-611f-4e98-ba95-a34cd0d45f50" containerName="mariadb-database-create" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.689622 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="751e5ae1-be9d-48e8-8cc1-e08012854203" containerName="mariadb-account-create-update" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.691092 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2t966" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.694342 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.694540 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.694837 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-pbtwq" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.703543 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2t966"] Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.798035 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6612de23-8b59-474c-9089-2559ecda8811-config-data\") pod \"nova-cell0-conductor-db-sync-2t966\" (UID: \"6612de23-8b59-474c-9089-2559ecda8811\") " pod="openstack/nova-cell0-conductor-db-sync-2t966" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.798410 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pwgh\" (UniqueName: \"kubernetes.io/projected/6612de23-8b59-474c-9089-2559ecda8811-kube-api-access-9pwgh\") pod \"nova-cell0-conductor-db-sync-2t966\" (UID: \"6612de23-8b59-474c-9089-2559ecda8811\") " pod="openstack/nova-cell0-conductor-db-sync-2t966" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.798513 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6612de23-8b59-474c-9089-2559ecda8811-scripts\") pod \"nova-cell0-conductor-db-sync-2t966\" (UID: \"6612de23-8b59-474c-9089-2559ecda8811\") " pod="openstack/nova-cell0-conductor-db-sync-2t966" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.798609 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6612de23-8b59-474c-9089-2559ecda8811-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-2t966\" (UID: \"6612de23-8b59-474c-9089-2559ecda8811\") " pod="openstack/nova-cell0-conductor-db-sync-2t966" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.841956 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13e19aaf-7cc3-40a6-852b-72b77b37cf1e" path="/var/lib/kubelet/pods/13e19aaf-7cc3-40a6-852b-72b77b37cf1e/volumes" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.900523 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pwgh\" (UniqueName: \"kubernetes.io/projected/6612de23-8b59-474c-9089-2559ecda8811-kube-api-access-9pwgh\") pod \"nova-cell0-conductor-db-sync-2t966\" (UID: \"6612de23-8b59-474c-9089-2559ecda8811\") " pod="openstack/nova-cell0-conductor-db-sync-2t966" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.901014 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6612de23-8b59-474c-9089-2559ecda8811-scripts\") pod \"nova-cell0-conductor-db-sync-2t966\" (UID: \"6612de23-8b59-474c-9089-2559ecda8811\") " pod="openstack/nova-cell0-conductor-db-sync-2t966" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.901926 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6612de23-8b59-474c-9089-2559ecda8811-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-2t966\" (UID: \"6612de23-8b59-474c-9089-2559ecda8811\") " pod="openstack/nova-cell0-conductor-db-sync-2t966" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.902120 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6612de23-8b59-474c-9089-2559ecda8811-config-data\") pod \"nova-cell0-conductor-db-sync-2t966\" (UID: \"6612de23-8b59-474c-9089-2559ecda8811\") " pod="openstack/nova-cell0-conductor-db-sync-2t966" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.915729 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6612de23-8b59-474c-9089-2559ecda8811-scripts\") pod \"nova-cell0-conductor-db-sync-2t966\" (UID: \"6612de23-8b59-474c-9089-2559ecda8811\") " pod="openstack/nova-cell0-conductor-db-sync-2t966" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.915804 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6612de23-8b59-474c-9089-2559ecda8811-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-2t966\" (UID: \"6612de23-8b59-474c-9089-2559ecda8811\") " pod="openstack/nova-cell0-conductor-db-sync-2t966" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.918615 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6612de23-8b59-474c-9089-2559ecda8811-config-data\") pod \"nova-cell0-conductor-db-sync-2t966\" (UID: \"6612de23-8b59-474c-9089-2559ecda8811\") " pod="openstack/nova-cell0-conductor-db-sync-2t966" Nov 26 00:46:57 crc kubenswrapper[4766]: I1126 00:46:57.920166 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pwgh\" (UniqueName: \"kubernetes.io/projected/6612de23-8b59-474c-9089-2559ecda8811-kube-api-access-9pwgh\") pod \"nova-cell0-conductor-db-sync-2t966\" (UID: \"6612de23-8b59-474c-9089-2559ecda8811\") " pod="openstack/nova-cell0-conductor-db-sync-2t966" Nov 26 00:46:58 crc kubenswrapper[4766]: I1126 00:46:58.024070 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2t966" Nov 26 00:46:58 crc kubenswrapper[4766]: I1126 00:46:58.571047 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2t966"] Nov 26 00:46:58 crc kubenswrapper[4766]: I1126 00:46:58.753875 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2t966" event={"ID":"6612de23-8b59-474c-9089-2559ecda8811","Type":"ContainerStarted","Data":"85f798eb9988271fde61fed062ea0947738774995634a8b5ecabca03419f94c3"} Nov 26 00:47:06 crc kubenswrapper[4766]: I1126 00:47:06.864939 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2t966" event={"ID":"6612de23-8b59-474c-9089-2559ecda8811","Type":"ContainerStarted","Data":"9846fffec4a9f47094a0f5e4313eea57458477b1bb7d993864e3aab65ad4b893"} Nov 26 00:47:06 crc kubenswrapper[4766]: I1126 00:47:06.893529 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-2t966" podStartSLOduration=2.065943652 podStartE2EDuration="9.893508015s" podCreationTimestamp="2025-11-26 00:46:57 +0000 UTC" firstStartedPulling="2025-11-26 00:46:58.570643644 +0000 UTC m=+1399.419414074" lastFinishedPulling="2025-11-26 00:47:06.398207997 +0000 UTC m=+1407.246978437" observedRunningTime="2025-11-26 00:47:06.885473463 +0000 UTC m=+1407.734243893" watchObservedRunningTime="2025-11-26 00:47:06.893508015 +0000 UTC m=+1407.742278465" Nov 26 00:47:16 crc kubenswrapper[4766]: I1126 00:47:16.983452 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rwk26"] Nov 26 00:47:16 crc kubenswrapper[4766]: I1126 00:47:16.992497 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rwk26" Nov 26 00:47:16 crc kubenswrapper[4766]: I1126 00:47:16.999475 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rwk26"] Nov 26 00:47:17 crc kubenswrapper[4766]: I1126 00:47:17.045001 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9786e879-74cf-4b41-8d2f-482ba9e8d54f-utilities\") pod \"community-operators-rwk26\" (UID: \"9786e879-74cf-4b41-8d2f-482ba9e8d54f\") " pod="openshift-marketplace/community-operators-rwk26" Nov 26 00:47:17 crc kubenswrapper[4766]: I1126 00:47:17.045355 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwjbj\" (UniqueName: \"kubernetes.io/projected/9786e879-74cf-4b41-8d2f-482ba9e8d54f-kube-api-access-xwjbj\") pod \"community-operators-rwk26\" (UID: \"9786e879-74cf-4b41-8d2f-482ba9e8d54f\") " pod="openshift-marketplace/community-operators-rwk26" Nov 26 00:47:17 crc kubenswrapper[4766]: I1126 00:47:17.045470 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9786e879-74cf-4b41-8d2f-482ba9e8d54f-catalog-content\") pod \"community-operators-rwk26\" (UID: \"9786e879-74cf-4b41-8d2f-482ba9e8d54f\") " pod="openshift-marketplace/community-operators-rwk26" Nov 26 00:47:17 crc kubenswrapper[4766]: I1126 00:47:17.147963 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9786e879-74cf-4b41-8d2f-482ba9e8d54f-catalog-content\") pod \"community-operators-rwk26\" (UID: \"9786e879-74cf-4b41-8d2f-482ba9e8d54f\") " pod="openshift-marketplace/community-operators-rwk26" Nov 26 00:47:17 crc kubenswrapper[4766]: I1126 00:47:17.148149 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9786e879-74cf-4b41-8d2f-482ba9e8d54f-utilities\") pod \"community-operators-rwk26\" (UID: \"9786e879-74cf-4b41-8d2f-482ba9e8d54f\") " pod="openshift-marketplace/community-operators-rwk26" Nov 26 00:47:17 crc kubenswrapper[4766]: I1126 00:47:17.148306 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwjbj\" (UniqueName: \"kubernetes.io/projected/9786e879-74cf-4b41-8d2f-482ba9e8d54f-kube-api-access-xwjbj\") pod \"community-operators-rwk26\" (UID: \"9786e879-74cf-4b41-8d2f-482ba9e8d54f\") " pod="openshift-marketplace/community-operators-rwk26" Nov 26 00:47:17 crc kubenswrapper[4766]: I1126 00:47:17.148466 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9786e879-74cf-4b41-8d2f-482ba9e8d54f-catalog-content\") pod \"community-operators-rwk26\" (UID: \"9786e879-74cf-4b41-8d2f-482ba9e8d54f\") " pod="openshift-marketplace/community-operators-rwk26" Nov 26 00:47:17 crc kubenswrapper[4766]: I1126 00:47:17.148955 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9786e879-74cf-4b41-8d2f-482ba9e8d54f-utilities\") pod \"community-operators-rwk26\" (UID: \"9786e879-74cf-4b41-8d2f-482ba9e8d54f\") " pod="openshift-marketplace/community-operators-rwk26" Nov 26 00:47:17 crc kubenswrapper[4766]: I1126 00:47:17.173562 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwjbj\" (UniqueName: \"kubernetes.io/projected/9786e879-74cf-4b41-8d2f-482ba9e8d54f-kube-api-access-xwjbj\") pod \"community-operators-rwk26\" (UID: \"9786e879-74cf-4b41-8d2f-482ba9e8d54f\") " pod="openshift-marketplace/community-operators-rwk26" Nov 26 00:47:17 crc kubenswrapper[4766]: I1126 00:47:17.346790 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rwk26" Nov 26 00:47:17 crc kubenswrapper[4766]: I1126 00:47:17.856729 4766 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podae9e2d8d-6ce6-490a-8463-a078626b551f"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podae9e2d8d-6ce6-490a-8463-a078626b551f] : Timed out while waiting for systemd to remove kubepods-besteffort-podae9e2d8d_6ce6_490a_8463_a078626b551f.slice" Nov 26 00:47:17 crc kubenswrapper[4766]: E1126 00:47:17.857112 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort podae9e2d8d-6ce6-490a-8463-a078626b551f] : unable to destroy cgroup paths for cgroup [kubepods besteffort podae9e2d8d-6ce6-490a-8463-a078626b551f] : Timed out while waiting for systemd to remove kubepods-besteffort-podae9e2d8d_6ce6_490a_8463_a078626b551f.slice" pod="openstack/ceilometer-0" podUID="ae9e2d8d-6ce6-490a-8463-a078626b551f" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.008815 4766 generic.go:334] "Generic (PLEG): container finished" podID="6612de23-8b59-474c-9089-2559ecda8811" containerID="9846fffec4a9f47094a0f5e4313eea57458477b1bb7d993864e3aab65ad4b893" exitCode=0 Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.008892 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.009579 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2t966" event={"ID":"6612de23-8b59-474c-9089-2559ecda8811","Type":"ContainerDied","Data":"9846fffec4a9f47094a0f5e4313eea57458477b1bb7d993864e3aab65ad4b893"} Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.072686 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.082331 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.096071 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.098787 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.105148 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.105201 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.112316 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.171105 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.171170 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.171206 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mp74x\" (UniqueName: \"kubernetes.io/projected/1424c325-de53-4623-8d9c-7b1ef990c76c-kube-api-access-mp74x\") pod \"ceilometer-0\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.171274 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1424c325-de53-4623-8d9c-7b1ef990c76c-run-httpd\") pod \"ceilometer-0\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.171458 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-config-data\") pod \"ceilometer-0\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.171515 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1424c325-de53-4623-8d9c-7b1ef990c76c-log-httpd\") pod \"ceilometer-0\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.171606 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-scripts\") pod \"ceilometer-0\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.202735 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rwk26"] Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.273011 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-config-data\") pod \"ceilometer-0\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.273340 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1424c325-de53-4623-8d9c-7b1ef990c76c-log-httpd\") pod \"ceilometer-0\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.273394 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-scripts\") pod \"ceilometer-0\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.273416 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.273448 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.273472 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mp74x\" (UniqueName: \"kubernetes.io/projected/1424c325-de53-4623-8d9c-7b1ef990c76c-kube-api-access-mp74x\") pod \"ceilometer-0\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.273521 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1424c325-de53-4623-8d9c-7b1ef990c76c-run-httpd\") pod \"ceilometer-0\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.275444 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1424c325-de53-4623-8d9c-7b1ef990c76c-log-httpd\") pod \"ceilometer-0\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.275546 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1424c325-de53-4623-8d9c-7b1ef990c76c-run-httpd\") pod \"ceilometer-0\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.280606 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-config-data\") pod \"ceilometer-0\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.281119 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-scripts\") pod \"ceilometer-0\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.281205 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.281857 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.303549 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mp74x\" (UniqueName: \"kubernetes.io/projected/1424c325-de53-4623-8d9c-7b1ef990c76c-kube-api-access-mp74x\") pod \"ceilometer-0\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.422085 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:47:18 crc kubenswrapper[4766]: I1126 00:47:18.933347 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:47:19 crc kubenswrapper[4766]: I1126 00:47:19.019510 4766 generic.go:334] "Generic (PLEG): container finished" podID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" containerID="9237bc93c46f5b46707c562126f17465bf9c732a639be92b586142a8d543f870" exitCode=0 Nov 26 00:47:19 crc kubenswrapper[4766]: I1126 00:47:19.019594 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwk26" event={"ID":"9786e879-74cf-4b41-8d2f-482ba9e8d54f","Type":"ContainerDied","Data":"9237bc93c46f5b46707c562126f17465bf9c732a639be92b586142a8d543f870"} Nov 26 00:47:19 crc kubenswrapper[4766]: I1126 00:47:19.019628 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwk26" event={"ID":"9786e879-74cf-4b41-8d2f-482ba9e8d54f","Type":"ContainerStarted","Data":"3f38432d89f6eb0024cd341789f516d200590be31b0c40465d29d2246d4c5698"} Nov 26 00:47:19 crc kubenswrapper[4766]: I1126 00:47:19.020926 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1424c325-de53-4623-8d9c-7b1ef990c76c","Type":"ContainerStarted","Data":"b5d8b6babb32f9dbbe353c4c81bfd8ccfe55484cc2b653c35fe5d9cf68972059"} Nov 26 00:47:19 crc kubenswrapper[4766]: I1126 00:47:19.572378 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2t966" Nov 26 00:47:19 crc kubenswrapper[4766]: I1126 00:47:19.699407 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6612de23-8b59-474c-9089-2559ecda8811-config-data\") pod \"6612de23-8b59-474c-9089-2559ecda8811\" (UID: \"6612de23-8b59-474c-9089-2559ecda8811\") " Nov 26 00:47:19 crc kubenswrapper[4766]: I1126 00:47:19.699624 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6612de23-8b59-474c-9089-2559ecda8811-scripts\") pod \"6612de23-8b59-474c-9089-2559ecda8811\" (UID: \"6612de23-8b59-474c-9089-2559ecda8811\") " Nov 26 00:47:19 crc kubenswrapper[4766]: I1126 00:47:19.699702 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pwgh\" (UniqueName: \"kubernetes.io/projected/6612de23-8b59-474c-9089-2559ecda8811-kube-api-access-9pwgh\") pod \"6612de23-8b59-474c-9089-2559ecda8811\" (UID: \"6612de23-8b59-474c-9089-2559ecda8811\") " Nov 26 00:47:19 crc kubenswrapper[4766]: I1126 00:47:19.699732 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6612de23-8b59-474c-9089-2559ecda8811-combined-ca-bundle\") pod \"6612de23-8b59-474c-9089-2559ecda8811\" (UID: \"6612de23-8b59-474c-9089-2559ecda8811\") " Nov 26 00:47:19 crc kubenswrapper[4766]: I1126 00:47:19.707795 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6612de23-8b59-474c-9089-2559ecda8811-scripts" (OuterVolumeSpecName: "scripts") pod "6612de23-8b59-474c-9089-2559ecda8811" (UID: "6612de23-8b59-474c-9089-2559ecda8811"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:47:19 crc kubenswrapper[4766]: I1126 00:47:19.708019 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6612de23-8b59-474c-9089-2559ecda8811-kube-api-access-9pwgh" (OuterVolumeSpecName: "kube-api-access-9pwgh") pod "6612de23-8b59-474c-9089-2559ecda8811" (UID: "6612de23-8b59-474c-9089-2559ecda8811"). InnerVolumeSpecName "kube-api-access-9pwgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:47:19 crc kubenswrapper[4766]: I1126 00:47:19.742411 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6612de23-8b59-474c-9089-2559ecda8811-config-data" (OuterVolumeSpecName: "config-data") pod "6612de23-8b59-474c-9089-2559ecda8811" (UID: "6612de23-8b59-474c-9089-2559ecda8811"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:47:19 crc kubenswrapper[4766]: I1126 00:47:19.749856 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6612de23-8b59-474c-9089-2559ecda8811-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6612de23-8b59-474c-9089-2559ecda8811" (UID: "6612de23-8b59-474c-9089-2559ecda8811"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:47:19 crc kubenswrapper[4766]: I1126 00:47:19.803456 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pwgh\" (UniqueName: \"kubernetes.io/projected/6612de23-8b59-474c-9089-2559ecda8811-kube-api-access-9pwgh\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:19 crc kubenswrapper[4766]: I1126 00:47:19.803488 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6612de23-8b59-474c-9089-2559ecda8811-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:19 crc kubenswrapper[4766]: I1126 00:47:19.803499 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6612de23-8b59-474c-9089-2559ecda8811-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:19 crc kubenswrapper[4766]: I1126 00:47:19.803509 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6612de23-8b59-474c-9089-2559ecda8811-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:19 crc kubenswrapper[4766]: I1126 00:47:19.843174 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae9e2d8d-6ce6-490a-8463-a078626b551f" path="/var/lib/kubelet/pods/ae9e2d8d-6ce6-490a-8463-a078626b551f/volumes" Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.032932 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1424c325-de53-4623-8d9c-7b1ef990c76c","Type":"ContainerStarted","Data":"882fe26d38afc2ce068558b9d91d2b7a05655774be1f6029d3621e48ad017258"} Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.036745 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwk26" event={"ID":"9786e879-74cf-4b41-8d2f-482ba9e8d54f","Type":"ContainerStarted","Data":"1d21e1aedc1c9ddc220987318fdec667060660941c075ff804b51171730506de"} Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.044021 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-2t966" event={"ID":"6612de23-8b59-474c-9089-2559ecda8811","Type":"ContainerDied","Data":"85f798eb9988271fde61fed062ea0947738774995634a8b5ecabca03419f94c3"} Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.044063 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85f798eb9988271fde61fed062ea0947738774995634a8b5ecabca03419f94c3" Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.044129 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-2t966" Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.144544 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 26 00:47:20 crc kubenswrapper[4766]: E1126 00:47:20.145172 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6612de23-8b59-474c-9089-2559ecda8811" containerName="nova-cell0-conductor-db-sync" Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.145198 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6612de23-8b59-474c-9089-2559ecda8811" containerName="nova-cell0-conductor-db-sync" Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.145458 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="6612de23-8b59-474c-9089-2559ecda8811" containerName="nova-cell0-conductor-db-sync" Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.146364 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.148637 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.148811 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-pbtwq" Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.156082 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.210990 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b33aa993-e31f-402f-9574-33bd30a942c4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b33aa993-e31f-402f-9574-33bd30a942c4\") " pod="openstack/nova-cell0-conductor-0" Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.211266 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b33aa993-e31f-402f-9574-33bd30a942c4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b33aa993-e31f-402f-9574-33bd30a942c4\") " pod="openstack/nova-cell0-conductor-0" Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.211459 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r4pg\" (UniqueName: \"kubernetes.io/projected/b33aa993-e31f-402f-9574-33bd30a942c4-kube-api-access-6r4pg\") pod \"nova-cell0-conductor-0\" (UID: \"b33aa993-e31f-402f-9574-33bd30a942c4\") " pod="openstack/nova-cell0-conductor-0" Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.317884 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r4pg\" (UniqueName: \"kubernetes.io/projected/b33aa993-e31f-402f-9574-33bd30a942c4-kube-api-access-6r4pg\") pod \"nova-cell0-conductor-0\" (UID: \"b33aa993-e31f-402f-9574-33bd30a942c4\") " pod="openstack/nova-cell0-conductor-0" Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.317988 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b33aa993-e31f-402f-9574-33bd30a942c4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b33aa993-e31f-402f-9574-33bd30a942c4\") " pod="openstack/nova-cell0-conductor-0" Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.318111 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b33aa993-e31f-402f-9574-33bd30a942c4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b33aa993-e31f-402f-9574-33bd30a942c4\") " pod="openstack/nova-cell0-conductor-0" Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.324265 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b33aa993-e31f-402f-9574-33bd30a942c4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b33aa993-e31f-402f-9574-33bd30a942c4\") " pod="openstack/nova-cell0-conductor-0" Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.334321 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b33aa993-e31f-402f-9574-33bd30a942c4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b33aa993-e31f-402f-9574-33bd30a942c4\") " pod="openstack/nova-cell0-conductor-0" Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.345076 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r4pg\" (UniqueName: \"kubernetes.io/projected/b33aa993-e31f-402f-9574-33bd30a942c4-kube-api-access-6r4pg\") pod \"nova-cell0-conductor-0\" (UID: \"b33aa993-e31f-402f-9574-33bd30a942c4\") " pod="openstack/nova-cell0-conductor-0" Nov 26 00:47:20 crc kubenswrapper[4766]: I1126 00:47:20.471827 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 26 00:47:21 crc kubenswrapper[4766]: I1126 00:47:21.041430 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 26 00:47:21 crc kubenswrapper[4766]: I1126 00:47:21.056918 4766 generic.go:334] "Generic (PLEG): container finished" podID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" containerID="1d21e1aedc1c9ddc220987318fdec667060660941c075ff804b51171730506de" exitCode=0 Nov 26 00:47:21 crc kubenswrapper[4766]: I1126 00:47:21.057000 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwk26" event={"ID":"9786e879-74cf-4b41-8d2f-482ba9e8d54f","Type":"ContainerDied","Data":"1d21e1aedc1c9ddc220987318fdec667060660941c075ff804b51171730506de"} Nov 26 00:47:21 crc kubenswrapper[4766]: I1126 00:47:21.057029 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwk26" event={"ID":"9786e879-74cf-4b41-8d2f-482ba9e8d54f","Type":"ContainerStarted","Data":"bceafd245f3339232ecc96719a07509eae7eaa5d91b5564c4c361ab21fbea67c"} Nov 26 00:47:21 crc kubenswrapper[4766]: I1126 00:47:21.058777 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b33aa993-e31f-402f-9574-33bd30a942c4","Type":"ContainerStarted","Data":"d45b11f819a53cbd8bb6258a86b33653b6c863af8fbf6d36413aa44af8bd16f2"} Nov 26 00:47:21 crc kubenswrapper[4766]: I1126 00:47:21.060930 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1424c325-de53-4623-8d9c-7b1ef990c76c","Type":"ContainerStarted","Data":"5666004e40f54d4aa8bc89c9c30f5e4a5bee32fb58429886e8ec14e9d09409b3"} Nov 26 00:47:22 crc kubenswrapper[4766]: I1126 00:47:22.073475 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1424c325-de53-4623-8d9c-7b1ef990c76c","Type":"ContainerStarted","Data":"b38fcdb7242e91ffaff5a2634e987f781d5930f298f5926742911caae74e04f3"} Nov 26 00:47:22 crc kubenswrapper[4766]: I1126 00:47:22.076551 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b33aa993-e31f-402f-9574-33bd30a942c4","Type":"ContainerStarted","Data":"9baffd183a09d4b0c6b993f4545e4b44d2b011400e4d8bedee2390c58514e4a6"} Nov 26 00:47:22 crc kubenswrapper[4766]: I1126 00:47:22.076883 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 26 00:47:22 crc kubenswrapper[4766]: I1126 00:47:22.105025 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rwk26" podStartSLOduration=4.698835557 podStartE2EDuration="6.104997699s" podCreationTimestamp="2025-11-26 00:47:16 +0000 UTC" firstStartedPulling="2025-11-26 00:47:19.021127728 +0000 UTC m=+1419.869898158" lastFinishedPulling="2025-11-26 00:47:20.42728987 +0000 UTC m=+1421.276060300" observedRunningTime="2025-11-26 00:47:22.093565503 +0000 UTC m=+1422.942335943" watchObservedRunningTime="2025-11-26 00:47:22.104997699 +0000 UTC m=+1422.953768129" Nov 26 00:47:22 crc kubenswrapper[4766]: I1126 00:47:22.118817 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.118793435 podStartE2EDuration="2.118793435s" podCreationTimestamp="2025-11-26 00:47:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:47:22.114147239 +0000 UTC m=+1422.962917679" watchObservedRunningTime="2025-11-26 00:47:22.118793435 +0000 UTC m=+1422.967563865" Nov 26 00:47:23 crc kubenswrapper[4766]: I1126 00:47:23.088542 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1424c325-de53-4623-8d9c-7b1ef990c76c","Type":"ContainerStarted","Data":"f32180c2d95fb156d27fc02459a5ea18ccf510d67a1536dde55da7e9d5262819"} Nov 26 00:47:23 crc kubenswrapper[4766]: I1126 00:47:23.110859 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.3292087399999999 podStartE2EDuration="5.110837274s" podCreationTimestamp="2025-11-26 00:47:18 +0000 UTC" firstStartedPulling="2025-11-26 00:47:18.948457336 +0000 UTC m=+1419.797227766" lastFinishedPulling="2025-11-26 00:47:22.73008587 +0000 UTC m=+1423.578856300" observedRunningTime="2025-11-26 00:47:23.10866414 +0000 UTC m=+1423.957434570" watchObservedRunningTime="2025-11-26 00:47:23.110837274 +0000 UTC m=+1423.959607714" Nov 26 00:47:23 crc kubenswrapper[4766]: I1126 00:47:23.649862 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:47:24 crc kubenswrapper[4766]: I1126 00:47:24.103884 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 26 00:47:25 crc kubenswrapper[4766]: I1126 00:47:25.112702 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" containerName="ceilometer-central-agent" containerID="cri-o://882fe26d38afc2ce068558b9d91d2b7a05655774be1f6029d3621e48ad017258" gracePeriod=30 Nov 26 00:47:25 crc kubenswrapper[4766]: I1126 00:47:25.112788 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" containerName="proxy-httpd" containerID="cri-o://f32180c2d95fb156d27fc02459a5ea18ccf510d67a1536dde55da7e9d5262819" gracePeriod=30 Nov 26 00:47:25 crc kubenswrapper[4766]: I1126 00:47:25.112821 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" containerName="sg-core" containerID="cri-o://b38fcdb7242e91ffaff5a2634e987f781d5930f298f5926742911caae74e04f3" gracePeriod=30 Nov 26 00:47:25 crc kubenswrapper[4766]: I1126 00:47:25.112851 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" containerName="ceilometer-notification-agent" containerID="cri-o://5666004e40f54d4aa8bc89c9c30f5e4a5bee32fb58429886e8ec14e9d09409b3" gracePeriod=30 Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.125757 4766 generic.go:334] "Generic (PLEG): container finished" podID="1424c325-de53-4623-8d9c-7b1ef990c76c" containerID="f32180c2d95fb156d27fc02459a5ea18ccf510d67a1536dde55da7e9d5262819" exitCode=0 Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.126076 4766 generic.go:334] "Generic (PLEG): container finished" podID="1424c325-de53-4623-8d9c-7b1ef990c76c" containerID="b38fcdb7242e91ffaff5a2634e987f781d5930f298f5926742911caae74e04f3" exitCode=2 Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.126088 4766 generic.go:334] "Generic (PLEG): container finished" podID="1424c325-de53-4623-8d9c-7b1ef990c76c" containerID="5666004e40f54d4aa8bc89c9c30f5e4a5bee32fb58429886e8ec14e9d09409b3" exitCode=0 Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.125838 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1424c325-de53-4623-8d9c-7b1ef990c76c","Type":"ContainerDied","Data":"f32180c2d95fb156d27fc02459a5ea18ccf510d67a1536dde55da7e9d5262819"} Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.126131 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1424c325-de53-4623-8d9c-7b1ef990c76c","Type":"ContainerDied","Data":"b38fcdb7242e91ffaff5a2634e987f781d5930f298f5926742911caae74e04f3"} Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.126151 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1424c325-de53-4623-8d9c-7b1ef990c76c","Type":"ContainerDied","Data":"5666004e40f54d4aa8bc89c9c30f5e4a5bee32fb58429886e8ec14e9d09409b3"} Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.162582 4766 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.162918 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186" gracePeriod=15 Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.162934 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b" gracePeriod=15 Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.162950 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5" gracePeriod=15 Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.162935 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319" gracePeriod=15 Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.163006 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36" gracePeriod=15 Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.167280 4766 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 26 00:47:26 crc kubenswrapper[4766]: E1126 00:47:26.168239 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.175018 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 26 00:47:26 crc kubenswrapper[4766]: E1126 00:47:26.175123 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.175189 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 26 00:47:26 crc kubenswrapper[4766]: E1126 00:47:26.175248 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.175305 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 26 00:47:26 crc kubenswrapper[4766]: E1126 00:47:26.175387 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.175446 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 26 00:47:26 crc kubenswrapper[4766]: E1126 00:47:26.175963 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.176046 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 26 00:47:26 crc kubenswrapper[4766]: E1126 00:47:26.176123 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.176192 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 26 00:47:26 crc kubenswrapper[4766]: E1126 00:47:26.176299 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.176380 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.176892 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.176988 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.177058 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.177121 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.177190 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.177255 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.181169 4766 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.182532 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: E1126 00:47:26.232509 4766 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.32:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.254481 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.254551 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.254581 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.254615 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.254696 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.254756 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.254802 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.254834 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.357122 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.357404 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.357563 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.357699 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.357817 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.357961 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.358101 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.358240 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.358340 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.358777 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.358931 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.359046 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.359103 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.359118 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.359133 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.359156 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: I1126 00:47:26.534040 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 00:47:26 crc kubenswrapper[4766]: E1126 00:47:26.566103 4766 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.32:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187b6804436575e9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-26 00:47:26.565561833 +0000 UTC m=+1427.414332263,LastTimestamp:2025-11-26 00:47:26.565561833 +0000 UTC m=+1427.414332263,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 26 00:47:27 crc kubenswrapper[4766]: E1126 00:47:27.027504 4766 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.32:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187b6804436575e9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-26 00:47:26.565561833 +0000 UTC m=+1427.414332263,LastTimestamp:2025-11-26 00:47:26.565561833 +0000 UTC m=+1427.414332263,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 26 00:47:27 crc kubenswrapper[4766]: I1126 00:47:27.146142 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 26 00:47:27 crc kubenswrapper[4766]: I1126 00:47:27.152181 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 26 00:47:27 crc kubenswrapper[4766]: I1126 00:47:27.153964 4766 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b" exitCode=0 Nov 26 00:47:27 crc kubenswrapper[4766]: I1126 00:47:27.153987 4766 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5" exitCode=0 Nov 26 00:47:27 crc kubenswrapper[4766]: I1126 00:47:27.153996 4766 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319" exitCode=0 Nov 26 00:47:27 crc kubenswrapper[4766]: I1126 00:47:27.154003 4766 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36" exitCode=2 Nov 26 00:47:27 crc kubenswrapper[4766]: I1126 00:47:27.154061 4766 scope.go:117] "RemoveContainer" containerID="34420959e6b5df9969c22215565521e36ddf07a944c15a20342dad5bfbb69115" Nov 26 00:47:27 crc kubenswrapper[4766]: I1126 00:47:27.156122 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"27937a6f6f497f5f7c3f3a89a54f4857db004c27ee4f74df829476bd1a716a89"} Nov 26 00:47:27 crc kubenswrapper[4766]: I1126 00:47:27.156168 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"add7c42a346acd8d772af3a4ba488def09ece74ce7425d4686a52fae27c7b583"} Nov 26 00:47:27 crc kubenswrapper[4766]: E1126 00:47:27.156884 4766 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.32:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 00:47:27 crc kubenswrapper[4766]: I1126 00:47:27.157923 4766 generic.go:334] "Generic (PLEG): container finished" podID="2bba0b1b-4848-4452-a535-c0cf5d13b30b" containerID="bf1673580fa8e1a522a2bf6426b11db0235234df253170289ee2b6d16def7445" exitCode=0 Nov 26 00:47:27 crc kubenswrapper[4766]: I1126 00:47:27.157956 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2bba0b1b-4848-4452-a535-c0cf5d13b30b","Type":"ContainerDied","Data":"bf1673580fa8e1a522a2bf6426b11db0235234df253170289ee2b6d16def7445"} Nov 26 00:47:27 crc kubenswrapper[4766]: I1126 00:47:27.347461 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rwk26" Nov 26 00:47:27 crc kubenswrapper[4766]: I1126 00:47:27.347787 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rwk26" Nov 26 00:47:27 crc kubenswrapper[4766]: I1126 00:47:27.403004 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rwk26" Nov 26 00:47:27 crc kubenswrapper[4766]: I1126 00:47:27.442303 4766 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 26 00:47:27 crc kubenswrapper[4766]: I1126 00:47:27.442572 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 26 00:47:28 crc kubenswrapper[4766]: I1126 00:47:28.171458 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 26 00:47:28 crc kubenswrapper[4766]: I1126 00:47:28.359595 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rwk26" Nov 26 00:47:28 crc kubenswrapper[4766]: E1126 00:47:28.865400 4766 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.32:6443: connect: connection refused" pod="openshift-image-registry/image-registry-66df7c8f76-s29xf" volumeName="registry-storage" Nov 26 00:47:28 crc kubenswrapper[4766]: I1126 00:47:28.865480 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.032335 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2bba0b1b-4848-4452-a535-c0cf5d13b30b-var-lock\") pod \"2bba0b1b-4848-4452-a535-c0cf5d13b30b\" (UID: \"2bba0b1b-4848-4452-a535-c0cf5d13b30b\") " Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.032461 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2bba0b1b-4848-4452-a535-c0cf5d13b30b-kubelet-dir\") pod \"2bba0b1b-4848-4452-a535-c0cf5d13b30b\" (UID: \"2bba0b1b-4848-4452-a535-c0cf5d13b30b\") " Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.032463 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2bba0b1b-4848-4452-a535-c0cf5d13b30b-var-lock" (OuterVolumeSpecName: "var-lock") pod "2bba0b1b-4848-4452-a535-c0cf5d13b30b" (UID: "2bba0b1b-4848-4452-a535-c0cf5d13b30b"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.032548 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2bba0b1b-4848-4452-a535-c0cf5d13b30b-kube-api-access\") pod \"2bba0b1b-4848-4452-a535-c0cf5d13b30b\" (UID: \"2bba0b1b-4848-4452-a535-c0cf5d13b30b\") " Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.032593 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2bba0b1b-4848-4452-a535-c0cf5d13b30b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2bba0b1b-4848-4452-a535-c0cf5d13b30b" (UID: "2bba0b1b-4848-4452-a535-c0cf5d13b30b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.033036 4766 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2bba0b1b-4848-4452-a535-c0cf5d13b30b-var-lock\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.033052 4766 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2bba0b1b-4848-4452-a535-c0cf5d13b30b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.039092 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bba0b1b-4848-4452-a535-c0cf5d13b30b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2bba0b1b-4848-4452-a535-c0cf5d13b30b" (UID: "2bba0b1b-4848-4452-a535-c0cf5d13b30b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.127551 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.128568 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.134453 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2bba0b1b-4848-4452-a535-c0cf5d13b30b-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.184267 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.185825 4766 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186" exitCode=0 Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.185896 4766 scope.go:117] "RemoveContainer" containerID="22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.185932 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.189663 4766 generic.go:334] "Generic (PLEG): container finished" podID="1424c325-de53-4623-8d9c-7b1ef990c76c" containerID="882fe26d38afc2ce068558b9d91d2b7a05655774be1f6029d3621e48ad017258" exitCode=0 Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.189680 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1424c325-de53-4623-8d9c-7b1ef990c76c","Type":"ContainerDied","Data":"882fe26d38afc2ce068558b9d91d2b7a05655774be1f6029d3621e48ad017258"} Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.191910 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2bba0b1b-4848-4452-a535-c0cf5d13b30b","Type":"ContainerDied","Data":"439388a00f68c195095ae7ba1fdbfc797d7bcd3ad14451123a18decdb7c2c6e0"} Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.191947 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="439388a00f68c195095ae7ba1fdbfc797d7bcd3ad14451123a18decdb7c2c6e0" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.191991 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.215989 4766 scope.go:117] "RemoveContainer" containerID="c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.235201 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.235457 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.235455 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.235506 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.235524 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.235611 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.236165 4766 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.236199 4766 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.236211 4766 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.242280 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.248400 4766 scope.go:117] "RemoveContainer" containerID="6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.273049 4766 scope.go:117] "RemoveContainer" containerID="0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.298038 4766 scope.go:117] "RemoveContainer" containerID="06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.325527 4766 scope.go:117] "RemoveContainer" containerID="90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.337106 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-config-data\") pod \"1424c325-de53-4623-8d9c-7b1ef990c76c\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.337470 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-combined-ca-bundle\") pod \"1424c325-de53-4623-8d9c-7b1ef990c76c\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.337726 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1424c325-de53-4623-8d9c-7b1ef990c76c-run-httpd\") pod \"1424c325-de53-4623-8d9c-7b1ef990c76c\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.337834 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-scripts\") pod \"1424c325-de53-4623-8d9c-7b1ef990c76c\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.337951 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mp74x\" (UniqueName: \"kubernetes.io/projected/1424c325-de53-4623-8d9c-7b1ef990c76c-kube-api-access-mp74x\") pod \"1424c325-de53-4623-8d9c-7b1ef990c76c\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.338030 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-sg-core-conf-yaml\") pod \"1424c325-de53-4623-8d9c-7b1ef990c76c\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.338090 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1424c325-de53-4623-8d9c-7b1ef990c76c-log-httpd\") pod \"1424c325-de53-4623-8d9c-7b1ef990c76c\" (UID: \"1424c325-de53-4623-8d9c-7b1ef990c76c\") " Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.338755 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1424c325-de53-4623-8d9c-7b1ef990c76c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1424c325-de53-4623-8d9c-7b1ef990c76c" (UID: "1424c325-de53-4623-8d9c-7b1ef990c76c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.338933 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1424c325-de53-4623-8d9c-7b1ef990c76c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1424c325-de53-4623-8d9c-7b1ef990c76c" (UID: "1424c325-de53-4623-8d9c-7b1ef990c76c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.339131 4766 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1424c325-de53-4623-8d9c-7b1ef990c76c-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.339159 4766 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1424c325-de53-4623-8d9c-7b1ef990c76c-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.342425 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-scripts" (OuterVolumeSpecName: "scripts") pod "1424c325-de53-4623-8d9c-7b1ef990c76c" (UID: "1424c325-de53-4623-8d9c-7b1ef990c76c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.344329 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1424c325-de53-4623-8d9c-7b1ef990c76c-kube-api-access-mp74x" (OuterVolumeSpecName: "kube-api-access-mp74x") pod "1424c325-de53-4623-8d9c-7b1ef990c76c" (UID: "1424c325-de53-4623-8d9c-7b1ef990c76c"). InnerVolumeSpecName "kube-api-access-mp74x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.359317 4766 scope.go:117] "RemoveContainer" containerID="22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b" Nov 26 00:47:29 crc kubenswrapper[4766]: E1126 00:47:29.359804 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\": container with ID starting with 22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b not found: ID does not exist" containerID="22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.359835 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b"} err="failed to get container status \"22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\": rpc error: code = NotFound desc = could not find container \"22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b\": container with ID starting with 22f8a2d301b0ac20fb2254cff7b9dd4163296ed8cf1dc745b3bbe9feef300c5b not found: ID does not exist" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.359855 4766 scope.go:117] "RemoveContainer" containerID="c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5" Nov 26 00:47:29 crc kubenswrapper[4766]: E1126 00:47:29.360024 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\": container with ID starting with c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5 not found: ID does not exist" containerID="c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.360045 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5"} err="failed to get container status \"c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\": rpc error: code = NotFound desc = could not find container \"c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5\": container with ID starting with c5a947e9b251d4fba6e8401a3de9f6aef1060be0a82bdd478e32b8a1f5817ea5 not found: ID does not exist" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.360057 4766 scope.go:117] "RemoveContainer" containerID="6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319" Nov 26 00:47:29 crc kubenswrapper[4766]: E1126 00:47:29.360216 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\": container with ID starting with 6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319 not found: ID does not exist" containerID="6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.360241 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319"} err="failed to get container status \"6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\": rpc error: code = NotFound desc = could not find container \"6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319\": container with ID starting with 6be3702f5cbc8f8e0f8bd7166e45a1765593a4924fa524e3dfe4be272c781319 not found: ID does not exist" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.360254 4766 scope.go:117] "RemoveContainer" containerID="0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36" Nov 26 00:47:29 crc kubenswrapper[4766]: E1126 00:47:29.360416 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\": container with ID starting with 0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36 not found: ID does not exist" containerID="0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.360438 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36"} err="failed to get container status \"0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\": rpc error: code = NotFound desc = could not find container \"0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36\": container with ID starting with 0b4cdd3d570c5f0487ffb478810b17b703fd693a36abb34ed14292b2f88faf36 not found: ID does not exist" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.360451 4766 scope.go:117] "RemoveContainer" containerID="06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186" Nov 26 00:47:29 crc kubenswrapper[4766]: E1126 00:47:29.360613 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\": container with ID starting with 06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186 not found: ID does not exist" containerID="06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.360678 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186"} err="failed to get container status \"06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\": rpc error: code = NotFound desc = could not find container \"06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186\": container with ID starting with 06ddc74c12461ca85da688eee37039248c48bef834b0b660edeb5811e3c80186 not found: ID does not exist" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.360698 4766 scope.go:117] "RemoveContainer" containerID="90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72" Nov 26 00:47:29 crc kubenswrapper[4766]: E1126 00:47:29.361002 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\": container with ID starting with 90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72 not found: ID does not exist" containerID="90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.361024 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72"} err="failed to get container status \"90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\": rpc error: code = NotFound desc = could not find container \"90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72\": container with ID starting with 90fe354f4099b9fd58872e4d0f1e4a25775b048f274384658d33ece63827fd72 not found: ID does not exist" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.374477 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1424c325-de53-4623-8d9c-7b1ef990c76c" (UID: "1424c325-de53-4623-8d9c-7b1ef990c76c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.422969 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1424c325-de53-4623-8d9c-7b1ef990c76c" (UID: "1424c325-de53-4623-8d9c-7b1ef990c76c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.440925 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mp74x\" (UniqueName: \"kubernetes.io/projected/1424c325-de53-4623-8d9c-7b1ef990c76c-kube-api-access-mp74x\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.440957 4766 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.440969 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.440979 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.447350 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-config-data" (OuterVolumeSpecName: "config-data") pod "1424c325-de53-4623-8d9c-7b1ef990c76c" (UID: "1424c325-de53-4623-8d9c-7b1ef990c76c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.543121 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1424c325-de53-4623-8d9c-7b1ef990c76c-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:29 crc kubenswrapper[4766]: I1126 00:47:29.890556 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 26 00:47:30 crc kubenswrapper[4766]: I1126 00:47:30.208363 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1424c325-de53-4623-8d9c-7b1ef990c76c","Type":"ContainerDied","Data":"b5d8b6babb32f9dbbe353c4c81bfd8ccfe55484cc2b653c35fe5d9cf68972059"} Nov 26 00:47:30 crc kubenswrapper[4766]: I1126 00:47:30.208398 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:47:30 crc kubenswrapper[4766]: I1126 00:47:30.208742 4766 scope.go:117] "RemoveContainer" containerID="f32180c2d95fb156d27fc02459a5ea18ccf510d67a1536dde55da7e9d5262819" Nov 26 00:47:30 crc kubenswrapper[4766]: I1126 00:47:30.231379 4766 scope.go:117] "RemoveContainer" containerID="b38fcdb7242e91ffaff5a2634e987f781d5930f298f5926742911caae74e04f3" Nov 26 00:47:30 crc kubenswrapper[4766]: I1126 00:47:30.251146 4766 scope.go:117] "RemoveContainer" containerID="5666004e40f54d4aa8bc89c9c30f5e4a5bee32fb58429886e8ec14e9d09409b3" Nov 26 00:47:30 crc kubenswrapper[4766]: I1126 00:47:30.269854 4766 scope.go:117] "RemoveContainer" containerID="882fe26d38afc2ce068558b9d91d2b7a05655774be1f6029d3621e48ad017258" Nov 26 00:47:30 crc kubenswrapper[4766]: I1126 00:47:30.425436 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" containerName="kube-state-metrics" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 26 00:47:30 crc kubenswrapper[4766]: I1126 00:47:30.513374 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 26 00:47:30 crc kubenswrapper[4766]: E1126 00:47:30.905538 4766 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openstack/swift-swift-storage-0: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/persistentvolumeclaims/swift-swift-storage-0\": dial tcp 38.102.83.32:6443: connect: connection refused" pod="openstack/swift-storage-0" volumeName="swift" Nov 26 00:47:31 crc kubenswrapper[4766]: I1126 00:47:31.186522 4766 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:31 crc kubenswrapper[4766]: I1126 00:47:31.187075 4766 status_manager.go:851] "Failed to get status for pod" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" pod="openstack/ceilometer-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/ceilometer-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:31 crc kubenswrapper[4766]: I1126 00:47:31.187292 4766 status_manager.go:851] "Failed to get status for pod" podUID="b33aa993-e31f-402f-9574-33bd30a942c4" pod="openstack/nova-cell0-conductor-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/nova-cell0-conductor-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:31 crc kubenswrapper[4766]: I1126 00:47:31.187478 4766 status_manager.go:851] "Failed to get status for pod" podUID="2bba0b1b-4848-4452-a535-c0cf5d13b30b" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:31 crc kubenswrapper[4766]: I1126 00:47:31.187644 4766 status_manager.go:851] "Failed to get status for pod" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" pod="openshift-marketplace/community-operators-rwk26" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rwk26\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:31 crc kubenswrapper[4766]: I1126 00:47:31.194920 4766 status_manager.go:851] "Failed to get status for pod" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" pod="openstack/ceilometer-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/ceilometer-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:31 crc kubenswrapper[4766]: I1126 00:47:31.195407 4766 status_manager.go:851] "Failed to get status for pod" podUID="b33aa993-e31f-402f-9574-33bd30a942c4" pod="openstack/nova-cell0-conductor-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/nova-cell0-conductor-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:31 crc kubenswrapper[4766]: I1126 00:47:31.195912 4766 status_manager.go:851] "Failed to get status for pod" podUID="2bba0b1b-4848-4452-a535-c0cf5d13b30b" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:31 crc kubenswrapper[4766]: I1126 00:47:31.197727 4766 status_manager.go:851] "Failed to get status for pod" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" pod="openshift-marketplace/community-operators-rwk26" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rwk26\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:32 crc kubenswrapper[4766]: E1126 00:47:32.177806 4766 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:32 crc kubenswrapper[4766]: E1126 00:47:32.178269 4766 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:32 crc kubenswrapper[4766]: E1126 00:47:32.178737 4766 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:32 crc kubenswrapper[4766]: E1126 00:47:32.179053 4766 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:32 crc kubenswrapper[4766]: E1126 00:47:32.179365 4766 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:32 crc kubenswrapper[4766]: I1126 00:47:32.179392 4766 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 26 00:47:32 crc kubenswrapper[4766]: E1126 00:47:32.179738 4766 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="200ms" Nov 26 00:47:32 crc kubenswrapper[4766]: E1126 00:47:32.380351 4766 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="400ms" Nov 26 00:47:32 crc kubenswrapper[4766]: E1126 00:47:32.590418 4766 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/f384c0de27a5e45c018a741a804b56020338eaf4d4069538edb27512b190da4d/diff" to get inode usage: stat /var/lib/containers/storage/overlay/f384c0de27a5e45c018a741a804b56020338eaf4d4069538edb27512b190da4d/diff: no such file or directory, extraDiskErr: Nov 26 00:47:32 crc kubenswrapper[4766]: E1126 00:47:32.677677 4766 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/cd640083bb642fc0dfbc85de1ed474e680bcb87b662979b3928db63b95181d56/diff" to get inode usage: stat /var/lib/containers/storage/overlay/cd640083bb642fc0dfbc85de1ed474e680bcb87b662979b3928db63b95181d56/diff: no such file or directory, extraDiskErr: Nov 26 00:47:32 crc kubenswrapper[4766]: E1126 00:47:32.781361 4766 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="800ms" Nov 26 00:47:32 crc kubenswrapper[4766]: E1126 00:47:32.844297 4766 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openstack/persistence-rabbitmq-cell1-server-0: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/persistentvolumeclaims/persistence-rabbitmq-cell1-server-0\": dial tcp 38.102.83.32:6443: connect: connection refused" pod="openstack/rabbitmq-cell1-server-0" volumeName="persistence" Nov 26 00:47:33 crc kubenswrapper[4766]: E1126 00:47:33.180356 4766 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/f0dfb7cf03a608bffd8fb0086a98236d110b79fabac4b3a33b945d32d95fca2c/diff" to get inode usage: stat /var/lib/containers/storage/overlay/f0dfb7cf03a608bffd8fb0086a98236d110b79fabac4b3a33b945d32d95fca2c/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_heat-engine-6dfcdbc5c7-znfmt_13e19aaf-7cc3-40a6-852b-72b77b37cf1e/heat-engine/0.log" to get inode usage: stat /var/log/pods/openstack_heat-engine-6dfcdbc5c7-znfmt_13e19aaf-7cc3-40a6-852b-72b77b37cf1e/heat-engine/0.log: no such file or directory Nov 26 00:47:33 crc kubenswrapper[4766]: E1126 00:47:33.582159 4766 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="1.6s" Nov 26 00:47:33 crc kubenswrapper[4766]: E1126 00:47:33.864764 4766 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-logging/storage-logging-loki-compactor-0: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-logging/persistentvolumeclaims/storage-logging-loki-compactor-0\": dial tcp 38.102.83.32:6443: connect: connection refused" pod="openshift-logging/logging-loki-compactor-0" volumeName="storage" Nov 26 00:47:34 crc kubenswrapper[4766]: E1126 00:47:34.614922 4766 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/570544ee2e7a7844e3077906dc5636d187b59b4eb1b7c2310b37d1a15d8975fc/diff" to get inode usage: stat /var/lib/containers/storage/overlay/570544ee2e7a7844e3077906dc5636d187b59b4eb1b7c2310b37d1a15d8975fc/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_ceilometer-0_ae9e2d8d-6ce6-490a-8463-a078626b551f/ceilometer-central-agent/0.log" to get inode usage: stat /var/log/pods/openstack_ceilometer-0_ae9e2d8d-6ce6-490a-8463-a078626b551f/ceilometer-central-agent/0.log: no such file or directory Nov 26 00:47:35 crc kubenswrapper[4766]: E1126 00:47:35.183404 4766 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="3.2s" Nov 26 00:47:36 crc kubenswrapper[4766]: E1126 00:47:36.162962 4766 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/71723bff6298158dee297b7aaa0d881760a12cf95a1aeea89c1c3681f290aa12/diff" to get inode usage: stat /var/lib/containers/storage/overlay/71723bff6298158dee297b7aaa0d881760a12cf95a1aeea89c1c3681f290aa12/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_ceilometer-0_ae9e2d8d-6ce6-490a-8463-a078626b551f/ceilometer-notification-agent/0.log" to get inode usage: stat /var/log/pods/openstack_ceilometer-0_ae9e2d8d-6ce6-490a-8463-a078626b551f/ceilometer-notification-agent/0.log: no such file or directory Nov 26 00:47:36 crc kubenswrapper[4766]: W1126 00:47:36.885943 4766 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-pod2bba0b1b_4848_4452_a535_c0cf5d13b30b.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-pod2bba0b1b_4848_4452_a535_c0cf5d13b30b.slice: no such file or directory Nov 26 00:47:36 crc kubenswrapper[4766]: W1126 00:47:36.886245 4766 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod619650be_59d2_47c8_815d_3337c156c83f.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod619650be_59d2_47c8_815d_3337c156c83f.slice: no such file or directory Nov 26 00:47:36 crc kubenswrapper[4766]: W1126 00:47:36.886267 4766 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2046b547_611f_4e98_ba95_a34cd0d45f50.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2046b547_611f_4e98_ba95_a34cd0d45f50.slice: no such file or directory Nov 26 00:47:36 crc kubenswrapper[4766]: W1126 00:47:36.886286 4766 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod751e5ae1_be9d_48e8_8cc1_e08012854203.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod751e5ae1_be9d_48e8_8cc1_e08012854203.slice: no such file or directory Nov 26 00:47:36 crc kubenswrapper[4766]: W1126 00:47:36.886307 4766 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb20f822d_71d5_4cb8_823b_8c261af60109.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb20f822d_71d5_4cb8_823b_8c261af60109.slice: no such file or directory Nov 26 00:47:36 crc kubenswrapper[4766]: W1126 00:47:36.886326 4766 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4794066a_0e54_4387_ae01_078abc184684.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4794066a_0e54_4387_ae01_078abc184684.slice: no such file or directory Nov 26 00:47:36 crc kubenswrapper[4766]: W1126 00:47:36.891479 4766 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d4b645a_ddab_4a29_8ffa_77e7873591c8.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d4b645a_ddab_4a29_8ffa_77e7873591c8.slice: no such file or directory Nov 26 00:47:36 crc kubenswrapper[4766]: W1126 00:47:36.893398 4766 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6612de23_8b59_474c_9089_2559ecda8811.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6612de23_8b59_474c_9089_2559ecda8811.slice: no such file or directory Nov 26 00:47:36 crc kubenswrapper[4766]: W1126 00:47:36.893793 4766 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1424c325_de53_4623_8d9c_7b1ef990c76c.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1424c325_de53_4623_8d9c_7b1ef990c76c.slice: no such file or directory Nov 26 00:47:37 crc kubenswrapper[4766]: E1126 00:47:37.028764 4766 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.32:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187b6804436575e9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-26 00:47:26.565561833 +0000 UTC m=+1427.414332263,LastTimestamp:2025-11-26 00:47:26.565561833 +0000 UTC m=+1427.414332263,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.283387 4766 generic.go:334] "Generic (PLEG): container finished" podID="16337a9e-8651-492e-9f39-21515275e3cb" containerID="0392ad20ec296c2eddec755e7b42f4a4b721787556d71d58364e5cb5612bf3e0" exitCode=137 Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.283444 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-856c5c6f86-fk66f" event={"ID":"16337a9e-8651-492e-9f39-21515275e3cb","Type":"ContainerDied","Data":"0392ad20ec296c2eddec755e7b42f4a4b721787556d71d58364e5cb5612bf3e0"} Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.285814 4766 generic.go:334] "Generic (PLEG): container finished" podID="d4442abe-9e23-4f4e-880c-89ca9d8ef785" containerID="9242e3ede9b06520f5b153cd3544ca4062ce8e5818ecaf3d453337c2fddf0419" exitCode=137 Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.285876 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" event={"ID":"d4442abe-9e23-4f4e-880c-89ca9d8ef785","Type":"ContainerDied","Data":"9242e3ede9b06520f5b153cd3544ca4062ce8e5818ecaf3d453337c2fddf0419"} Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.285901 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" event={"ID":"d4442abe-9e23-4f4e-880c-89ca9d8ef785","Type":"ContainerDied","Data":"21ccd534a3c8245de0c1ae495219c0024a5d7281a517c44b7030c1563c7f24d3"} Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.285912 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21ccd534a3c8245de0c1ae495219c0024a5d7281a517c44b7030c1563c7f24d3" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.287915 4766 generic.go:334] "Generic (PLEG): container finished" podID="91d3e9c8-8bbc-43b9-947b-71c4bb474b21" containerID="5c707b0b94e22890a9fb6d29850435a4aca9ff411003ed90468f5dcf5adef6c1" exitCode=1 Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.287955 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" event={"ID":"91d3e9c8-8bbc-43b9-947b-71c4bb474b21","Type":"ContainerDied","Data":"5c707b0b94e22890a9fb6d29850435a4aca9ff411003ed90468f5dcf5adef6c1"} Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.288687 4766 scope.go:117] "RemoveContainer" containerID="5c707b0b94e22890a9fb6d29850435a4aca9ff411003ed90468f5dcf5adef6c1" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.288785 4766 status_manager.go:851] "Failed to get status for pod" podUID="b33aa993-e31f-402f-9574-33bd30a942c4" pod="openstack/nova-cell0-conductor-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/nova-cell0-conductor-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.288968 4766 status_manager.go:851] "Failed to get status for pod" podUID="2bba0b1b-4848-4452-a535-c0cf5d13b30b" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.289296 4766 status_manager.go:851] "Failed to get status for pod" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" pod="openshift-marketplace/community-operators-rwk26" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rwk26\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.289723 4766 status_manager.go:851] "Failed to get status for pod" podUID="91d3e9c8-8bbc-43b9-947b-71c4bb474b21" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/octavia-operator-controller-manager-64cdc6ff96-bfw5f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.289938 4766 status_manager.go:851] "Failed to get status for pod" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" pod="openstack/ceilometer-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/ceilometer-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.346038 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.346891 4766 status_manager.go:851] "Failed to get status for pod" podUID="d4442abe-9e23-4f4e-880c-89ca9d8ef785" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/heat-cfnapi-b8c9d9569-lrlx6\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.347427 4766 status_manager.go:851] "Failed to get status for pod" podUID="b33aa993-e31f-402f-9574-33bd30a942c4" pod="openstack/nova-cell0-conductor-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/nova-cell0-conductor-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.347879 4766 status_manager.go:851] "Failed to get status for pod" podUID="2bba0b1b-4848-4452-a535-c0cf5d13b30b" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.348223 4766 status_manager.go:851] "Failed to get status for pod" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" pod="openshift-marketplace/community-operators-rwk26" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rwk26\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.348524 4766 status_manager.go:851] "Failed to get status for pod" podUID="91d3e9c8-8bbc-43b9-947b-71c4bb474b21" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/octavia-operator-controller-manager-64cdc6ff96-bfw5f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.348816 4766 status_manager.go:851] "Failed to get status for pod" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" pod="openstack/ceilometer-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/ceilometer-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.476898 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-856c5c6f86-fk66f" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.477773 4766 status_manager.go:851] "Failed to get status for pod" podUID="91d3e9c8-8bbc-43b9-947b-71c4bb474b21" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/octavia-operator-controller-manager-64cdc6ff96-bfw5f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.478088 4766 status_manager.go:851] "Failed to get status for pod" podUID="16337a9e-8651-492e-9f39-21515275e3cb" pod="openstack/heat-api-856c5c6f86-fk66f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/heat-api-856c5c6f86-fk66f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.478258 4766 status_manager.go:851] "Failed to get status for pod" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" pod="openstack/ceilometer-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/ceilometer-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.478454 4766 status_manager.go:851] "Failed to get status for pod" podUID="d4442abe-9e23-4f4e-880c-89ca9d8ef785" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/heat-cfnapi-b8c9d9569-lrlx6\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.478603 4766 status_manager.go:851] "Failed to get status for pod" podUID="b33aa993-e31f-402f-9574-33bd30a942c4" pod="openstack/nova-cell0-conductor-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/nova-cell0-conductor-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.478863 4766 status_manager.go:851] "Failed to get status for pod" podUID="2bba0b1b-4848-4452-a535-c0cf5d13b30b" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.479082 4766 status_manager.go:851] "Failed to get status for pod" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" pod="openshift-marketplace/community-operators-rwk26" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rwk26\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.535067 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4442abe-9e23-4f4e-880c-89ca9d8ef785-config-data-custom\") pod \"d4442abe-9e23-4f4e-880c-89ca9d8ef785\" (UID: \"d4442abe-9e23-4f4e-880c-89ca9d8ef785\") " Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.535351 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5m9s\" (UniqueName: \"kubernetes.io/projected/d4442abe-9e23-4f4e-880c-89ca9d8ef785-kube-api-access-v5m9s\") pod \"d4442abe-9e23-4f4e-880c-89ca9d8ef785\" (UID: \"d4442abe-9e23-4f4e-880c-89ca9d8ef785\") " Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.535539 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4442abe-9e23-4f4e-880c-89ca9d8ef785-config-data\") pod \"d4442abe-9e23-4f4e-880c-89ca9d8ef785\" (UID: \"d4442abe-9e23-4f4e-880c-89ca9d8ef785\") " Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.535753 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4442abe-9e23-4f4e-880c-89ca9d8ef785-combined-ca-bundle\") pod \"d4442abe-9e23-4f4e-880c-89ca9d8ef785\" (UID: \"d4442abe-9e23-4f4e-880c-89ca9d8ef785\") " Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.541001 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4442abe-9e23-4f4e-880c-89ca9d8ef785-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d4442abe-9e23-4f4e-880c-89ca9d8ef785" (UID: "d4442abe-9e23-4f4e-880c-89ca9d8ef785"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.541947 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4442abe-9e23-4f4e-880c-89ca9d8ef785-kube-api-access-v5m9s" (OuterVolumeSpecName: "kube-api-access-v5m9s") pod "d4442abe-9e23-4f4e-880c-89ca9d8ef785" (UID: "d4442abe-9e23-4f4e-880c-89ca9d8ef785"). InnerVolumeSpecName "kube-api-access-v5m9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.573322 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4442abe-9e23-4f4e-880c-89ca9d8ef785-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4442abe-9e23-4f4e-880c-89ca9d8ef785" (UID: "d4442abe-9e23-4f4e-880c-89ca9d8ef785"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.600472 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4442abe-9e23-4f4e-880c-89ca9d8ef785-config-data" (OuterVolumeSpecName: "config-data") pod "d4442abe-9e23-4f4e-880c-89ca9d8ef785" (UID: "d4442abe-9e23-4f4e-880c-89ca9d8ef785"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.637392 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16337a9e-8651-492e-9f39-21515275e3cb-config-data\") pod \"16337a9e-8651-492e-9f39-21515275e3cb\" (UID: \"16337a9e-8651-492e-9f39-21515275e3cb\") " Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.637504 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16337a9e-8651-492e-9f39-21515275e3cb-combined-ca-bundle\") pod \"16337a9e-8651-492e-9f39-21515275e3cb\" (UID: \"16337a9e-8651-492e-9f39-21515275e3cb\") " Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.637568 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16337a9e-8651-492e-9f39-21515275e3cb-config-data-custom\") pod \"16337a9e-8651-492e-9f39-21515275e3cb\" (UID: \"16337a9e-8651-492e-9f39-21515275e3cb\") " Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.637759 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmtqd\" (UniqueName: \"kubernetes.io/projected/16337a9e-8651-492e-9f39-21515275e3cb-kube-api-access-vmtqd\") pod \"16337a9e-8651-492e-9f39-21515275e3cb\" (UID: \"16337a9e-8651-492e-9f39-21515275e3cb\") " Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.638293 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4442abe-9e23-4f4e-880c-89ca9d8ef785-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.638311 4766 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4442abe-9e23-4f4e-880c-89ca9d8ef785-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.638322 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5m9s\" (UniqueName: \"kubernetes.io/projected/d4442abe-9e23-4f4e-880c-89ca9d8ef785-kube-api-access-v5m9s\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.638335 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4442abe-9e23-4f4e-880c-89ca9d8ef785-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.642016 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16337a9e-8651-492e-9f39-21515275e3cb-kube-api-access-vmtqd" (OuterVolumeSpecName: "kube-api-access-vmtqd") pod "16337a9e-8651-492e-9f39-21515275e3cb" (UID: "16337a9e-8651-492e-9f39-21515275e3cb"). InnerVolumeSpecName "kube-api-access-vmtqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.643839 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16337a9e-8651-492e-9f39-21515275e3cb-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "16337a9e-8651-492e-9f39-21515275e3cb" (UID: "16337a9e-8651-492e-9f39-21515275e3cb"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.667942 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16337a9e-8651-492e-9f39-21515275e3cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16337a9e-8651-492e-9f39-21515275e3cb" (UID: "16337a9e-8651-492e-9f39-21515275e3cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.695467 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16337a9e-8651-492e-9f39-21515275e3cb-config-data" (OuterVolumeSpecName: "config-data") pod "16337a9e-8651-492e-9f39-21515275e3cb" (UID: "16337a9e-8651-492e-9f39-21515275e3cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.741070 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16337a9e-8651-492e-9f39-21515275e3cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.741102 4766 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16337a9e-8651-492e-9f39-21515275e3cb-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.741112 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmtqd\" (UniqueName: \"kubernetes.io/projected/16337a9e-8651-492e-9f39-21515275e3cb-kube-api-access-vmtqd\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.741124 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16337a9e-8651-492e-9f39-21515275e3cb-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.826967 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.828352 4766 status_manager.go:851] "Failed to get status for pod" podUID="2bba0b1b-4848-4452-a535-c0cf5d13b30b" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.828702 4766 status_manager.go:851] "Failed to get status for pod" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" pod="openshift-marketplace/community-operators-rwk26" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rwk26\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.828932 4766 status_manager.go:851] "Failed to get status for pod" podUID="91d3e9c8-8bbc-43b9-947b-71c4bb474b21" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/octavia-operator-controller-manager-64cdc6ff96-bfw5f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.829111 4766 status_manager.go:851] "Failed to get status for pod" podUID="16337a9e-8651-492e-9f39-21515275e3cb" pod="openstack/heat-api-856c5c6f86-fk66f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/heat-api-856c5c6f86-fk66f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.829270 4766 status_manager.go:851] "Failed to get status for pod" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" pod="openstack/ceilometer-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/ceilometer-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.829427 4766 status_manager.go:851] "Failed to get status for pod" podUID="d4442abe-9e23-4f4e-880c-89ca9d8ef785" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/heat-cfnapi-b8c9d9569-lrlx6\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.829665 4766 status_manager.go:851] "Failed to get status for pod" podUID="b33aa993-e31f-402f-9574-33bd30a942c4" pod="openstack/nova-cell0-conductor-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/nova-cell0-conductor-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.848014 4766 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.848046 4766 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:47:37 crc kubenswrapper[4766]: E1126 00:47:37.848361 4766 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:37 crc kubenswrapper[4766]: I1126 00:47:37.848792 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.314301 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-856c5c6f86-fk66f" event={"ID":"16337a9e-8651-492e-9f39-21515275e3cb","Type":"ContainerDied","Data":"30c3c0b146e53a755a19931c1c4a4daedb4f44e55542080a7073ec8139c7ccd5"} Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.314327 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-856c5c6f86-fk66f" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.314570 4766 scope.go:117] "RemoveContainer" containerID="0392ad20ec296c2eddec755e7b42f4a4b721787556d71d58364e5cb5612bf3e0" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.315386 4766 status_manager.go:851] "Failed to get status for pod" podUID="d4442abe-9e23-4f4e-880c-89ca9d8ef785" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/heat-cfnapi-b8c9d9569-lrlx6\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.315673 4766 status_manager.go:851] "Failed to get status for pod" podUID="b33aa993-e31f-402f-9574-33bd30a942c4" pod="openstack/nova-cell0-conductor-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/nova-cell0-conductor-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.316010 4766 status_manager.go:851] "Failed to get status for pod" podUID="2bba0b1b-4848-4452-a535-c0cf5d13b30b" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.316343 4766 status_manager.go:851] "Failed to get status for pod" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" pod="openshift-marketplace/community-operators-rwk26" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rwk26\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.316708 4766 status_manager.go:851] "Failed to get status for pod" podUID="91d3e9c8-8bbc-43b9-947b-71c4bb474b21" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/octavia-operator-controller-manager-64cdc6ff96-bfw5f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.316981 4766 status_manager.go:851] "Failed to get status for pod" podUID="16337a9e-8651-492e-9f39-21515275e3cb" pod="openstack/heat-api-856c5c6f86-fk66f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/heat-api-856c5c6f86-fk66f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.317168 4766 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="99201497eca72eaecfe602de122913170926aa70bd8a14227a7e70208dab2552" exitCode=0 Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.317189 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"99201497eca72eaecfe602de122913170926aa70bd8a14227a7e70208dab2552"} Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.317216 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f79a59b6ed08b5c0134b4b116d488b44464e3b0ccc26c9d4342f3bfc7ada4a1e"} Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.317250 4766 status_manager.go:851] "Failed to get status for pod" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" pod="openstack/ceilometer-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/ceilometer-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.317481 4766 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.317497 4766 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.317775 4766 status_manager.go:851] "Failed to get status for pod" podUID="d4442abe-9e23-4f4e-880c-89ca9d8ef785" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/heat-cfnapi-b8c9d9569-lrlx6\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: E1126 00:47:38.317848 4766 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.318104 4766 status_manager.go:851] "Failed to get status for pod" podUID="b33aa993-e31f-402f-9574-33bd30a942c4" pod="openstack/nova-cell0-conductor-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/nova-cell0-conductor-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.318536 4766 status_manager.go:851] "Failed to get status for pod" podUID="2bba0b1b-4848-4452-a535-c0cf5d13b30b" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.319308 4766 status_manager.go:851] "Failed to get status for pod" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" pod="openshift-marketplace/community-operators-rwk26" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rwk26\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.319629 4766 status_manager.go:851] "Failed to get status for pod" podUID="91d3e9c8-8bbc-43b9-947b-71c4bb474b21" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/octavia-operator-controller-manager-64cdc6ff96-bfw5f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.320104 4766 status_manager.go:851] "Failed to get status for pod" podUID="16337a9e-8651-492e-9f39-21515275e3cb" pod="openstack/heat-api-856c5c6f86-fk66f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/heat-api-856c5c6f86-fk66f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.320607 4766 status_manager.go:851] "Failed to get status for pod" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" pod="openstack/ceilometer-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/ceilometer-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.321241 4766 status_manager.go:851] "Failed to get status for pod" podUID="d4442abe-9e23-4f4e-880c-89ca9d8ef785" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/heat-cfnapi-b8c9d9569-lrlx6\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.321810 4766 status_manager.go:851] "Failed to get status for pod" podUID="b33aa993-e31f-402f-9574-33bd30a942c4" pod="openstack/nova-cell0-conductor-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/nova-cell0-conductor-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.321848 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" event={"ID":"08bd3753-3b70-44f8-92ff-98bedd88f662","Type":"ContainerDied","Data":"c8697279442bcd764b62b25e4990741b03dd670e88bec4559ad22b620126eab6"} Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.321828 4766 generic.go:334] "Generic (PLEG): container finished" podID="08bd3753-3b70-44f8-92ff-98bedd88f662" containerID="c8697279442bcd764b62b25e4990741b03dd670e88bec4559ad22b620126eab6" exitCode=1 Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.322203 4766 status_manager.go:851] "Failed to get status for pod" podUID="2bba0b1b-4848-4452-a535-c0cf5d13b30b" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.322278 4766 scope.go:117] "RemoveContainer" containerID="c8697279442bcd764b62b25e4990741b03dd670e88bec4559ad22b620126eab6" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.322744 4766 status_manager.go:851] "Failed to get status for pod" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" pod="openshift-marketplace/community-operators-rwk26" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rwk26\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.323091 4766 status_manager.go:851] "Failed to get status for pod" podUID="91d3e9c8-8bbc-43b9-947b-71c4bb474b21" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/octavia-operator-controller-manager-64cdc6ff96-bfw5f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.323365 4766 status_manager.go:851] "Failed to get status for pod" podUID="16337a9e-8651-492e-9f39-21515275e3cb" pod="openstack/heat-api-856c5c6f86-fk66f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/heat-api-856c5c6f86-fk66f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.323586 4766 status_manager.go:851] "Failed to get status for pod" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" pod="openstack/ceilometer-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/ceilometer-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.324051 4766 status_manager.go:851] "Failed to get status for pod" podUID="91d3e9c8-8bbc-43b9-947b-71c4bb474b21" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/octavia-operator-controller-manager-64cdc6ff96-bfw5f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.324404 4766 status_manager.go:851] "Failed to get status for pod" podUID="16337a9e-8651-492e-9f39-21515275e3cb" pod="openstack/heat-api-856c5c6f86-fk66f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/heat-api-856c5c6f86-fk66f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.324906 4766 status_manager.go:851] "Failed to get status for pod" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" pod="openstack/ceilometer-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/ceilometer-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.325263 4766 status_manager.go:851] "Failed to get status for pod" podUID="d4442abe-9e23-4f4e-880c-89ca9d8ef785" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/heat-cfnapi-b8c9d9569-lrlx6\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.325740 4766 status_manager.go:851] "Failed to get status for pod" podUID="b33aa993-e31f-402f-9574-33bd30a942c4" pod="openstack/nova-cell0-conductor-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/nova-cell0-conductor-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.326349 4766 status_manager.go:851] "Failed to get status for pod" podUID="08bd3753-3b70-44f8-92ff-98bedd88f662" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/ironic-operator-controller-manager-67cb4dc6d4-qtnxq\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.326690 4766 status_manager.go:851] "Failed to get status for pod" podUID="2bba0b1b-4848-4452-a535-c0cf5d13b30b" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.326930 4766 status_manager.go:851] "Failed to get status for pod" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" pod="openshift-marketplace/community-operators-rwk26" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rwk26\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.327106 4766 generic.go:334] "Generic (PLEG): container finished" podID="91d3e9c8-8bbc-43b9-947b-71c4bb474b21" containerID="571408ef2aef1a0e6b0f38d750793734c21c57756cc64b5fd32d672f6b4e07df" exitCode=1 Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.327171 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" event={"ID":"91d3e9c8-8bbc-43b9-947b-71c4bb474b21","Type":"ContainerDied","Data":"571408ef2aef1a0e6b0f38d750793734c21c57756cc64b5fd32d672f6b4e07df"} Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.327936 4766 scope.go:117] "RemoveContainer" containerID="571408ef2aef1a0e6b0f38d750793734c21c57756cc64b5fd32d672f6b4e07df" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.328017 4766 status_manager.go:851] "Failed to get status for pod" podUID="d4442abe-9e23-4f4e-880c-89ca9d8ef785" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/heat-cfnapi-b8c9d9569-lrlx6\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.328308 4766 status_manager.go:851] "Failed to get status for pod" podUID="b33aa993-e31f-402f-9574-33bd30a942c4" pod="openstack/nova-cell0-conductor-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/nova-cell0-conductor-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.328563 4766 status_manager.go:851] "Failed to get status for pod" podUID="08bd3753-3b70-44f8-92ff-98bedd88f662" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/ironic-operator-controller-manager-67cb4dc6d4-qtnxq\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: E1126 00:47:38.328762 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=octavia-operator-controller-manager-64cdc6ff96-bfw5f_openstack-operators(91d3e9c8-8bbc-43b9-947b-71c4bb474b21)\"" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" podUID="91d3e9c8-8bbc-43b9-947b-71c4bb474b21" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.329083 4766 status_manager.go:851] "Failed to get status for pod" podUID="2bba0b1b-4848-4452-a535-c0cf5d13b30b" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.329371 4766 status_manager.go:851] "Failed to get status for pod" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" pod="openshift-marketplace/community-operators-rwk26" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rwk26\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.329833 4766 status_manager.go:851] "Failed to get status for pod" podUID="91d3e9c8-8bbc-43b9-947b-71c4bb474b21" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/octavia-operator-controller-manager-64cdc6ff96-bfw5f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.329990 4766 generic.go:334] "Generic (PLEG): container finished" podID="fd3a6275-9f47-4799-bc15-96a51b93f31b" containerID="bd98c26b9f9e66fab6812e421b2adedc526f01a40be3a9040a7e9c259b0ff68f" exitCode=1 Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.330066 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" event={"ID":"fd3a6275-9f47-4799-bc15-96a51b93f31b","Type":"ContainerDied","Data":"bd98c26b9f9e66fab6812e421b2adedc526f01a40be3a9040a7e9c259b0ff68f"} Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.330150 4766 status_manager.go:851] "Failed to get status for pod" podUID="16337a9e-8651-492e-9f39-21515275e3cb" pod="openstack/heat-api-856c5c6f86-fk66f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/heat-api-856c5c6f86-fk66f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.330447 4766 status_manager.go:851] "Failed to get status for pod" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" pod="openstack/ceilometer-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/ceilometer-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.330795 4766 status_manager.go:851] "Failed to get status for pod" podUID="d4442abe-9e23-4f4e-880c-89ca9d8ef785" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/heat-cfnapi-b8c9d9569-lrlx6\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.330870 4766 scope.go:117] "RemoveContainer" containerID="bd98c26b9f9e66fab6812e421b2adedc526f01a40be3a9040a7e9c259b0ff68f" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.331261 4766 status_manager.go:851] "Failed to get status for pod" podUID="08bd3753-3b70-44f8-92ff-98bedd88f662" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/ironic-operator-controller-manager-67cb4dc6d4-qtnxq\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.331635 4766 status_manager.go:851] "Failed to get status for pod" podUID="b33aa993-e31f-402f-9574-33bd30a942c4" pod="openstack/nova-cell0-conductor-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/nova-cell0-conductor-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.331994 4766 status_manager.go:851] "Failed to get status for pod" podUID="2bba0b1b-4848-4452-a535-c0cf5d13b30b" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.332531 4766 status_manager.go:851] "Failed to get status for pod" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" pod="openshift-marketplace/community-operators-rwk26" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rwk26\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.332917 4766 status_manager.go:851] "Failed to get status for pod" podUID="fd3a6275-9f47-4799-bc15-96a51b93f31b" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-764c8cf56f-vrwpt\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.333250 4766 status_manager.go:851] "Failed to get status for pod" podUID="91d3e9c8-8bbc-43b9-947b-71c4bb474b21" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/octavia-operator-controller-manager-64cdc6ff96-bfw5f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.333393 4766 generic.go:334] "Generic (PLEG): container finished" podID="b90f7164-a5e9-452f-8733-f549f15e17ec" containerID="29a28daa257ad0dbe49f0e720598861e7a54f0846b11c442b98c25eeb4f4d57f" exitCode=1 Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.333463 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.333465 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" event={"ID":"b90f7164-a5e9-452f-8733-f549f15e17ec","Type":"ContainerDied","Data":"29a28daa257ad0dbe49f0e720598861e7a54f0846b11c442b98c25eeb4f4d57f"} Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.333733 4766 status_manager.go:851] "Failed to get status for pod" podUID="16337a9e-8651-492e-9f39-21515275e3cb" pod="openstack/heat-api-856c5c6f86-fk66f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/heat-api-856c5c6f86-fk66f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.334039 4766 status_manager.go:851] "Failed to get status for pod" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" pod="openstack/ceilometer-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/ceilometer-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.334375 4766 scope.go:117] "RemoveContainer" containerID="29a28daa257ad0dbe49f0e720598861e7a54f0846b11c442b98c25eeb4f4d57f" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.334538 4766 status_manager.go:851] "Failed to get status for pod" podUID="b90f7164-a5e9-452f-8733-f549f15e17ec" pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operators-redhat/pods/loki-operator-controller-manager-7bbf4564c5-m77vx\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.334962 4766 status_manager.go:851] "Failed to get status for pod" podUID="d4442abe-9e23-4f4e-880c-89ca9d8ef785" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/heat-cfnapi-b8c9d9569-lrlx6\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.335422 4766 status_manager.go:851] "Failed to get status for pod" podUID="08bd3753-3b70-44f8-92ff-98bedd88f662" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/ironic-operator-controller-manager-67cb4dc6d4-qtnxq\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.335796 4766 status_manager.go:851] "Failed to get status for pod" podUID="b33aa993-e31f-402f-9574-33bd30a942c4" pod="openstack/nova-cell0-conductor-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/nova-cell0-conductor-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.336040 4766 status_manager.go:851] "Failed to get status for pod" podUID="2bba0b1b-4848-4452-a535-c0cf5d13b30b" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.336371 4766 status_manager.go:851] "Failed to get status for pod" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" pod="openshift-marketplace/community-operators-rwk26" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rwk26\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.336622 4766 status_manager.go:851] "Failed to get status for pod" podUID="fd3a6275-9f47-4799-bc15-96a51b93f31b" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-764c8cf56f-vrwpt\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.336909 4766 status_manager.go:851] "Failed to get status for pod" podUID="91d3e9c8-8bbc-43b9-947b-71c4bb474b21" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/octavia-operator-controller-manager-64cdc6ff96-bfw5f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.337248 4766 status_manager.go:851] "Failed to get status for pod" podUID="16337a9e-8651-492e-9f39-21515275e3cb" pod="openstack/heat-api-856c5c6f86-fk66f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/heat-api-856c5c6f86-fk66f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.337590 4766 status_manager.go:851] "Failed to get status for pod" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" pod="openstack/ceilometer-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/ceilometer-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.338221 4766 status_manager.go:851] "Failed to get status for pod" podUID="91d3e9c8-8bbc-43b9-947b-71c4bb474b21" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/octavia-operator-controller-manager-64cdc6ff96-bfw5f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.338612 4766 status_manager.go:851] "Failed to get status for pod" podUID="16337a9e-8651-492e-9f39-21515275e3cb" pod="openstack/heat-api-856c5c6f86-fk66f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/heat-api-856c5c6f86-fk66f\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.338974 4766 status_manager.go:851] "Failed to get status for pod" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" pod="openstack/ceilometer-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/ceilometer-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.339392 4766 status_manager.go:851] "Failed to get status for pod" podUID="b90f7164-a5e9-452f-8733-f549f15e17ec" pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operators-redhat/pods/loki-operator-controller-manager-7bbf4564c5-m77vx\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.339761 4766 status_manager.go:851] "Failed to get status for pod" podUID="d4442abe-9e23-4f4e-880c-89ca9d8ef785" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/heat-cfnapi-b8c9d9569-lrlx6\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.340108 4766 status_manager.go:851] "Failed to get status for pod" podUID="b33aa993-e31f-402f-9574-33bd30a942c4" pod="openstack/nova-cell0-conductor-0" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack/pods/nova-cell0-conductor-0\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.340437 4766 status_manager.go:851] "Failed to get status for pod" podUID="08bd3753-3b70-44f8-92ff-98bedd88f662" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/ironic-operator-controller-manager-67cb4dc6d4-qtnxq\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.340814 4766 status_manager.go:851] "Failed to get status for pod" podUID="2bba0b1b-4848-4452-a535-c0cf5d13b30b" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.341091 4766 status_manager.go:851] "Failed to get status for pod" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" pod="openshift-marketplace/community-operators-rwk26" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rwk26\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.341400 4766 status_manager.go:851] "Failed to get status for pod" podUID="fd3a6275-9f47-4799-bc15-96a51b93f31b" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-764c8cf56f-vrwpt\": dial tcp 38.102.83.32:6443: connect: connection refused" Nov 26 00:47:38 crc kubenswrapper[4766]: I1126 00:47:38.366131 4766 scope.go:117] "RemoveContainer" containerID="5c707b0b94e22890a9fb6d29850435a4aca9ff411003ed90468f5dcf5adef6c1" Nov 26 00:47:38 crc kubenswrapper[4766]: E1126 00:47:38.384502 4766 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="6.4s" Nov 26 00:47:39 crc kubenswrapper[4766]: I1126 00:47:39.348360 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 26 00:47:39 crc kubenswrapper[4766]: I1126 00:47:39.348685 4766 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e" exitCode=1 Nov 26 00:47:39 crc kubenswrapper[4766]: I1126 00:47:39.348753 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e"} Nov 26 00:47:39 crc kubenswrapper[4766]: I1126 00:47:39.352888 4766 scope.go:117] "RemoveContainer" containerID="cac37cb66d9c637b04481304141cbc82571363a72514523ef34c7917ac1d769e" Nov 26 00:47:39 crc kubenswrapper[4766]: I1126 00:47:39.360400 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" event={"ID":"b90f7164-a5e9-452f-8733-f549f15e17ec","Type":"ContainerStarted","Data":"96767efcc93da98e71dd334f384a0d650d4713297574d5e00527145ab7decdbb"} Nov 26 00:47:39 crc kubenswrapper[4766]: I1126 00:47:39.360676 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" Nov 26 00:47:39 crc kubenswrapper[4766]: I1126 00:47:39.364789 4766 generic.go:334] "Generic (PLEG): container finished" podID="fd3a6275-9f47-4799-bc15-96a51b93f31b" containerID="0db2683335300616e0363c5bc5ce375b5fd490a89a25d39263b201e2816d69ec" exitCode=1 Nov 26 00:47:39 crc kubenswrapper[4766]: I1126 00:47:39.364880 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" event={"ID":"fd3a6275-9f47-4799-bc15-96a51b93f31b","Type":"ContainerDied","Data":"0db2683335300616e0363c5bc5ce375b5fd490a89a25d39263b201e2816d69ec"} Nov 26 00:47:39 crc kubenswrapper[4766]: I1126 00:47:39.364925 4766 scope.go:117] "RemoveContainer" containerID="bd98c26b9f9e66fab6812e421b2adedc526f01a40be3a9040a7e9c259b0ff68f" Nov 26 00:47:39 crc kubenswrapper[4766]: I1126 00:47:39.365453 4766 scope.go:117] "RemoveContainer" containerID="0db2683335300616e0363c5bc5ce375b5fd490a89a25d39263b201e2816d69ec" Nov 26 00:47:39 crc kubenswrapper[4766]: E1126 00:47:39.365742 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=metallb-operator-controller-manager-764c8cf56f-vrwpt_metallb-system(fd3a6275-9f47-4799-bc15-96a51b93f31b)\"" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" podUID="fd3a6275-9f47-4799-bc15-96a51b93f31b" Nov 26 00:47:39 crc kubenswrapper[4766]: I1126 00:47:39.374920 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"24b4e3d3f660d4666bfaab0d03c9c89ef19e561a02984118c0923a121dd7cc98"} Nov 26 00:47:39 crc kubenswrapper[4766]: I1126 00:47:39.375071 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8fdfc3a820d2e3af2ed8a2c6ff2a872693ccac6689a3fa8dcb4d52e9dff74b4d"} Nov 26 00:47:39 crc kubenswrapper[4766]: I1126 00:47:39.375087 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2ab2ee09efa5daa61d59df94941665772ea28374d3da13b000e388de81bc1303"} Nov 26 00:47:39 crc kubenswrapper[4766]: I1126 00:47:39.377270 4766 generic.go:334] "Generic (PLEG): container finished" podID="08bd3753-3b70-44f8-92ff-98bedd88f662" containerID="a2540ad4bcec8692a937d7029a6100f622c38a2889d9078548875630e3de8431" exitCode=1 Nov 26 00:47:39 crc kubenswrapper[4766]: I1126 00:47:39.377300 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" event={"ID":"08bd3753-3b70-44f8-92ff-98bedd88f662","Type":"ContainerDied","Data":"a2540ad4bcec8692a937d7029a6100f622c38a2889d9078548875630e3de8431"} Nov 26 00:47:39 crc kubenswrapper[4766]: I1126 00:47:39.377964 4766 scope.go:117] "RemoveContainer" containerID="a2540ad4bcec8692a937d7029a6100f622c38a2889d9078548875630e3de8431" Nov 26 00:47:39 crc kubenswrapper[4766]: E1126 00:47:39.378198 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=ironic-operator-controller-manager-67cb4dc6d4-qtnxq_openstack-operators(08bd3753-3b70-44f8-92ff-98bedd88f662)\"" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" podUID="08bd3753-3b70-44f8-92ff-98bedd88f662" Nov 26 00:47:39 crc kubenswrapper[4766]: I1126 00:47:39.531989 4766 scope.go:117] "RemoveContainer" containerID="c8697279442bcd764b62b25e4990741b03dd670e88bec4559ad22b620126eab6" Nov 26 00:47:40 crc kubenswrapper[4766]: I1126 00:47:40.398196 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 26 00:47:40 crc kubenswrapper[4766]: I1126 00:47:40.398687 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b0571bdb3c4006736a79e1db00f3608748b2f4f5e355a766ba30c7d23c29b4c8"} Nov 26 00:47:40 crc kubenswrapper[4766]: I1126 00:47:40.409069 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2d78726ae058e267e3d7d1c34088af172ce9f680b42b65f4e6599200dc66da4c"} Nov 26 00:47:40 crc kubenswrapper[4766]: I1126 00:47:40.409152 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ffc97aa724027a54864be3066b974909a07c8599e1df16945cbd6d4e4b216091"} Nov 26 00:47:40 crc kubenswrapper[4766]: I1126 00:47:40.409474 4766 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:47:40 crc kubenswrapper[4766]: I1126 00:47:40.409494 4766 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:47:40 crc kubenswrapper[4766]: I1126 00:47:40.441970 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" containerName="kube-state-metrics" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 26 00:47:41 crc kubenswrapper[4766]: I1126 00:47:41.997282 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" Nov 26 00:47:41 crc kubenswrapper[4766]: I1126 00:47:41.998319 4766 scope.go:117] "RemoveContainer" containerID="a2540ad4bcec8692a937d7029a6100f622c38a2889d9078548875630e3de8431" Nov 26 00:47:41 crc kubenswrapper[4766]: E1126 00:47:41.998641 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=ironic-operator-controller-manager-67cb4dc6d4-qtnxq_openstack-operators(08bd3753-3b70-44f8-92ff-98bedd88f662)\"" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" podUID="08bd3753-3b70-44f8-92ff-98bedd88f662" Nov 26 00:47:42 crc kubenswrapper[4766]: I1126 00:47:42.200956 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" Nov 26 00:47:42 crc kubenswrapper[4766]: I1126 00:47:42.201801 4766 scope.go:117] "RemoveContainer" containerID="571408ef2aef1a0e6b0f38d750793734c21c57756cc64b5fd32d672f6b4e07df" Nov 26 00:47:42 crc kubenswrapper[4766]: E1126 00:47:42.202099 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=octavia-operator-controller-manager-64cdc6ff96-bfw5f_openstack-operators(91d3e9c8-8bbc-43b9-947b-71c4bb474b21)\"" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" podUID="91d3e9c8-8bbc-43b9-947b-71c4bb474b21" Nov 26 00:47:42 crc kubenswrapper[4766]: I1126 00:47:42.849057 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:42 crc kubenswrapper[4766]: I1126 00:47:42.849132 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:42 crc kubenswrapper[4766]: I1126 00:47:42.854491 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:44 crc kubenswrapper[4766]: I1126 00:47:44.413760 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" Nov 26 00:47:44 crc kubenswrapper[4766]: I1126 00:47:44.414757 4766 scope.go:117] "RemoveContainer" containerID="0db2683335300616e0363c5bc5ce375b5fd490a89a25d39263b201e2816d69ec" Nov 26 00:47:44 crc kubenswrapper[4766]: E1126 00:47:44.414976 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=metallb-operator-controller-manager-764c8cf56f-vrwpt_metallb-system(fd3a6275-9f47-4799-bc15-96a51b93f31b)\"" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" podUID="fd3a6275-9f47-4799-bc15-96a51b93f31b" Nov 26 00:47:44 crc kubenswrapper[4766]: I1126 00:47:44.903356 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-7bbf4564c5-m77vx" Nov 26 00:47:45 crc kubenswrapper[4766]: I1126 00:47:45.421030 4766 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:45 crc kubenswrapper[4766]: I1126 00:47:45.462293 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:45 crc kubenswrapper[4766]: I1126 00:47:45.462337 4766 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:47:45 crc kubenswrapper[4766]: I1126 00:47:45.462367 4766 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:47:45 crc kubenswrapper[4766]: I1126 00:47:45.466369 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:46 crc kubenswrapper[4766]: I1126 00:47:46.473086 4766 generic.go:334] "Generic (PLEG): container finished" podID="a2e7c098-bf9d-4d99-84bf-adfb49ee6b12" containerID="973fb49329012ed9c8a7426429d28b69e518c0f2507cbfeacc05ff401e64aa19" exitCode=1 Nov 26 00:47:46 crc kubenswrapper[4766]: I1126 00:47:46.473168 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" event={"ID":"a2e7c098-bf9d-4d99-84bf-adfb49ee6b12","Type":"ContainerDied","Data":"973fb49329012ed9c8a7426429d28b69e518c0f2507cbfeacc05ff401e64aa19"} Nov 26 00:47:46 crc kubenswrapper[4766]: I1126 00:47:46.474034 4766 scope.go:117] "RemoveContainer" containerID="973fb49329012ed9c8a7426429d28b69e518c0f2507cbfeacc05ff401e64aa19" Nov 26 00:47:46 crc kubenswrapper[4766]: I1126 00:47:46.474581 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_71bb4a3aecc4ba5b26c4b7318770ce13/kube-apiserver-check-endpoints/0.log" Nov 26 00:47:46 crc kubenswrapper[4766]: I1126 00:47:46.475966 4766 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="2d78726ae058e267e3d7d1c34088af172ce9f680b42b65f4e6599200dc66da4c" exitCode=255 Nov 26 00:47:46 crc kubenswrapper[4766]: I1126 00:47:46.476049 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"2d78726ae058e267e3d7d1c34088af172ce9f680b42b65f4e6599200dc66da4c"} Nov 26 00:47:46 crc kubenswrapper[4766]: I1126 00:47:46.476481 4766 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:47:46 crc kubenswrapper[4766]: I1126 00:47:46.476499 4766 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:47:46 crc kubenswrapper[4766]: I1126 00:47:46.478218 4766 generic.go:334] "Generic (PLEG): container finished" podID="cc87f631-ba4e-4943-b51d-bf1012cca754" containerID="f7cdc453d3359bcc3627ad6f9177b79e5434324bdb9b07dd2d35a4e00b037322" exitCode=1 Nov 26 00:47:46 crc kubenswrapper[4766]: I1126 00:47:46.478298 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" event={"ID":"cc87f631-ba4e-4943-b51d-bf1012cca754","Type":"ContainerDied","Data":"f7cdc453d3359bcc3627ad6f9177b79e5434324bdb9b07dd2d35a4e00b037322"} Nov 26 00:47:46 crc kubenswrapper[4766]: I1126 00:47:46.479065 4766 scope.go:117] "RemoveContainer" containerID="f7cdc453d3359bcc3627ad6f9177b79e5434324bdb9b07dd2d35a4e00b037322" Nov 26 00:47:46 crc kubenswrapper[4766]: I1126 00:47:46.482524 4766 generic.go:334] "Generic (PLEG): container finished" podID="2bb23a3b-1f4a-409f-9d9c-0afe73a33352" containerID="2f60c1d7aa6d4d7875d73101d3c8630a065629b9bbc495ba4eb16c81c1076282" exitCode=1 Nov 26 00:47:46 crc kubenswrapper[4766]: I1126 00:47:46.482582 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" event={"ID":"2bb23a3b-1f4a-409f-9d9c-0afe73a33352","Type":"ContainerDied","Data":"2f60c1d7aa6d4d7875d73101d3c8630a065629b9bbc495ba4eb16c81c1076282"} Nov 26 00:47:46 crc kubenswrapper[4766]: I1126 00:47:46.483332 4766 scope.go:117] "RemoveContainer" containerID="2f60c1d7aa6d4d7875d73101d3c8630a065629b9bbc495ba4eb16c81c1076282" Nov 26 00:47:46 crc kubenswrapper[4766]: I1126 00:47:46.490126 4766 scope.go:117] "RemoveContainer" containerID="2d78726ae058e267e3d7d1c34088af172ce9f680b42b65f4e6599200dc66da4c" Nov 26 00:47:46 crc kubenswrapper[4766]: I1126 00:47:46.490493 4766 generic.go:334] "Generic (PLEG): container finished" podID="5284ad70-a36c-4666-8202-84633d83f50b" containerID="adbc6b083454dd91829f6392ff5941d676cb8c4d44bbae613815465810d69a3e" exitCode=1 Nov 26 00:47:46 crc kubenswrapper[4766]: I1126 00:47:46.490531 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" event={"ID":"5284ad70-a36c-4666-8202-84633d83f50b","Type":"ContainerDied","Data":"adbc6b083454dd91829f6392ff5941d676cb8c4d44bbae613815465810d69a3e"} Nov 26 00:47:46 crc kubenswrapper[4766]: I1126 00:47:46.491290 4766 scope.go:117] "RemoveContainer" containerID="adbc6b083454dd91829f6392ff5941d676cb8c4d44bbae613815465810d69a3e" Nov 26 00:47:46 crc kubenswrapper[4766]: I1126 00:47:46.724838 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 00:47:46 crc kubenswrapper[4766]: I1126 00:47:46.729537 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.306805 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-b8c9d9569-lrlx6" podUID="d4442abe-9e23-4f4e-880c-89ca9d8ef785" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.0.195:8000/healthcheck\": dial tcp 10.217.0.195:8000: i/o timeout (Client.Timeout exceeded while awaiting headers)" Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.425437 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-856c5c6f86-fk66f" podUID="16337a9e-8651-492e-9f39-21515275e3cb" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.197:8004/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.504900 4766 generic.go:334] "Generic (PLEG): container finished" podID="cc87f631-ba4e-4943-b51d-bf1012cca754" containerID="b6766bbab2729ba50956a4e98254bd688babe7e7ba968cdd1ea57ef3ea991f0e" exitCode=1 Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.505059 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" event={"ID":"cc87f631-ba4e-4943-b51d-bf1012cca754","Type":"ContainerDied","Data":"b6766bbab2729ba50956a4e98254bd688babe7e7ba968cdd1ea57ef3ea991f0e"} Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.505403 4766 scope.go:117] "RemoveContainer" containerID="f7cdc453d3359bcc3627ad6f9177b79e5434324bdb9b07dd2d35a4e00b037322" Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.506059 4766 scope.go:117] "RemoveContainer" containerID="b6766bbab2729ba50956a4e98254bd688babe7e7ba968cdd1ea57ef3ea991f0e" Nov 26 00:47:47 crc kubenswrapper[4766]: E1126 00:47:47.506422 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=barbican-operator-controller-manager-7b64f4fb85-gbdhz_openstack-operators(cc87f631-ba4e-4943-b51d-bf1012cca754)\"" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" podUID="cc87f631-ba4e-4943-b51d-bf1012cca754" Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.509313 4766 generic.go:334] "Generic (PLEG): container finished" podID="2bb23a3b-1f4a-409f-9d9c-0afe73a33352" containerID="5d7d2b957bbb4841fc8ab537bef654cacd24c482167164e785eed6faba12e21f" exitCode=1 Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.514575 4766 generic.go:334] "Generic (PLEG): container finished" podID="04e53d09-de84-4682-b17b-61f9ea7bf152" containerID="e2d340c49871daeed42f4ca0db9efe9f2362548fc9ff8ab87a1d99962b1a93c3" exitCode=1 Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.524878 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" event={"ID":"2bb23a3b-1f4a-409f-9d9c-0afe73a33352","Type":"ContainerDied","Data":"5d7d2b957bbb4841fc8ab537bef654cacd24c482167164e785eed6faba12e21f"} Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.524952 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" event={"ID":"04e53d09-de84-4682-b17b-61f9ea7bf152","Type":"ContainerDied","Data":"e2d340c49871daeed42f4ca0db9efe9f2362548fc9ff8ab87a1d99962b1a93c3"} Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.525584 4766 scope.go:117] "RemoveContainer" containerID="e2d340c49871daeed42f4ca0db9efe9f2362548fc9ff8ab87a1d99962b1a93c3" Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.526615 4766 scope.go:117] "RemoveContainer" containerID="5d7d2b957bbb4841fc8ab537bef654cacd24c482167164e785eed6faba12e21f" Nov 26 00:47:47 crc kubenswrapper[4766]: E1126 00:47:47.527043 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=neutron-operator-controller-manager-6fdcddb789-wdzgw_openstack-operators(2bb23a3b-1f4a-409f-9d9c-0afe73a33352)\"" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" podUID="2bb23a3b-1f4a-409f-9d9c-0afe73a33352" Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.528420 4766 generic.go:334] "Generic (PLEG): container finished" podID="c3cec209-fdef-44a6-90d0-2d06cfd7e5cc" containerID="6361803ac24912c90abb031d2a560bba65c8b89ddf1b832d483be9a3151edb53" exitCode=1 Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.528792 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" event={"ID":"c3cec209-fdef-44a6-90d0-2d06cfd7e5cc","Type":"ContainerDied","Data":"6361803ac24912c90abb031d2a560bba65c8b89ddf1b832d483be9a3151edb53"} Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.529377 4766 scope.go:117] "RemoveContainer" containerID="6361803ac24912c90abb031d2a560bba65c8b89ddf1b832d483be9a3151edb53" Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.538602 4766 generic.go:334] "Generic (PLEG): container finished" podID="5284ad70-a36c-4666-8202-84633d83f50b" containerID="c4b508111353f64ce1118a8ac912e3d512d2f23b317dd6a0483c4fe2e0126067" exitCode=1 Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.538684 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" event={"ID":"5284ad70-a36c-4666-8202-84633d83f50b","Type":"ContainerDied","Data":"c4b508111353f64ce1118a8ac912e3d512d2f23b317dd6a0483c4fe2e0126067"} Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.539121 4766 scope.go:117] "RemoveContainer" containerID="c4b508111353f64ce1118a8ac912e3d512d2f23b317dd6a0483c4fe2e0126067" Nov 26 00:47:47 crc kubenswrapper[4766]: E1126 00:47:47.539481 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=horizon-operator-controller-manager-5d494799bf-h4mrn_openstack-operators(5284ad70-a36c-4666-8202-84633d83f50b)\"" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" podUID="5284ad70-a36c-4666-8202-84633d83f50b" Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.542791 4766 generic.go:334] "Generic (PLEG): container finished" podID="62c72069-cf63-4244-a55e-0028aaaed380" containerID="8f4c6185fb7a867b4fb5901e3bfa554b6d34fb4271abaf7dfaee99cdc49113be" exitCode=1 Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.542863 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7d5fdccdc4-5sh97" event={"ID":"62c72069-cf63-4244-a55e-0028aaaed380","Type":"ContainerDied","Data":"8f4c6185fb7a867b4fb5901e3bfa554b6d34fb4271abaf7dfaee99cdc49113be"} Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.543476 4766 scope.go:117] "RemoveContainer" containerID="8f4c6185fb7a867b4fb5901e3bfa554b6d34fb4271abaf7dfaee99cdc49113be" Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.548272 4766 generic.go:334] "Generic (PLEG): container finished" podID="751b8e60-46d4-4f5e-b62d-800f641ccd40" containerID="529ad9f55b98547ef2e983ec1a27611aee84050866339d4ad48ffac041ae848e" exitCode=1 Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.548334 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" event={"ID":"751b8e60-46d4-4f5e-b62d-800f641ccd40","Type":"ContainerDied","Data":"529ad9f55b98547ef2e983ec1a27611aee84050866339d4ad48ffac041ae848e"} Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.548985 4766 scope.go:117] "RemoveContainer" containerID="529ad9f55b98547ef2e983ec1a27611aee84050866339d4ad48ffac041ae848e" Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.558882 4766 generic.go:334] "Generic (PLEG): container finished" podID="a2e7c098-bf9d-4d99-84bf-adfb49ee6b12" containerID="3a5bf6a88bd14fa3df408b57df7d3930b7c4f71f2d70788e373686f0aad92b30" exitCode=1 Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.558962 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" event={"ID":"a2e7c098-bf9d-4d99-84bf-adfb49ee6b12","Type":"ContainerDied","Data":"3a5bf6a88bd14fa3df408b57df7d3930b7c4f71f2d70788e373686f0aad92b30"} Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.562310 4766 scope.go:117] "RemoveContainer" containerID="3a5bf6a88bd14fa3df408b57df7d3930b7c4f71f2d70788e373686f0aad92b30" Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.563134 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_71bb4a3aecc4ba5b26c4b7318770ce13/kube-apiserver-check-endpoints/0.log" Nov 26 00:47:47 crc kubenswrapper[4766]: E1126 00:47:47.564432 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=ovn-operator-controller-manager-56897c768d-9zhql_openstack-operators(a2e7c098-bf9d-4d99-84bf-adfb49ee6b12)\"" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" podUID="a2e7c098-bf9d-4d99-84bf-adfb49ee6b12" Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.568038 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c9d7eda97f15fe6f9a75ca8a20ea0e2e59d89ec2c8fa0c399aa26ed106e094ec"} Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.568248 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.568284 4766 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.568487 4766 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.569970 4766 generic.go:334] "Generic (PLEG): container finished" podID="1ecbd1d2-cb7f-432f-bc75-903d0643ea29" containerID="8a2e74efad7d541f04c8dc635f49437996e65e1ca2c77cfd523ba5f59dee0347" exitCode=1 Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.570038 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" event={"ID":"1ecbd1d2-cb7f-432f-bc75-903d0643ea29","Type":"ContainerDied","Data":"8a2e74efad7d541f04c8dc635f49437996e65e1ca2c77cfd523ba5f59dee0347"} Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.571449 4766 scope.go:117] "RemoveContainer" containerID="8a2e74efad7d541f04c8dc635f49437996e65e1ca2c77cfd523ba5f59dee0347" Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.585115 4766 generic.go:334] "Generic (PLEG): container finished" podID="40780971-20ed-4b4c-8d19-e74f6582ef8c" containerID="d771a6533701416a700cc1e18bc1cd25b44543865ae2251cd56deca2cba74e69" exitCode=1 Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.585453 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-glkl6" event={"ID":"40780971-20ed-4b4c-8d19-e74f6582ef8c","Type":"ContainerDied","Data":"d771a6533701416a700cc1e18bc1cd25b44543865ae2251cd56deca2cba74e69"} Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.585987 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.586429 4766 scope.go:117] "RemoveContainer" containerID="d771a6533701416a700cc1e18bc1cd25b44543865ae2251cd56deca2cba74e69" Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.589064 4766 scope.go:117] "RemoveContainer" containerID="2f60c1d7aa6d4d7875d73101d3c8630a065629b9bbc495ba4eb16c81c1076282" Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.861006 4766 scope.go:117] "RemoveContainer" containerID="adbc6b083454dd91829f6392ff5941d676cb8c4d44bbae613815465810d69a3e" Nov 26 00:47:47 crc kubenswrapper[4766]: I1126 00:47:47.894108 4766 scope.go:117] "RemoveContainer" containerID="973fb49329012ed9c8a7426429d28b69e518c0f2507cbfeacc05ff401e64aa19" Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.597539 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_71bb4a3aecc4ba5b26c4b7318770ce13/kube-apiserver-check-endpoints/1.log" Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.599144 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_71bb4a3aecc4ba5b26c4b7318770ce13/kube-apiserver-check-endpoints/0.log" Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.600512 4766 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="c9d7eda97f15fe6f9a75ca8a20ea0e2e59d89ec2c8fa0c399aa26ed106e094ec" exitCode=255 Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.600575 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"c9d7eda97f15fe6f9a75ca8a20ea0e2e59d89ec2c8fa0c399aa26ed106e094ec"} Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.600615 4766 scope.go:117] "RemoveContainer" containerID="2d78726ae058e267e3d7d1c34088af172ce9f680b42b65f4e6599200dc66da4c" Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.601052 4766 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.601135 4766 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.613599 4766 scope.go:117] "RemoveContainer" containerID="c9d7eda97f15fe6f9a75ca8a20ea0e2e59d89ec2c8fa0c399aa26ed106e094ec" Nov 26 00:47:48 crc kubenswrapper[4766]: E1126 00:47:48.614025 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(71bb4a3aecc4ba5b26c4b7318770ce13)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.615408 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7d5fdccdc4-5sh97" event={"ID":"62c72069-cf63-4244-a55e-0028aaaed380","Type":"ContainerStarted","Data":"943ff5eb47eb575cf22bb294cbd1a849f1039dc40bd1d7a9475f0a955da539fb"} Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.616153 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7d5fdccdc4-5sh97" Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.627383 4766 generic.go:334] "Generic (PLEG): container finished" podID="749a25e6-23eb-4fdd-8e15-ac2874cdee7f" containerID="e8d48edf8701473cece2bdce0d7107d87dd054530b4e46cc47521fd87e92ac5a" exitCode=1 Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.627461 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" event={"ID":"749a25e6-23eb-4fdd-8e15-ac2874cdee7f","Type":"ContainerDied","Data":"e8d48edf8701473cece2bdce0d7107d87dd054530b4e46cc47521fd87e92ac5a"} Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.628399 4766 scope.go:117] "RemoveContainer" containerID="e8d48edf8701473cece2bdce0d7107d87dd054530b4e46cc47521fd87e92ac5a" Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.633692 4766 generic.go:334] "Generic (PLEG): container finished" podID="8969652f-14c9-4618-ad2d-1b3fd1ad3627" containerID="7ca895f781b91bd56703eeb8cbdcc0e9ef3d10ec65456779ca8985abe2fadd88" exitCode=1 Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.633789 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" event={"ID":"8969652f-14c9-4618-ad2d-1b3fd1ad3627","Type":"ContainerDied","Data":"7ca895f781b91bd56703eeb8cbdcc0e9ef3d10ec65456779ca8985abe2fadd88"} Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.635532 4766 scope.go:117] "RemoveContainer" containerID="7ca895f781b91bd56703eeb8cbdcc0e9ef3d10ec65456779ca8985abe2fadd88" Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.637576 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-glkl6" event={"ID":"40780971-20ed-4b4c-8d19-e74f6582ef8c","Type":"ContainerStarted","Data":"583010fbc99cceb4713cbae2f15a5899788059edb08d70d1fa51399f4790ee58"} Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.638685 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-glkl6" Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.643129 4766 generic.go:334] "Generic (PLEG): container finished" podID="751b8e60-46d4-4f5e-b62d-800f641ccd40" containerID="37c98bbe0d1a1d7dec9f2442d2c16ebe49c7485725817b532cbc38bae4365252" exitCode=1 Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.643197 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" event={"ID":"751b8e60-46d4-4f5e-b62d-800f641ccd40","Type":"ContainerDied","Data":"37c98bbe0d1a1d7dec9f2442d2c16ebe49c7485725817b532cbc38bae4365252"} Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.644059 4766 scope.go:117] "RemoveContainer" containerID="37c98bbe0d1a1d7dec9f2442d2c16ebe49c7485725817b532cbc38bae4365252" Nov 26 00:47:48 crc kubenswrapper[4766]: E1126 00:47:48.644389 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=watcher-operator-controller-manager-656dcb59d4-2hpg8_openstack-operators(751b8e60-46d4-4f5e-b62d-800f641ccd40)\"" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" podUID="751b8e60-46d4-4f5e-b62d-800f641ccd40" Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.650182 4766 generic.go:334] "Generic (PLEG): container finished" podID="04e53d09-de84-4682-b17b-61f9ea7bf152" containerID="089ff57a19d572ad607d9b8aefbeca178c8046089d9d0f1a2b360a9cd51d8e1b" exitCode=1 Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.650232 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" event={"ID":"04e53d09-de84-4682-b17b-61f9ea7bf152","Type":"ContainerDied","Data":"089ff57a19d572ad607d9b8aefbeca178c8046089d9d0f1a2b360a9cd51d8e1b"} Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.650895 4766 scope.go:117] "RemoveContainer" containerID="089ff57a19d572ad607d9b8aefbeca178c8046089d9d0f1a2b360a9cd51d8e1b" Nov 26 00:47:48 crc kubenswrapper[4766]: E1126 00:47:48.651131 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=telemetry-operator-controller-manager-5ccf99cfdd-9cfft_openstack-operators(04e53d09-de84-4682-b17b-61f9ea7bf152)\"" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" podUID="04e53d09-de84-4682-b17b-61f9ea7bf152" Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.653096 4766 generic.go:334] "Generic (PLEG): container finished" podID="c3cec209-fdef-44a6-90d0-2d06cfd7e5cc" containerID="f4f0be8205cb3c087b28e7451a802ae7e7912ffbab226d356ad6b10cb884ef49" exitCode=1 Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.653173 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" event={"ID":"c3cec209-fdef-44a6-90d0-2d06cfd7e5cc","Type":"ContainerDied","Data":"f4f0be8205cb3c087b28e7451a802ae7e7912ffbab226d356ad6b10cb884ef49"} Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.653605 4766 scope.go:117] "RemoveContainer" containerID="f4f0be8205cb3c087b28e7451a802ae7e7912ffbab226d356ad6b10cb884ef49" Nov 26 00:47:48 crc kubenswrapper[4766]: E1126 00:47:48.653935 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=infra-operator-controller-manager-57548d458d-x6h4v_openstack-operators(c3cec209-fdef-44a6-90d0-2d06cfd7e5cc)\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" podUID="c3cec209-fdef-44a6-90d0-2d06cfd7e5cc" Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.655968 4766 generic.go:334] "Generic (PLEG): container finished" podID="3d0d59b2-0486-40a2-ae60-b738a4f76dc4" containerID="282e0f3e513c0848d9385350f540459f60d5cca52c05ea2585f15b371786e987" exitCode=1 Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.656056 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" event={"ID":"3d0d59b2-0486-40a2-ae60-b738a4f76dc4","Type":"ContainerDied","Data":"282e0f3e513c0848d9385350f540459f60d5cca52c05ea2585f15b371786e987"} Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.656690 4766 scope.go:117] "RemoveContainer" containerID="282e0f3e513c0848d9385350f540459f60d5cca52c05ea2585f15b371786e987" Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.659135 4766 generic.go:334] "Generic (PLEG): container finished" podID="1ecbd1d2-cb7f-432f-bc75-903d0643ea29" containerID="5f7d72a8709e834d6b16398a4a94f1cdb096b975fefcb4fa32830decc9e3a380" exitCode=1 Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.659390 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" event={"ID":"1ecbd1d2-cb7f-432f-bc75-903d0643ea29","Type":"ContainerDied","Data":"5f7d72a8709e834d6b16398a4a94f1cdb096b975fefcb4fa32830decc9e3a380"} Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.660221 4766 scope.go:117] "RemoveContainer" containerID="5f7d72a8709e834d6b16398a4a94f1cdb096b975fefcb4fa32830decc9e3a380" Nov 26 00:47:48 crc kubenswrapper[4766]: E1126 00:47:48.660994 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=heat-operator-controller-manager-5b77f656f-tkj9b_openstack-operators(1ecbd1d2-cb7f-432f-bc75-903d0643ea29)\"" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" podUID="1ecbd1d2-cb7f-432f-bc75-903d0643ea29" Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.814706 4766 scope.go:117] "RemoveContainer" containerID="529ad9f55b98547ef2e983ec1a27611aee84050866339d4ad48ffac041ae848e" Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.892622 4766 scope.go:117] "RemoveContainer" containerID="e2d340c49871daeed42f4ca0db9efe9f2362548fc9ff8ab87a1d99962b1a93c3" Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.938351 4766 scope.go:117] "RemoveContainer" containerID="6361803ac24912c90abb031d2a560bba65c8b89ddf1b832d483be9a3151edb53" Nov 26 00:47:48 crc kubenswrapper[4766]: I1126 00:47:48.971136 4766 scope.go:117] "RemoveContainer" containerID="8a2e74efad7d541f04c8dc635f49437996e65e1ca2c77cfd523ba5f59dee0347" Nov 26 00:47:49 crc kubenswrapper[4766]: I1126 00:47:49.669544 4766 generic.go:334] "Generic (PLEG): container finished" podID="749a25e6-23eb-4fdd-8e15-ac2874cdee7f" containerID="502141878d9557d404939737707070c203f202ba93a6d5aa04240a14a618e040" exitCode=1 Nov 26 00:47:49 crc kubenswrapper[4766]: I1126 00:47:49.669612 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" event={"ID":"749a25e6-23eb-4fdd-8e15-ac2874cdee7f","Type":"ContainerDied","Data":"502141878d9557d404939737707070c203f202ba93a6d5aa04240a14a618e040"} Nov 26 00:47:49 crc kubenswrapper[4766]: I1126 00:47:49.669970 4766 scope.go:117] "RemoveContainer" containerID="e8d48edf8701473cece2bdce0d7107d87dd054530b4e46cc47521fd87e92ac5a" Nov 26 00:47:49 crc kubenswrapper[4766]: I1126 00:47:49.670561 4766 scope.go:117] "RemoveContainer" containerID="502141878d9557d404939737707070c203f202ba93a6d5aa04240a14a618e040" Nov 26 00:47:49 crc kubenswrapper[4766]: E1126 00:47:49.670932 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=mariadb-operator-controller-manager-66f4dd4bc7-z865g_openstack-operators(749a25e6-23eb-4fdd-8e15-ac2874cdee7f)\"" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" podUID="749a25e6-23eb-4fdd-8e15-ac2874cdee7f" Nov 26 00:47:49 crc kubenswrapper[4766]: I1126 00:47:49.679121 4766 generic.go:334] "Generic (PLEG): container finished" podID="3d0d59b2-0486-40a2-ae60-b738a4f76dc4" containerID="ccb6a5c90a0ad287ea4fde750c373ddb97fe844e0b2024e905b73d849ad7c7d6" exitCode=1 Nov 26 00:47:49 crc kubenswrapper[4766]: I1126 00:47:49.679211 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" event={"ID":"3d0d59b2-0486-40a2-ae60-b738a4f76dc4","Type":"ContainerDied","Data":"ccb6a5c90a0ad287ea4fde750c373ddb97fe844e0b2024e905b73d849ad7c7d6"} Nov 26 00:47:49 crc kubenswrapper[4766]: I1126 00:47:49.680097 4766 scope.go:117] "RemoveContainer" containerID="ccb6a5c90a0ad287ea4fde750c373ddb97fe844e0b2024e905b73d849ad7c7d6" Nov 26 00:47:49 crc kubenswrapper[4766]: E1126 00:47:49.680744 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=keystone-operator-controller-manager-7b4567c7cf-fcr76_openstack-operators(3d0d59b2-0486-40a2-ae60-b738a4f76dc4)\"" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" podUID="3d0d59b2-0486-40a2-ae60-b738a4f76dc4" Nov 26 00:47:49 crc kubenswrapper[4766]: I1126 00:47:49.684487 4766 generic.go:334] "Generic (PLEG): container finished" podID="91ee52b5-0e2e-4732-87bf-809e378d33d3" containerID="0a0ef28abd9114b5dfc1b827a78e10c6f12f44b41063b3480c57b485a5daea67" exitCode=1 Nov 26 00:47:49 crc kubenswrapper[4766]: I1126 00:47:49.684541 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw" event={"ID":"91ee52b5-0e2e-4732-87bf-809e378d33d3","Type":"ContainerDied","Data":"0a0ef28abd9114b5dfc1b827a78e10c6f12f44b41063b3480c57b485a5daea67"} Nov 26 00:47:49 crc kubenswrapper[4766]: I1126 00:47:49.685178 4766 scope.go:117] "RemoveContainer" containerID="0a0ef28abd9114b5dfc1b827a78e10c6f12f44b41063b3480c57b485a5daea67" Nov 26 00:47:49 crc kubenswrapper[4766]: I1126 00:47:49.695026 4766 generic.go:334] "Generic (PLEG): container finished" podID="8969652f-14c9-4618-ad2d-1b3fd1ad3627" containerID="69ed96c12e25467c32b6ad99ca52ad4f46ecc9f3be096414f5b261a00874e013" exitCode=1 Nov 26 00:47:49 crc kubenswrapper[4766]: I1126 00:47:49.695090 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" event={"ID":"8969652f-14c9-4618-ad2d-1b3fd1ad3627","Type":"ContainerDied","Data":"69ed96c12e25467c32b6ad99ca52ad4f46ecc9f3be096414f5b261a00874e013"} Nov 26 00:47:49 crc kubenswrapper[4766]: I1126 00:47:49.695596 4766 scope.go:117] "RemoveContainer" containerID="69ed96c12e25467c32b6ad99ca52ad4f46ecc9f3be096414f5b261a00874e013" Nov 26 00:47:49 crc kubenswrapper[4766]: E1126 00:47:49.695968 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=nova-operator-controller-manager-79556f57fc-dm7xb_openstack-operators(8969652f-14c9-4618-ad2d-1b3fd1ad3627)\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" podUID="8969652f-14c9-4618-ad2d-1b3fd1ad3627" Nov 26 00:47:49 crc kubenswrapper[4766]: I1126 00:47:49.698373 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_71bb4a3aecc4ba5b26c4b7318770ce13/kube-apiserver-check-endpoints/1.log" Nov 26 00:47:49 crc kubenswrapper[4766]: I1126 00:47:49.701352 4766 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:47:49 crc kubenswrapper[4766]: I1126 00:47:49.701380 4766 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:47:49 crc kubenswrapper[4766]: I1126 00:47:49.705312 4766 scope.go:117] "RemoveContainer" containerID="c9d7eda97f15fe6f9a75ca8a20ea0e2e59d89ec2c8fa0c399aa26ed106e094ec" Nov 26 00:47:49 crc kubenswrapper[4766]: E1126 00:47:49.705599 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(71bb4a3aecc4ba5b26c4b7318770ce13)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Nov 26 00:47:49 crc kubenswrapper[4766]: I1126 00:47:49.743771 4766 scope.go:117] "RemoveContainer" containerID="282e0f3e513c0848d9385350f540459f60d5cca52c05ea2585f15b371786e987" Nov 26 00:47:49 crc kubenswrapper[4766]: I1126 00:47:49.861424 4766 scope.go:117] "RemoveContainer" containerID="7ca895f781b91bd56703eeb8cbdcc0e9ef3d10ec65456779ca8985abe2fadd88" Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.422890 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" containerName="kube-state-metrics" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.423105 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/kube-state-metrics-0" Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.423797 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-state-metrics" containerStatusID={"Type":"cri-o","ID":"ec396350a1da311b232393c00a8651565f79130843bddb720cbd326d3df4f5dd"} pod="openstack/kube-state-metrics-0" containerMessage="Container kube-state-metrics failed liveness probe, will be restarted" Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.423835 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" containerName="kube-state-metrics" containerID="cri-o://ec396350a1da311b232393c00a8651565f79130843bddb720cbd326d3df4f5dd" gracePeriod=30 Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.713408 4766 generic.go:334] "Generic (PLEG): container finished" podID="91ee52b5-0e2e-4732-87bf-809e378d33d3" containerID="9e0ffa5e614b56b5972b597f5d7cf1d69c9755a6edfc1db6ac05034cb4e3d62d" exitCode=1 Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.713488 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw" event={"ID":"91ee52b5-0e2e-4732-87bf-809e378d33d3","Type":"ContainerDied","Data":"9e0ffa5e614b56b5972b597f5d7cf1d69c9755a6edfc1db6ac05034cb4e3d62d"} Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.713551 4766 scope.go:117] "RemoveContainer" containerID="0a0ef28abd9114b5dfc1b827a78e10c6f12f44b41063b3480c57b485a5daea67" Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.714400 4766 scope.go:117] "RemoveContainer" containerID="9e0ffa5e614b56b5972b597f5d7cf1d69c9755a6edfc1db6ac05034cb4e3d62d" Nov 26 00:47:50 crc kubenswrapper[4766]: E1126 00:47:50.715043 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=operator pod=rabbitmq-cluster-operator-manager-668c99d594-mx9vw_openstack-operators(91ee52b5-0e2e-4732-87bf-809e378d33d3)\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw" podUID="91ee52b5-0e2e-4732-87bf-809e378d33d3" Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.728327 4766 generic.go:334] "Generic (PLEG): container finished" podID="00a32e6c-edcc-4625-ac35-26fed19f1638" containerID="90575fc710573493e63d6004e10aaf0090eed703ac12ecf62af5a2bbbe5bff23" exitCode=1 Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.728819 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" event={"ID":"00a32e6c-edcc-4625-ac35-26fed19f1638","Type":"ContainerDied","Data":"90575fc710573493e63d6004e10aaf0090eed703ac12ecf62af5a2bbbe5bff23"} Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.729739 4766 scope.go:117] "RemoveContainer" containerID="90575fc710573493e63d6004e10aaf0090eed703ac12ecf62af5a2bbbe5bff23" Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.731046 4766 generic.go:334] "Generic (PLEG): container finished" podID="d00b9c60-0fb7-49c6-944a-fac598cc22a4" containerID="35bfe8abff8194805e0514db2803bd9c014d0cb93366a09b57b6684dc470002d" exitCode=1 Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.731148 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" event={"ID":"d00b9c60-0fb7-49c6-944a-fac598cc22a4","Type":"ContainerDied","Data":"35bfe8abff8194805e0514db2803bd9c014d0cb93366a09b57b6684dc470002d"} Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.731698 4766 scope.go:117] "RemoveContainer" containerID="35bfe8abff8194805e0514db2803bd9c014d0cb93366a09b57b6684dc470002d" Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.733509 4766 generic.go:334] "Generic (PLEG): container finished" podID="0e6cfef3-86af-4cf0-ab06-b3eef687836b" containerID="b27130c628e9b6ed27e2271bbe305fc8f86a893172736a4bf2cd4a9053f2dc60" exitCode=1 Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.733558 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" event={"ID":"0e6cfef3-86af-4cf0-ab06-b3eef687836b","Type":"ContainerDied","Data":"b27130c628e9b6ed27e2271bbe305fc8f86a893172736a4bf2cd4a9053f2dc60"} Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.735229 4766 scope.go:117] "RemoveContainer" containerID="b27130c628e9b6ed27e2271bbe305fc8f86a893172736a4bf2cd4a9053f2dc60" Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.737447 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" event={"ID":"31b1b73b-1d10-49ac-8549-b202c965b63e","Type":"ContainerDied","Data":"f1706c458c961181fe0f0079d94cf794bd65e747edbd15a93fd1d4c83ed950ce"} Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.738202 4766 scope.go:117] "RemoveContainer" containerID="f1706c458c961181fe0f0079d94cf794bd65e747edbd15a93fd1d4c83ed950ce" Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.737397 4766 generic.go:334] "Generic (PLEG): container finished" podID="31b1b73b-1d10-49ac-8549-b202c965b63e" containerID="f1706c458c961181fe0f0079d94cf794bd65e747edbd15a93fd1d4c83ed950ce" exitCode=1 Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.746525 4766 generic.go:334] "Generic (PLEG): container finished" podID="97419f32-8768-495f-a652-8a1e8701546f" containerID="f4b0af14fec9ed185546e3a6bc7c56dad525b00ceaba8d4801f280f5b12cb82e" exitCode=1 Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.746619 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" event={"ID":"97419f32-8768-495f-a652-8a1e8701546f","Type":"ContainerDied","Data":"f4b0af14fec9ed185546e3a6bc7c56dad525b00ceaba8d4801f280f5b12cb82e"} Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.747718 4766 scope.go:117] "RemoveContainer" containerID="f4b0af14fec9ed185546e3a6bc7c56dad525b00ceaba8d4801f280f5b12cb82e" Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.749532 4766 generic.go:334] "Generic (PLEG): container finished" podID="a16c5563-750f-4e91-93bd-4044c656cb88" containerID="7a8ccd36fe913294b5a86afc0b6c563f7643d1547af8414850c17e3801762226" exitCode=1 Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.749632 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn" event={"ID":"a16c5563-750f-4e91-93bd-4044c656cb88","Type":"ContainerDied","Data":"7a8ccd36fe913294b5a86afc0b6c563f7643d1547af8414850c17e3801762226"} Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.750519 4766 scope.go:117] "RemoveContainer" containerID="7a8ccd36fe913294b5a86afc0b6c563f7643d1547af8414850c17e3801762226" Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.751836 4766 generic.go:334] "Generic (PLEG): container finished" podID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" containerID="ec396350a1da311b232393c00a8651565f79130843bddb720cbd326d3df4f5dd" exitCode=2 Nov 26 00:47:50 crc kubenswrapper[4766]: I1126 00:47:50.751872 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"aa692ff7-2356-4ff4-b9fe-08884e4081cf","Type":"ContainerDied","Data":"ec396350a1da311b232393c00a8651565f79130843bddb720cbd326d3df4f5dd"} Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.468185 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.468496 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.505064 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.505468 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.614763 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" podUID="45353fc4-9955-460b-9276-9792dfdaf8f4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.97:8081/readyz\": dial tcp 10.217.0.97:8081: connect: connection refused" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.614779 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" podUID="45353fc4-9955-460b-9276-9792dfdaf8f4" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.97:8081/healthz\": dial tcp 10.217.0.97:8081: connect: connection refused" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.688872 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.689186 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.689996 4766 scope.go:117] "RemoveContainer" containerID="c4b508111353f64ce1118a8ac912e3d512d2f23b317dd6a0483c4fe2e0126067" Nov 26 00:47:51 crc kubenswrapper[4766]: E1126 00:47:51.690436 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=horizon-operator-controller-manager-5d494799bf-h4mrn_openstack-operators(5284ad70-a36c-4666-8202-84633d83f50b)\"" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" podUID="5284ad70-a36c-4666-8202-84633d83f50b" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.761704 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.761754 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.762600 4766 scope.go:117] "RemoveContainer" containerID="b6766bbab2729ba50956a4e98254bd688babe7e7ba968cdd1ea57ef3ea991f0e" Nov 26 00:47:51 crc kubenswrapper[4766]: E1126 00:47:51.762966 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=barbican-operator-controller-manager-7b64f4fb85-gbdhz_openstack-operators(cc87f631-ba4e-4943-b51d-bf1012cca754)\"" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" podUID="cc87f631-ba4e-4943-b51d-bf1012cca754" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.765700 4766 generic.go:334] "Generic (PLEG): container finished" podID="150d5aa9-549d-4b43-b5c3-e6c81e942146" containerID="8fa8e16236e2c4d7e14a8760e1cba9898e839125d1530d58220f2a9afcd34ba3" exitCode=1 Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.765776 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" event={"ID":"150d5aa9-549d-4b43-b5c3-e6c81e942146","Type":"ContainerDied","Data":"8fa8e16236e2c4d7e14a8760e1cba9898e839125d1530d58220f2a9afcd34ba3"} Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.766318 4766 scope.go:117] "RemoveContainer" containerID="8fa8e16236e2c4d7e14a8760e1cba9898e839125d1530d58220f2a9afcd34ba3" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.768301 4766 generic.go:334] "Generic (PLEG): container finished" podID="45353fc4-9955-460b-9276-9792dfdaf8f4" containerID="07f0c7157997d9fbe76f790c8c21c65a71f1d39b9ab836201135a61d80cb901e" exitCode=1 Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.768393 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" event={"ID":"45353fc4-9955-460b-9276-9792dfdaf8f4","Type":"ContainerDied","Data":"07f0c7157997d9fbe76f790c8c21c65a71f1d39b9ab836201135a61d80cb901e"} Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.769072 4766 scope.go:117] "RemoveContainer" containerID="07f0c7157997d9fbe76f790c8c21c65a71f1d39b9ab836201135a61d80cb901e" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.771818 4766 generic.go:334] "Generic (PLEG): container finished" podID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" containerID="15d323458b686dcde64b0ff56a735c5e5d229471e9e5d22670ddf11771f0f923" exitCode=1 Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.771871 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"aa692ff7-2356-4ff4-b9fe-08884e4081cf","Type":"ContainerDied","Data":"15d323458b686dcde64b0ff56a735c5e5d229471e9e5d22670ddf11771f0f923"} Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.771892 4766 scope.go:117] "RemoveContainer" containerID="ec396350a1da311b232393c00a8651565f79130843bddb720cbd326d3df4f5dd" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.772213 4766 scope.go:117] "RemoveContainer" containerID="15d323458b686dcde64b0ff56a735c5e5d229471e9e5d22670ddf11771f0f923" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.780006 4766 generic.go:334] "Generic (PLEG): container finished" podID="0e6cfef3-86af-4cf0-ab06-b3eef687836b" containerID="432b9d83f070437cf10cedbd328af7913bf474cb30fe8c3a32ef30213f4f821d" exitCode=1 Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.780085 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" event={"ID":"0e6cfef3-86af-4cf0-ab06-b3eef687836b","Type":"ContainerDied","Data":"432b9d83f070437cf10cedbd328af7913bf474cb30fe8c3a32ef30213f4f821d"} Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.781886 4766 scope.go:117] "RemoveContainer" containerID="432b9d83f070437cf10cedbd328af7913bf474cb30fe8c3a32ef30213f4f821d" Nov 26 00:47:51 crc kubenswrapper[4766]: E1126 00:47:51.782748 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=designate-operator-controller-manager-955677c94-dqtcj_openstack-operators(0e6cfef3-86af-4cf0-ab06-b3eef687836b)\"" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" podUID="0e6cfef3-86af-4cf0-ab06-b3eef687836b" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.796452 4766 generic.go:334] "Generic (PLEG): container finished" podID="31b1b73b-1d10-49ac-8549-b202c965b63e" containerID="bc24cd655f5baac2de315e4c87491860ef20a923a154f22b51ad9e0ca6825752" exitCode=1 Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.796715 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" event={"ID":"31b1b73b-1d10-49ac-8549-b202c965b63e","Type":"ContainerDied","Data":"bc24cd655f5baac2de315e4c87491860ef20a923a154f22b51ad9e0ca6825752"} Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.798089 4766 scope.go:117] "RemoveContainer" containerID="bc24cd655f5baac2de315e4c87491860ef20a923a154f22b51ad9e0ca6825752" Nov 26 00:47:51 crc kubenswrapper[4766]: E1126 00:47:51.798859 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=cinder-operator-controller-manager-6b7f75547b-jgz6c_openstack-operators(31b1b73b-1d10-49ac-8549-b202c965b63e)\"" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" podUID="31b1b73b-1d10-49ac-8549-b202c965b63e" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.805884 4766 generic.go:334] "Generic (PLEG): container finished" podID="97419f32-8768-495f-a652-8a1e8701546f" containerID="8ce23e89748eaf63d2ede32d0d0f586ce2050c357cf7ddf77eebed117cb8a3da" exitCode=1 Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.805952 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" event={"ID":"97419f32-8768-495f-a652-8a1e8701546f","Type":"ContainerDied","Data":"8ce23e89748eaf63d2ede32d0d0f586ce2050c357cf7ddf77eebed117cb8a3da"} Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.807005 4766 scope.go:117] "RemoveContainer" containerID="8ce23e89748eaf63d2ede32d0d0f586ce2050c357cf7ddf77eebed117cb8a3da" Nov 26 00:47:51 crc kubenswrapper[4766]: E1126 00:47:51.807361 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=placement-operator-controller-manager-57988cc5b5-xlxgj_openstack-operators(97419f32-8768-495f-a652-8a1e8701546f)\"" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" podUID="97419f32-8768-495f-a652-8a1e8701546f" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.812165 4766 generic.go:334] "Generic (PLEG): container finished" podID="00a32e6c-edcc-4625-ac35-26fed19f1638" containerID="0c41858ad420d326d413541187c4331ae0d60df9b0ce9ce39bd4e1c69d9e0536" exitCode=1 Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.812418 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" event={"ID":"00a32e6c-edcc-4625-ac35-26fed19f1638","Type":"ContainerDied","Data":"0c41858ad420d326d413541187c4331ae0d60df9b0ce9ce39bd4e1c69d9e0536"} Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.813346 4766 scope.go:117] "RemoveContainer" containerID="0c41858ad420d326d413541187c4331ae0d60df9b0ce9ce39bd4e1c69d9e0536" Nov 26 00:47:51 crc kubenswrapper[4766]: E1126 00:47:51.813805 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=swift-operator-controller-manager-d77b94747-2t9w2_openstack-operators(00a32e6c-edcc-4625-ac35-26fed19f1638)\"" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" podUID="00a32e6c-edcc-4625-ac35-26fed19f1638" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.815957 4766 generic.go:334] "Generic (PLEG): container finished" podID="a16c5563-750f-4e91-93bd-4044c656cb88" containerID="624e9ac049eb714839dab1ba6bf0a2082221a591770ac0e439ad6f283f024f43" exitCode=1 Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.816062 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn" event={"ID":"a16c5563-750f-4e91-93bd-4044c656cb88","Type":"ContainerDied","Data":"624e9ac049eb714839dab1ba6bf0a2082221a591770ac0e439ad6f283f024f43"} Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.817704 4766 scope.go:117] "RemoveContainer" containerID="624e9ac049eb714839dab1ba6bf0a2082221a591770ac0e439ad6f283f024f43" Nov 26 00:47:51 crc kubenswrapper[4766]: E1126 00:47:51.818070 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=manila-operator-controller-manager-5d499bf58b-rs4kn_openstack-operators(a16c5563-750f-4e91-93bd-4044c656cb88)\"" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn" podUID="a16c5563-750f-4e91-93bd-4044c656cb88" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.820456 4766 generic.go:334] "Generic (PLEG): container finished" podID="d00b9c60-0fb7-49c6-944a-fac598cc22a4" containerID="05fe629429cc17682b24434acacc6d08fc24ebc0d20e7008292b71d64a9c5529" exitCode=1 Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.820545 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" event={"ID":"d00b9c60-0fb7-49c6-944a-fac598cc22a4","Type":"ContainerDied","Data":"05fe629429cc17682b24434acacc6d08fc24ebc0d20e7008292b71d64a9c5529"} Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.824721 4766 scope.go:117] "RemoveContainer" containerID="05fe629429cc17682b24434acacc6d08fc24ebc0d20e7008292b71d64a9c5529" Nov 26 00:47:51 crc kubenswrapper[4766]: E1126 00:47:51.825051 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=openstack-operator-controller-manager-86674bdb4b-vq98m_openstack-operators(d00b9c60-0fb7-49c6-944a-fac598cc22a4)\"" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" podUID="d00b9c60-0fb7-49c6-944a-fac598cc22a4" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.825409 4766 scope.go:117] "RemoveContainer" containerID="c4b508111353f64ce1118a8ac912e3d512d2f23b317dd6a0483c4fe2e0126067" Nov 26 00:47:51 crc kubenswrapper[4766]: E1126 00:47:51.825716 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=horizon-operator-controller-manager-5d494799bf-h4mrn_openstack-operators(5284ad70-a36c-4666-8202-84633d83f50b)\"" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" podUID="5284ad70-a36c-4666-8202-84633d83f50b" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.851205 4766 scope.go:117] "RemoveContainer" containerID="b27130c628e9b6ed27e2271bbe305fc8f86a893172736a4bf2cd4a9053f2dc60" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.857557 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.857600 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.858349 4766 scope.go:117] "RemoveContainer" containerID="5f7d72a8709e834d6b16398a4a94f1cdb096b975fefcb4fa32830decc9e3a380" Nov 26 00:47:51 crc kubenswrapper[4766]: E1126 00:47:51.858613 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=heat-operator-controller-manager-5b77f656f-tkj9b_openstack-operators(1ecbd1d2-cb7f-432f-bc75-903d0643ea29)\"" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" podUID="1ecbd1d2-cb7f-432f-bc75-903d0643ea29" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.991398 4766 scope.go:117] "RemoveContainer" containerID="f1706c458c961181fe0f0079d94cf794bd65e747edbd15a93fd1d4c83ed950ce" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.996993 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" Nov 26 00:47:51 crc kubenswrapper[4766]: I1126 00:47:51.997738 4766 scope.go:117] "RemoveContainer" containerID="a2540ad4bcec8692a937d7029a6100f622c38a2889d9078548875630e3de8431" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.029022 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.029060 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.029734 4766 scope.go:117] "RemoveContainer" containerID="ccb6a5c90a0ad287ea4fde750c373ddb97fe844e0b2024e905b73d849ad7c7d6" Nov 26 00:47:52 crc kubenswrapper[4766]: E1126 00:47:52.029956 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=keystone-operator-controller-manager-7b4567c7cf-fcr76_openstack-operators(3d0d59b2-0486-40a2-ae60-b738a4f76dc4)\"" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" podUID="3d0d59b2-0486-40a2-ae60-b738a4f76dc4" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.078989 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.079082 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.107027 4766 scope.go:117] "RemoveContainer" containerID="f4b0af14fec9ed185546e3a6bc7c56dad525b00ceaba8d4801f280f5b12cb82e" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.110773 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.110922 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.111628 4766 scope.go:117] "RemoveContainer" containerID="502141878d9557d404939737707070c203f202ba93a6d5aa04240a14a618e040" Nov 26 00:47:52 crc kubenswrapper[4766]: E1126 00:47:52.111935 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=mariadb-operator-controller-manager-66f4dd4bc7-z865g_openstack-operators(749a25e6-23eb-4fdd-8e15-ac2874cdee7f)\"" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" podUID="749a25e6-23eb-4fdd-8e15-ac2874cdee7f" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.145765 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.146103 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.147268 4766 scope.go:117] "RemoveContainer" containerID="5d7d2b957bbb4841fc8ab537bef654cacd24c482167164e785eed6faba12e21f" Nov 26 00:47:52 crc kubenswrapper[4766]: E1126 00:47:52.147694 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=neutron-operator-controller-manager-6fdcddb789-wdzgw_openstack-operators(2bb23a3b-1f4a-409f-9d9c-0afe73a33352)\"" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" podUID="2bb23a3b-1f4a-409f-9d9c-0afe73a33352" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.151359 4766 scope.go:117] "RemoveContainer" containerID="90575fc710573493e63d6004e10aaf0090eed703ac12ecf62af5a2bbbe5bff23" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.163717 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.163982 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.164838 4766 scope.go:117] "RemoveContainer" containerID="69ed96c12e25467c32b6ad99ca52ad4f46ecc9f3be096414f5b261a00874e013" Nov 26 00:47:52 crc kubenswrapper[4766]: E1126 00:47:52.165167 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=nova-operator-controller-manager-79556f57fc-dm7xb_openstack-operators(8969652f-14c9-4618-ad2d-1b3fd1ad3627)\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" podUID="8969652f-14c9-4618-ad2d-1b3fd1ad3627" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.201769 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.202415 4766 scope.go:117] "RemoveContainer" containerID="571408ef2aef1a0e6b0f38d750793734c21c57756cc64b5fd32d672f6b4e07df" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.214601 4766 scope.go:117] "RemoveContainer" containerID="7a8ccd36fe913294b5a86afc0b6c563f7643d1547af8414850c17e3801762226" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.221874 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.222430 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.223271 4766 scope.go:117] "RemoveContainer" containerID="3a5bf6a88bd14fa3df408b57df7d3930b7c4f71f2d70788e373686f0aad92b30" Nov 26 00:47:52 crc kubenswrapper[4766]: E1126 00:47:52.223551 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=ovn-operator-controller-manager-56897c768d-9zhql_openstack-operators(a2e7c098-bf9d-4d99-84bf-adfb49ee6b12)\"" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" podUID="a2e7c098-bf9d-4d99-84bf-adfb49ee6b12" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.243461 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.243510 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.262291 4766 scope.go:117] "RemoveContainer" containerID="35bfe8abff8194805e0514db2803bd9c014d0cb93366a09b57b6684dc470002d" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.353875 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.354328 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.564093 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.564155 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.564987 4766 scope.go:117] "RemoveContainer" containerID="089ff57a19d572ad607d9b8aefbeca178c8046089d9d0f1a2b360a9cd51d8e1b" Nov 26 00:47:52 crc kubenswrapper[4766]: E1126 00:47:52.565371 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=telemetry-operator-controller-manager-5ccf99cfdd-9cfft_openstack-operators(04e53d09-de84-4682-b17b-61f9ea7bf152)\"" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" podUID="04e53d09-de84-4682-b17b-61f9ea7bf152" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.628304 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.628571 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.629730 4766 scope.go:117] "RemoveContainer" containerID="37c98bbe0d1a1d7dec9f2442d2c16ebe49c7485725817b532cbc38bae4365252" Nov 26 00:47:52 crc kubenswrapper[4766]: E1126 00:47:52.630183 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=watcher-operator-controller-manager-656dcb59d4-2hpg8_openstack-operators(751b8e60-46d4-4f5e-b62d-800f641ccd40)\"" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" podUID="751b8e60-46d4-4f5e-b62d-800f641ccd40" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.741149 4766 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-v4fsg" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.840540 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" event={"ID":"150d5aa9-549d-4b43-b5c3-e6c81e942146","Type":"ContainerStarted","Data":"62c2f89183d5e95cf65e4a38a68273e334f91a4f2567c2245816862d52d8b830"} Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.841584 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.843813 4766 scope.go:117] "RemoveContainer" containerID="8ce23e89748eaf63d2ede32d0d0f586ce2050c357cf7ddf77eebed117cb8a3da" Nov 26 00:47:52 crc kubenswrapper[4766]: E1126 00:47:52.844118 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=placement-operator-controller-manager-57988cc5b5-xlxgj_openstack-operators(97419f32-8768-495f-a652-8a1e8701546f)\"" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" podUID="97419f32-8768-495f-a652-8a1e8701546f" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.845576 4766 generic.go:334] "Generic (PLEG): container finished" podID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" containerID="fc1da18b2031a9683dec239eb1c313b14139ca65098b96f7294b63e1da1f01ac" exitCode=1 Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.845693 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"aa692ff7-2356-4ff4-b9fe-08884e4081cf","Type":"ContainerDied","Data":"fc1da18b2031a9683dec239eb1c313b14139ca65098b96f7294b63e1da1f01ac"} Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.845780 4766 scope.go:117] "RemoveContainer" containerID="15d323458b686dcde64b0ff56a735c5e5d229471e9e5d22670ddf11771f0f923" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.846215 4766 scope.go:117] "RemoveContainer" containerID="fc1da18b2031a9683dec239eb1c313b14139ca65098b96f7294b63e1da1f01ac" Nov 26 00:47:52 crc kubenswrapper[4766]: E1126 00:47:52.846493 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-state-metrics pod=kube-state-metrics-0_openstack(aa692ff7-2356-4ff4-b9fe-08884e4081cf)\"" pod="openstack/kube-state-metrics-0" podUID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.851670 4766 generic.go:334] "Generic (PLEG): container finished" podID="45353fc4-9955-460b-9276-9792dfdaf8f4" containerID="05a3f0845ceb0e6ddd27c7537b01703da19f1edc1685d2a382208a93ba7d72a2" exitCode=1 Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.851735 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" event={"ID":"45353fc4-9955-460b-9276-9792dfdaf8f4","Type":"ContainerDied","Data":"05a3f0845ceb0e6ddd27c7537b01703da19f1edc1685d2a382208a93ba7d72a2"} Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.852369 4766 scope.go:117] "RemoveContainer" containerID="05a3f0845ceb0e6ddd27c7537b01703da19f1edc1685d2a382208a93ba7d72a2" Nov 26 00:47:52 crc kubenswrapper[4766]: E1126 00:47:52.852630 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=glance-operator-controller-manager-589cbd6b5b-k68q9_openstack-operators(45353fc4-9955-460b-9276-9792dfdaf8f4)\"" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" podUID="45353fc4-9955-460b-9276-9792dfdaf8f4" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.857722 4766 generic.go:334] "Generic (PLEG): container finished" podID="08bd3753-3b70-44f8-92ff-98bedd88f662" containerID="f852f3819cc2e23f970fe6d5410e8b916b08f77990d1cf12bdeb13de0d480f2d" exitCode=1 Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.857786 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" event={"ID":"08bd3753-3b70-44f8-92ff-98bedd88f662","Type":"ContainerDied","Data":"f852f3819cc2e23f970fe6d5410e8b916b08f77990d1cf12bdeb13de0d480f2d"} Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.858453 4766 scope.go:117] "RemoveContainer" containerID="f852f3819cc2e23f970fe6d5410e8b916b08f77990d1cf12bdeb13de0d480f2d" Nov 26 00:47:52 crc kubenswrapper[4766]: E1126 00:47:52.858777 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=ironic-operator-controller-manager-67cb4dc6d4-qtnxq_openstack-operators(08bd3753-3b70-44f8-92ff-98bedd88f662)\"" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" podUID="08bd3753-3b70-44f8-92ff-98bedd88f662" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.861056 4766 scope.go:117] "RemoveContainer" containerID="0c41858ad420d326d413541187c4331ae0d60df9b0ce9ce39bd4e1c69d9e0536" Nov 26 00:47:52 crc kubenswrapper[4766]: E1126 00:47:52.861259 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=swift-operator-controller-manager-d77b94747-2t9w2_openstack-operators(00a32e6c-edcc-4625-ac35-26fed19f1638)\"" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" podUID="00a32e6c-edcc-4625-ac35-26fed19f1638" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.863056 4766 generic.go:334] "Generic (PLEG): container finished" podID="91d3e9c8-8bbc-43b9-947b-71c4bb474b21" containerID="e8fbaea76abae029d1a7e07d17d13ba1ef29a28831a8d0e6f6fc04e295ff065e" exitCode=1 Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.863117 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" event={"ID":"91d3e9c8-8bbc-43b9-947b-71c4bb474b21","Type":"ContainerDied","Data":"e8fbaea76abae029d1a7e07d17d13ba1ef29a28831a8d0e6f6fc04e295ff065e"} Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.863719 4766 scope.go:117] "RemoveContainer" containerID="e8fbaea76abae029d1a7e07d17d13ba1ef29a28831a8d0e6f6fc04e295ff065e" Nov 26 00:47:52 crc kubenswrapper[4766]: E1126 00:47:52.863957 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=octavia-operator-controller-manager-64cdc6ff96-bfw5f_openstack-operators(91d3e9c8-8bbc-43b9-947b-71c4bb474b21)\"" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" podUID="91d3e9c8-8bbc-43b9-947b-71c4bb474b21" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.866120 4766 scope.go:117] "RemoveContainer" containerID="624e9ac049eb714839dab1ba6bf0a2082221a591770ac0e439ad6f283f024f43" Nov 26 00:47:52 crc kubenswrapper[4766]: E1126 00:47:52.866357 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=manila-operator-controller-manager-5d499bf58b-rs4kn_openstack-operators(a16c5563-750f-4e91-93bd-4044c656cb88)\"" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn" podUID="a16c5563-750f-4e91-93bd-4044c656cb88" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.869449 4766 scope.go:117] "RemoveContainer" containerID="432b9d83f070437cf10cedbd328af7913bf474cb30fe8c3a32ef30213f4f821d" Nov 26 00:47:52 crc kubenswrapper[4766]: E1126 00:47:52.869680 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=designate-operator-controller-manager-955677c94-dqtcj_openstack-operators(0e6cfef3-86af-4cf0-ab06-b3eef687836b)\"" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" podUID="0e6cfef3-86af-4cf0-ab06-b3eef687836b" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.878713 4766 scope.go:117] "RemoveContainer" containerID="502141878d9557d404939737707070c203f202ba93a6d5aa04240a14a618e040" Nov 26 00:47:52 crc kubenswrapper[4766]: E1126 00:47:52.878969 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=mariadb-operator-controller-manager-66f4dd4bc7-z865g_openstack-operators(749a25e6-23eb-4fdd-8e15-ac2874cdee7f)\"" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" podUID="749a25e6-23eb-4fdd-8e15-ac2874cdee7f" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.879292 4766 scope.go:117] "RemoveContainer" containerID="bc24cd655f5baac2de315e4c87491860ef20a923a154f22b51ad9e0ca6825752" Nov 26 00:47:52 crc kubenswrapper[4766]: E1126 00:47:52.879506 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=cinder-operator-controller-manager-6b7f75547b-jgz6c_openstack-operators(31b1b73b-1d10-49ac-8549-b202c965b63e)\"" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" podUID="31b1b73b-1d10-49ac-8549-b202c965b63e" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.879866 4766 scope.go:117] "RemoveContainer" containerID="69ed96c12e25467c32b6ad99ca52ad4f46ecc9f3be096414f5b261a00874e013" Nov 26 00:47:52 crc kubenswrapper[4766]: E1126 00:47:52.880103 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=nova-operator-controller-manager-79556f57fc-dm7xb_openstack-operators(8969652f-14c9-4618-ad2d-1b3fd1ad3627)\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" podUID="8969652f-14c9-4618-ad2d-1b3fd1ad3627" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.880515 4766 scope.go:117] "RemoveContainer" containerID="37c98bbe0d1a1d7dec9f2442d2c16ebe49c7485725817b532cbc38bae4365252" Nov 26 00:47:52 crc kubenswrapper[4766]: E1126 00:47:52.880800 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=watcher-operator-controller-manager-656dcb59d4-2hpg8_openstack-operators(751b8e60-46d4-4f5e-b62d-800f641ccd40)\"" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" podUID="751b8e60-46d4-4f5e-b62d-800f641ccd40" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.880987 4766 scope.go:117] "RemoveContainer" containerID="3a5bf6a88bd14fa3df408b57df7d3930b7c4f71f2d70788e373686f0aad92b30" Nov 26 00:47:52 crc kubenswrapper[4766]: E1126 00:47:52.881445 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=ovn-operator-controller-manager-56897c768d-9zhql_openstack-operators(a2e7c098-bf9d-4d99-84bf-adfb49ee6b12)\"" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" podUID="a2e7c098-bf9d-4d99-84bf-adfb49ee6b12" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.915954 4766 scope.go:117] "RemoveContainer" containerID="07f0c7157997d9fbe76f790c8c21c65a71f1d39b9ab836201135a61d80cb901e" Nov 26 00:47:52 crc kubenswrapper[4766]: I1126 00:47:52.986527 4766 scope.go:117] "RemoveContainer" containerID="a2540ad4bcec8692a937d7029a6100f622c38a2889d9078548875630e3de8431" Nov 26 00:47:53 crc kubenswrapper[4766]: I1126 00:47:53.062358 4766 scope.go:117] "RemoveContainer" containerID="571408ef2aef1a0e6b0f38d750793734c21c57756cc64b5fd32d672f6b4e07df" Nov 26 00:47:53 crc kubenswrapper[4766]: I1126 00:47:53.842377 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:47:53 crc kubenswrapper[4766]: I1126 00:47:53.843128 4766 scope.go:117] "RemoveContainer" containerID="f4f0be8205cb3c087b28e7451a802ae7e7912ffbab226d356ad6b10cb884ef49" Nov 26 00:47:53 crc kubenswrapper[4766]: E1126 00:47:53.843417 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=infra-operator-controller-manager-57548d458d-x6h4v_openstack-operators(c3cec209-fdef-44a6-90d0-2d06cfd7e5cc)\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" podUID="c3cec209-fdef-44a6-90d0-2d06cfd7e5cc" Nov 26 00:47:53 crc kubenswrapper[4766]: I1126 00:47:53.895738 4766 scope.go:117] "RemoveContainer" containerID="fc1da18b2031a9683dec239eb1c313b14139ca65098b96f7294b63e1da1f01ac" Nov 26 00:47:53 crc kubenswrapper[4766]: E1126 00:47:53.896016 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-state-metrics pod=kube-state-metrics-0_openstack(aa692ff7-2356-4ff4-b9fe-08884e4081cf)\"" pod="openstack/kube-state-metrics-0" podUID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" Nov 26 00:47:53 crc kubenswrapper[4766]: I1126 00:47:53.897136 4766 scope.go:117] "RemoveContainer" containerID="8ce23e89748eaf63d2ede32d0d0f586ce2050c357cf7ddf77eebed117cb8a3da" Nov 26 00:47:53 crc kubenswrapper[4766]: I1126 00:47:53.897328 4766 scope.go:117] "RemoveContainer" containerID="624e9ac049eb714839dab1ba6bf0a2082221a591770ac0e439ad6f283f024f43" Nov 26 00:47:53 crc kubenswrapper[4766]: E1126 00:47:53.897342 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=placement-operator-controller-manager-57988cc5b5-xlxgj_openstack-operators(97419f32-8768-495f-a652-8a1e8701546f)\"" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" podUID="97419f32-8768-495f-a652-8a1e8701546f" Nov 26 00:47:53 crc kubenswrapper[4766]: E1126 00:47:53.897601 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=manila-operator-controller-manager-5d499bf58b-rs4kn_openstack-operators(a16c5563-750f-4e91-93bd-4044c656cb88)\"" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn" podUID="a16c5563-750f-4e91-93bd-4044c656cb88" Nov 26 00:47:54 crc kubenswrapper[4766]: I1126 00:47:54.798992 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 26 00:47:55 crc kubenswrapper[4766]: I1126 00:47:55.809857 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 26 00:47:56 crc kubenswrapper[4766]: I1126 00:47:56.137058 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 26 00:47:56 crc kubenswrapper[4766]: I1126 00:47:56.141906 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 26 00:47:56 crc kubenswrapper[4766]: I1126 00:47:56.425746 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 26 00:47:56 crc kubenswrapper[4766]: I1126 00:47:56.503729 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 26 00:47:56 crc kubenswrapper[4766]: I1126 00:47:56.729816 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Nov 26 00:47:57 crc kubenswrapper[4766]: I1126 00:47:57.029316 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 26 00:47:57 crc kubenswrapper[4766]: I1126 00:47:57.344943 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 26 00:47:57 crc kubenswrapper[4766]: I1126 00:47:57.389774 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-http" Nov 26 00:47:57 crc kubenswrapper[4766]: I1126 00:47:57.434165 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 26 00:47:57 crc kubenswrapper[4766]: I1126 00:47:57.771169 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 26 00:47:57 crc kubenswrapper[4766]: I1126 00:47:57.849358 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:47:57 crc kubenswrapper[4766]: I1126 00:47:57.849999 4766 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:47:57 crc kubenswrapper[4766]: I1126 00:47:57.850031 4766 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:47:57 crc kubenswrapper[4766]: I1126 00:47:57.855633 4766 scope.go:117] "RemoveContainer" containerID="c9d7eda97f15fe6f9a75ca8a20ea0e2e59d89ec2c8fa0c399aa26ed106e094ec" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.030242 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.095489 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.155116 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.203698 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.220127 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.243676 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.244396 4766 scope.go:117] "RemoveContainer" containerID="05fe629429cc17682b24434acacc6d08fc24ebc0d20e7008292b71d64a9c5529" Nov 26 00:47:58 crc kubenswrapper[4766]: E1126 00:47:58.244677 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=openstack-operator-controller-manager-86674bdb4b-vq98m_openstack-operators(d00b9c60-0fb7-49c6-944a-fac598cc22a4)\"" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" podUID="d00b9c60-0fb7-49c6-944a-fac598cc22a4" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.245236 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.296734 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.311573 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.368783 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-http" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.375067 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.386839 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-k8x7n" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.440275 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.532768 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.591466 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.628265 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.632900 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.702964 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.709622 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-44jtp" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.790522 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.827292 4766 scope.go:117] "RemoveContainer" containerID="0db2683335300616e0363c5bc5ce375b5fd490a89a25d39263b201e2816d69ec" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.944808 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.961467 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_71bb4a3aecc4ba5b26c4b7318770ce13/kube-apiserver-check-endpoints/2.log" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.962111 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_71bb4a3aecc4ba5b26c4b7318770ce13/kube-apiserver-check-endpoints/1.log" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.963640 4766 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="0c594df0d9660266fff1417a3f4f615f9c3a9b37b90eeef4b4d18d0d955a98a2" exitCode=255 Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.963679 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"0c594df0d9660266fff1417a3f4f615f9c3a9b37b90eeef4b4d18d0d955a98a2"} Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.963733 4766 scope.go:117] "RemoveContainer" containerID="c9d7eda97f15fe6f9a75ca8a20ea0e2e59d89ec2c8fa0c399aa26ed106e094ec" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.963991 4766 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.964009 4766 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.968048 4766 scope.go:117] "RemoveContainer" containerID="0c594df0d9660266fff1417a3f4f615f9c3a9b37b90eeef4b4d18d0d955a98a2" Nov 26 00:47:58 crc kubenswrapper[4766]: E1126 00:47:58.968378 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(71bb4a3aecc4ba5b26c4b7318770ce13)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.989017 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 26 00:47:58 crc kubenswrapper[4766]: I1126 00:47:58.999313 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 26 00:47:59 crc kubenswrapper[4766]: I1126 00:47:59.052086 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 26 00:47:59 crc kubenswrapper[4766]: I1126 00:47:59.077880 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7d5fdccdc4-5sh97" Nov 26 00:47:59 crc kubenswrapper[4766]: I1126 00:47:59.137480 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 26 00:47:59 crc kubenswrapper[4766]: I1126 00:47:59.284340 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 26 00:47:59 crc kubenswrapper[4766]: I1126 00:47:59.308282 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 26 00:47:59 crc kubenswrapper[4766]: I1126 00:47:59.382827 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 26 00:47:59 crc kubenswrapper[4766]: I1126 00:47:59.473082 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 26 00:47:59 crc kubenswrapper[4766]: I1126 00:47:59.552236 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 26 00:47:59 crc kubenswrapper[4766]: I1126 00:47:59.553384 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 26 00:47:59 crc kubenswrapper[4766]: I1126 00:47:59.609714 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 26 00:47:59 crc kubenswrapper[4766]: I1126 00:47:59.630490 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 26 00:47:59 crc kubenswrapper[4766]: I1126 00:47:59.668181 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-grpc" Nov 26 00:47:59 crc kubenswrapper[4766]: I1126 00:47:59.755539 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 26 00:47:59 crc kubenswrapper[4766]: I1126 00:47:59.868920 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 26 00:47:59 crc kubenswrapper[4766]: I1126 00:47:59.935540 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 26 00:47:59 crc kubenswrapper[4766]: I1126 00:47:59.975604 4766 generic.go:334] "Generic (PLEG): container finished" podID="fd3a6275-9f47-4799-bc15-96a51b93f31b" containerID="6396f3caca85c89b1192e1772820e1b1f5623132fa5baf5b852754febe7243de" exitCode=1 Nov 26 00:47:59 crc kubenswrapper[4766]: I1126 00:47:59.975698 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" event={"ID":"fd3a6275-9f47-4799-bc15-96a51b93f31b","Type":"ContainerDied","Data":"6396f3caca85c89b1192e1772820e1b1f5623132fa5baf5b852754febe7243de"} Nov 26 00:47:59 crc kubenswrapper[4766]: I1126 00:47:59.975779 4766 scope.go:117] "RemoveContainer" containerID="0db2683335300616e0363c5bc5ce375b5fd490a89a25d39263b201e2816d69ec" Nov 26 00:47:59 crc kubenswrapper[4766]: I1126 00:47:59.976429 4766 scope.go:117] "RemoveContainer" containerID="6396f3caca85c89b1192e1772820e1b1f5623132fa5baf5b852754febe7243de" Nov 26 00:47:59 crc kubenswrapper[4766]: E1126 00:47:59.976742 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=metallb-operator-controller-manager-764c8cf56f-vrwpt_metallb-system(fd3a6275-9f47-4799-bc15-96a51b93f31b)\"" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" podUID="fd3a6275-9f47-4799-bc15-96a51b93f31b" Nov 26 00:47:59 crc kubenswrapper[4766]: I1126 00:47:59.979417 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_71bb4a3aecc4ba5b26c4b7318770ce13/kube-apiserver-check-endpoints/2.log" Nov 26 00:47:59 crc kubenswrapper[4766]: I1126 00:47:59.999729 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.104912 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.117004 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.151715 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.192488 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.279085 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.283217 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.292204 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.294721 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.351560 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.358114 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-5b4dx" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.399403 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.420356 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.420416 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/kube-state-metrics-0" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.421127 4766 scope.go:117] "RemoveContainer" containerID="fc1da18b2031a9683dec239eb1c313b14139ca65098b96f7294b63e1da1f01ac" Nov 26 00:48:00 crc kubenswrapper[4766]: E1126 00:48:00.421370 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-state-metrics pod=kube-state-metrics-0_openstack(aa692ff7-2356-4ff4-b9fe-08884e4081cf)\"" pod="openstack/kube-state-metrics-0" podUID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.424788 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.435168 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.437513 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.481849 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-glg62" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.509352 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.557521 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.682267 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.691584 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.705640 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 26 00:48:00 crc kubenswrapper[4766]: I1126 00:48:00.924543 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.100610 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.142179 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.235161 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.239357 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.239792 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-vjfll" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.240224 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.317610 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.347025 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.398319 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.404848 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-t2z57" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.434642 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.455161 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.459315 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.468761 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.469417 4766 scope.go:117] "RemoveContainer" containerID="bc24cd655f5baac2de315e4c87491860ef20a923a154f22b51ad9e0ca6825752" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.490828 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.505049 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.505850 4766 scope.go:117] "RemoveContainer" containerID="432b9d83f070437cf10cedbd328af7913bf474cb30fe8c3a32ef30213f4f821d" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.515728 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.613771 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.614541 4766 scope.go:117] "RemoveContainer" containerID="05a3f0845ceb0e6ddd27c7537b01703da19f1edc1685d2a382208a93ba7d72a2" Nov 26 00:48:01 crc kubenswrapper[4766]: E1126 00:48:01.614913 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=glance-operator-controller-manager-589cbd6b5b-k68q9_openstack-operators(45353fc4-9955-460b-9276-9792dfdaf8f4)\"" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" podUID="45353fc4-9955-460b-9276-9792dfdaf8f4" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.619625 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.619832 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.658184 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-wcjrw" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.679369 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway-ca-bundle" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.770793 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards-sa-dockercfg-lb8kj" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.780960 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.838877 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.841432 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-rjlnb" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.861974 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.878775 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.917278 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-s3" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.961792 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.982644 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.997919 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" Nov 26 00:48:01 crc kubenswrapper[4766]: I1126 00:48:01.998761 4766 scope.go:117] "RemoveContainer" containerID="f852f3819cc2e23f970fe6d5410e8b916b08f77990d1cf12bdeb13de0d480f2d" Nov 26 00:48:01 crc kubenswrapper[4766]: E1126 00:48:01.999017 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=ironic-operator-controller-manager-67cb4dc6d4-qtnxq_openstack-operators(08bd3753-3b70-44f8-92ff-98bedd88f662)\"" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" podUID="08bd3753-3b70-44f8-92ff-98bedd88f662" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.008367 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.016850 4766 generic.go:334] "Generic (PLEG): container finished" podID="0e6cfef3-86af-4cf0-ab06-b3eef687836b" containerID="11900098c4ce70adf98b70c668f6fd3114393fbfb5ebae1324aa13e7994d8001" exitCode=1 Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.016924 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" event={"ID":"0e6cfef3-86af-4cf0-ab06-b3eef687836b","Type":"ContainerDied","Data":"11900098c4ce70adf98b70c668f6fd3114393fbfb5ebae1324aa13e7994d8001"} Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.016974 4766 scope.go:117] "RemoveContainer" containerID="432b9d83f070437cf10cedbd328af7913bf474cb30fe8c3a32ef30213f4f821d" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.018474 4766 scope.go:117] "RemoveContainer" containerID="11900098c4ce70adf98b70c668f6fd3114393fbfb5ebae1324aa13e7994d8001" Nov 26 00:48:02 crc kubenswrapper[4766]: E1126 00:48:02.024458 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=designate-operator-controller-manager-955677c94-dqtcj_openstack-operators(0e6cfef3-86af-4cf0-ab06-b3eef687836b)\"" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" podUID="0e6cfef3-86af-4cf0-ab06-b3eef687836b" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.030188 4766 generic.go:334] "Generic (PLEG): container finished" podID="31b1b73b-1d10-49ac-8549-b202c965b63e" containerID="00bc96df29ffd032af13f751f10f9ebd94973bdaab39c23fa02a669782573988" exitCode=1 Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.030285 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" event={"ID":"31b1b73b-1d10-49ac-8549-b202c965b63e","Type":"ContainerDied","Data":"00bc96df29ffd032af13f751f10f9ebd94973bdaab39c23fa02a669782573988"} Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.031555 4766 scope.go:117] "RemoveContainer" containerID="00bc96df29ffd032af13f751f10f9ebd94973bdaab39c23fa02a669782573988" Nov 26 00:48:02 crc kubenswrapper[4766]: E1126 00:48:02.032190 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=cinder-operator-controller-manager-6b7f75547b-jgz6c_openstack-operators(31b1b73b-1d10-49ac-8549-b202c965b63e)\"" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" podUID="31b1b73b-1d10-49ac-8549-b202c965b63e" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.107726 4766 scope.go:117] "RemoveContainer" containerID="bc24cd655f5baac2de315e4c87491860ef20a923a154f22b51ad9e0ca6825752" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.123466 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-n8pfm" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.134881 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.158280 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.183852 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.193187 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.198373 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.200304 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.201155 4766 scope.go:117] "RemoveContainer" containerID="e8fbaea76abae029d1a7e07d17d13ba1ef29a28831a8d0e6f6fc04e295ff065e" Nov 26 00:48:02 crc kubenswrapper[4766]: E1126 00:48:02.201486 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=octavia-operator-controller-manager-64cdc6ff96-bfw5f_openstack-operators(91d3e9c8-8bbc-43b9-947b-71c4bb474b21)\"" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" podUID="91d3e9c8-8bbc-43b9-947b-71c4bb474b21" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.248287 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.263689 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.300604 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.320751 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.339805 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.379244 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.428559 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.495431 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-tf6mh" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.505357 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.511148 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-dq898" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.514072 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.519337 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.525909 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.539344 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-config" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.540745 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-qwbfd" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.563067 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.604233 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-glkl6" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.622733 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.640526 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.652739 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.668482 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.717108 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.735729 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.800989 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.818598 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.826425 4766 scope.go:117] "RemoveContainer" containerID="c4b508111353f64ce1118a8ac912e3d512d2f23b317dd6a0483c4fe2e0126067" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.826730 4766 scope.go:117] "RemoveContainer" containerID="5f7d72a8709e834d6b16398a4a94f1cdb096b975fefcb4fa32830decc9e3a380" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.841223 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.864064 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.882922 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-pnfcn" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.933461 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-f6npz" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.945619 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.965984 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Nov 26 00:48:02 crc kubenswrapper[4766]: I1126 00:48:02.983161 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.018281 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.046907 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.050316 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.056264 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.100965 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.137348 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.146948 4766 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-hxpmg" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.157418 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-457zd" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.246205 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-grpc" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.277411 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-m2b9k" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.285386 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.289510 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.295279 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.390085 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"kube-root-ca.crt" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.391353 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.396141 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.405484 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.407392 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.416486 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-c8mws" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.417614 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.430014 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.443396 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.486632 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.493607 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.493840 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.503365 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.517373 4766 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-wt42m" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.553294 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.566128 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-kkdmz" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.678151 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-5zcn8" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.678169 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-rw628" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.703699 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-94zcs" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.733326 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.734867 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.745958 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.747734 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.771106 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.782293 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.798511 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.816194 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.827809 4766 scope.go:117] "RemoveContainer" containerID="69ed96c12e25467c32b6ad99ca52ad4f46ecc9f3be096414f5b261a00874e013" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.838740 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.839180 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.846147 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.847303 4766 scope.go:117] "RemoveContainer" containerID="f4f0be8205cb3c087b28e7451a802ae7e7912ffbab226d356ad6b10cb884ef49" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.870488 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.881752 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.916052 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.927431 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.939513 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.952359 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.964334 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.996459 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-grpc" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.998385 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 26 00:48:03 crc kubenswrapper[4766]: I1126 00:48:03.998514 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.006041 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.034329 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.066715 4766 generic.go:334] "Generic (PLEG): container finished" podID="1ecbd1d2-cb7f-432f-bc75-903d0643ea29" containerID="e483976d00f30841115f915004cb491c322867e1677e1242b465493dd4116df1" exitCode=1 Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.066841 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" event={"ID":"1ecbd1d2-cb7f-432f-bc75-903d0643ea29","Type":"ContainerDied","Data":"e483976d00f30841115f915004cb491c322867e1677e1242b465493dd4116df1"} Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.066892 4766 scope.go:117] "RemoveContainer" containerID="5f7d72a8709e834d6b16398a4a94f1cdb096b975fefcb4fa32830decc9e3a380" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.072219 4766 scope.go:117] "RemoveContainer" containerID="e483976d00f30841115f915004cb491c322867e1677e1242b465493dd4116df1" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.085203 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.087601 4766 generic.go:334] "Generic (PLEG): container finished" podID="5284ad70-a36c-4666-8202-84633d83f50b" containerID="077353144c2eb33b5eb96b9173cb8e009bf57db150620c2e8afe527ca503c06c" exitCode=1 Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.087711 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" event={"ID":"5284ad70-a36c-4666-8202-84633d83f50b","Type":"ContainerDied","Data":"077353144c2eb33b5eb96b9173cb8e009bf57db150620c2e8afe527ca503c06c"} Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.089049 4766 scope.go:117] "RemoveContainer" containerID="077353144c2eb33b5eb96b9173cb8e009bf57db150620c2e8afe527ca503c06c" Nov 26 00:48:04 crc kubenswrapper[4766]: E1126 00:48:04.096534 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=horizon-operator-controller-manager-5d494799bf-h4mrn_openstack-operators(5284ad70-a36c-4666-8202-84633d83f50b)\"" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" podUID="5284ad70-a36c-4666-8202-84633d83f50b" Nov 26 00:48:04 crc kubenswrapper[4766]: E1126 00:48:04.109568 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=heat-operator-controller-manager-5b77f656f-tkj9b_openstack-operators(1ecbd1d2-cb7f-432f-bc75-903d0643ea29)\"" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" podUID="1ecbd1d2-cb7f-432f-bc75-903d0643ea29" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.111334 4766 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.131949 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-b8c9d9569-lrlx6","openstack/heat-api-856c5c6f86-fk66f","openshift-kube-apiserver/kube-apiserver-crc","openstack/ceilometer-0"] Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.132508 4766 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.132535 4766 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e7f9d435-86f6-40ca-9e87-a1b4e44c18f3" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.135723 4766 scope.go:117] "RemoveContainer" containerID="0c594df0d9660266fff1417a3f4f615f9c3a9b37b90eeef4b4d18d0d955a98a2" Nov 26 00:48:04 crc kubenswrapper[4766]: E1126 00:48:04.136076 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(71bb4a3aecc4ba5b26c4b7318770ce13)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.149907 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.172768 4766 scope.go:117] "RemoveContainer" containerID="c4b508111353f64ce1118a8ac912e3d512d2f23b317dd6a0483c4fe2e0126067" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.219638 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.226950 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.239521 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ksmzm" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.240333 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.267912 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.278135 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.332877 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.339072 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-bfcr8" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.385534 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.414021 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.414774 4766 scope.go:117] "RemoveContainer" containerID="6396f3caca85c89b1192e1772820e1b1f5623132fa5baf5b852754febe7243de" Nov 26 00:48:04 crc kubenswrapper[4766]: E1126 00:48:04.415034 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=metallb-operator-controller-manager-764c8cf56f-vrwpt_metallb-system(fd3a6275-9f47-4799-bc15-96a51b93f31b)\"" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" podUID="fd3a6275-9f47-4799-bc15-96a51b93f31b" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.424845 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.496608 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-x9jkp" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.540149 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.559675 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.576119 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.673058 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.741043 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-gjrqk" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.769739 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-87dbv" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.774076 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-grpc" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.826685 4766 scope.go:117] "RemoveContainer" containerID="502141878d9557d404939737707070c203f202ba93a6d5aa04240a14a618e040" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.826813 4766 scope.go:117] "RemoveContainer" containerID="3a5bf6a88bd14fa3df408b57df7d3930b7c4f71f2d70788e373686f0aad92b30" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.826929 4766 scope.go:117] "RemoveContainer" containerID="5d7d2b957bbb4841fc8ab537bef654cacd24c482167164e785eed6faba12e21f" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.859590 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.904116 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.947018 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-6kd8s" Nov 26 00:48:04 crc kubenswrapper[4766]: I1126 00:48:04.953050 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-dockercfg-4qhbv" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.006045 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.032087 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.034595 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.043980 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.079557 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.101407 4766 generic.go:334] "Generic (PLEG): container finished" podID="8969652f-14c9-4618-ad2d-1b3fd1ad3627" containerID="40b84f98ff2e7a66714b3171a6fb601ec8a52d1fcc8b4f251028c0ec0684b6d2" exitCode=1 Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.101438 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" event={"ID":"8969652f-14c9-4618-ad2d-1b3fd1ad3627","Type":"ContainerDied","Data":"40b84f98ff2e7a66714b3171a6fb601ec8a52d1fcc8b4f251028c0ec0684b6d2"} Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.101553 4766 scope.go:117] "RemoveContainer" containerID="69ed96c12e25467c32b6ad99ca52ad4f46ecc9f3be096414f5b261a00874e013" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.102357 4766 scope.go:117] "RemoveContainer" containerID="40b84f98ff2e7a66714b3171a6fb601ec8a52d1fcc8b4f251028c0ec0684b6d2" Nov 26 00:48:05 crc kubenswrapper[4766]: E1126 00:48:05.102684 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=nova-operator-controller-manager-79556f57fc-dm7xb_openstack-operators(8969652f-14c9-4618-ad2d-1b3fd1ad3627)\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" podUID="8969652f-14c9-4618-ad2d-1b3fd1ad3627" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.141120 4766 generic.go:334] "Generic (PLEG): container finished" podID="c3cec209-fdef-44a6-90d0-2d06cfd7e5cc" containerID="9ed08fb921ec77ec6e1e8ed96b51841858e7437516717a99a7033a0ee465972b" exitCode=1 Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.141218 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" event={"ID":"c3cec209-fdef-44a6-90d0-2d06cfd7e5cc","Type":"ContainerDied","Data":"9ed08fb921ec77ec6e1e8ed96b51841858e7437516717a99a7033a0ee465972b"} Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.141866 4766 scope.go:117] "RemoveContainer" containerID="9ed08fb921ec77ec6e1e8ed96b51841858e7437516717a99a7033a0ee465972b" Nov 26 00:48:05 crc kubenswrapper[4766]: E1126 00:48:05.142226 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=infra-operator-controller-manager-57548d458d-x6h4v_openstack-operators(c3cec209-fdef-44a6-90d0-2d06cfd7e5cc)\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" podUID="c3cec209-fdef-44a6-90d0-2d06cfd7e5cc" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.153761 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-cwfw4" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.178999 4766 scope.go:117] "RemoveContainer" containerID="f4f0be8205cb3c087b28e7451a802ae7e7912ffbab226d356ad6b10cb884ef49" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.207237 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-8w9h8" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.221476 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.240525 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.263237 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.299178 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.323899 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.344072 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-pc94r" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.370221 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.375568 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.381358 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.415790 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.417763 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.430185 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.445197 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.453304 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.491947 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.539992 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.618200 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.628236 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.645699 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-llt2n" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.645936 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-kdksf" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.669509 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.694513 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.702031 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-grpc" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.726629 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-fwb2c" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.779141 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-rd2kk" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.791566 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.804398 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.826890 4766 scope.go:117] "RemoveContainer" containerID="9e0ffa5e614b56b5972b597f5d7cf1d69c9755a6edfc1db6ac05034cb4e3d62d" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.827093 4766 scope.go:117] "RemoveContainer" containerID="089ff57a19d572ad607d9b8aefbeca178c8046089d9d0f1a2b360a9cd51d8e1b" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.827685 4766 scope.go:117] "RemoveContainer" containerID="b6766bbab2729ba50956a4e98254bd688babe7e7ba968cdd1ea57ef3ea991f0e" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.840586 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" path="/var/lib/kubelet/pods/1424c325-de53-4623-8d9c-7b1ef990c76c/volumes" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.841481 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16337a9e-8651-492e-9f39-21515275e3cb" path="/var/lib/kubelet/pods/16337a9e-8651-492e-9f39-21515275e3cb/volumes" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.842089 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4442abe-9e23-4f4e-880c-89ca9d8ef785" path="/var/lib/kubelet/pods/d4442abe-9e23-4f4e-880c-89ca9d8ef785/volumes" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.859012 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.884693 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 26 00:48:05 crc kubenswrapper[4766]: I1126 00:48:05.938669 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.046680 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.079691 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.084718 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.126736 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.151318 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.160043 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.161867 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.163894 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" event={"ID":"cc87f631-ba4e-4943-b51d-bf1012cca754","Type":"ContainerStarted","Data":"cea7bdd56843f2b2b035ccfb43d7a95792ec34d2199a87ccd3f29a3a8286b383"} Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.164165 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.167892 4766 generic.go:334] "Generic (PLEG): container finished" podID="2bb23a3b-1f4a-409f-9d9c-0afe73a33352" containerID="e2387b88677d93c567a1d130a1054097ae8e786e58aed741f4a8aa02b969a6b3" exitCode=1 Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.167950 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" event={"ID":"2bb23a3b-1f4a-409f-9d9c-0afe73a33352","Type":"ContainerDied","Data":"e2387b88677d93c567a1d130a1054097ae8e786e58aed741f4a8aa02b969a6b3"} Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.167983 4766 scope.go:117] "RemoveContainer" containerID="5d7d2b957bbb4841fc8ab537bef654cacd24c482167164e785eed6faba12e21f" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.168507 4766 scope.go:117] "RemoveContainer" containerID="e2387b88677d93c567a1d130a1054097ae8e786e58aed741f4a8aa02b969a6b3" Nov 26 00:48:06 crc kubenswrapper[4766]: E1126 00:48:06.168764 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=neutron-operator-controller-manager-6fdcddb789-wdzgw_openstack-operators(2bb23a3b-1f4a-409f-9d9c-0afe73a33352)\"" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" podUID="2bb23a3b-1f4a-409f-9d9c-0afe73a33352" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.171633 4766 generic.go:334] "Generic (PLEG): container finished" podID="749a25e6-23eb-4fdd-8e15-ac2874cdee7f" containerID="2616d6e260dc8764705bb96f63f0717a0cfc0738acfaa392b4301101a7942fe2" exitCode=1 Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.171739 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" event={"ID":"749a25e6-23eb-4fdd-8e15-ac2874cdee7f","Type":"ContainerDied","Data":"2616d6e260dc8764705bb96f63f0717a0cfc0738acfaa392b4301101a7942fe2"} Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.174831 4766 scope.go:117] "RemoveContainer" containerID="2616d6e260dc8764705bb96f63f0717a0cfc0738acfaa392b4301101a7942fe2" Nov 26 00:48:06 crc kubenswrapper[4766]: E1126 00:48:06.175639 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=mariadb-operator-controller-manager-66f4dd4bc7-z865g_openstack-operators(749a25e6-23eb-4fdd-8e15-ac2874cdee7f)\"" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" podUID="749a25e6-23eb-4fdd-8e15-ac2874cdee7f" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.176032 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" event={"ID":"04e53d09-de84-4682-b17b-61f9ea7bf152","Type":"ContainerStarted","Data":"b1b0c5aa8f40d0f2c07265a853eba57d2333036b22f55004315b3f36d17ce754"} Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.176994 4766 scope.go:117] "RemoveContainer" containerID="b1b0c5aa8f40d0f2c07265a853eba57d2333036b22f55004315b3f36d17ce754" Nov 26 00:48:06 crc kubenswrapper[4766]: E1126 00:48:06.177350 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=telemetry-operator-controller-manager-5ccf99cfdd-9cfft_openstack-operators(04e53d09-de84-4682-b17b-61f9ea7bf152)\"" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" podUID="04e53d09-de84-4682-b17b-61f9ea7bf152" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.190450 4766 generic.go:334] "Generic (PLEG): container finished" podID="a2e7c098-bf9d-4d99-84bf-adfb49ee6b12" containerID="31e4e2562c7a94a89c0e4b9d27edc8f0f6845cbd15437a89c00febff1c4aa709" exitCode=1 Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.190543 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" event={"ID":"a2e7c098-bf9d-4d99-84bf-adfb49ee6b12","Type":"ContainerDied","Data":"31e4e2562c7a94a89c0e4b9d27edc8f0f6845cbd15437a89c00febff1c4aa709"} Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.191608 4766 scope.go:117] "RemoveContainer" containerID="31e4e2562c7a94a89c0e4b9d27edc8f0f6845cbd15437a89c00febff1c4aa709" Nov 26 00:48:06 crc kubenswrapper[4766]: E1126 00:48:06.192029 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=ovn-operator-controller-manager-56897c768d-9zhql_openstack-operators(a2e7c098-bf9d-4d99-84bf-adfb49ee6b12)\"" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" podUID="a2e7c098-bf9d-4d99-84bf-adfb49ee6b12" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.197611 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw" event={"ID":"91ee52b5-0e2e-4732-87bf-809e378d33d3","Type":"ContainerStarted","Data":"0566d4933fc09678c5b262e6d0a2f71deb7b2668a246a353a87132378e3d1324"} Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.199357 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-8fcdz" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.209937 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.250046 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.252985 4766 scope.go:117] "RemoveContainer" containerID="502141878d9557d404939737707070c203f202ba93a6d5aa04240a14a618e040" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.273566 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.286570 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-58h9w" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.302953 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.313504 4766 scope.go:117] "RemoveContainer" containerID="3a5bf6a88bd14fa3df408b57df7d3930b7c4f71f2d70788e373686f0aad92b30" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.336576 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.340979 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.378809 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.513099 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.551064 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.603509 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.629311 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.674753 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.679766 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.703613 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.717291 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.781448 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.781875 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.783891 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.792872 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.810706 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.825025 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-http" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.827370 4766 scope.go:117] "RemoveContainer" containerID="0c41858ad420d326d413541187c4331ae0d60df9b0ce9ce39bd4e1c69d9e0536" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.860009 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.895573 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.907215 4766 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.911119 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.929884 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.929901 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:48:06 crc kubenswrapper[4766]: E1126 00:48:06.930639 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" containerName="ceilometer-central-agent" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.930676 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" containerName="ceilometer-central-agent" Nov 26 00:48:06 crc kubenswrapper[4766]: E1126 00:48:06.930691 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" containerName="sg-core" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.930698 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" containerName="sg-core" Nov 26 00:48:06 crc kubenswrapper[4766]: E1126 00:48:06.930714 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" containerName="proxy-httpd" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.930720 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" containerName="proxy-httpd" Nov 26 00:48:06 crc kubenswrapper[4766]: E1126 00:48:06.930757 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" containerName="ceilometer-notification-agent" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.930763 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" containerName="ceilometer-notification-agent" Nov 26 00:48:06 crc kubenswrapper[4766]: E1126 00:48:06.930778 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4442abe-9e23-4f4e-880c-89ca9d8ef785" containerName="heat-cfnapi" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.930784 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4442abe-9e23-4f4e-880c-89ca9d8ef785" containerName="heat-cfnapi" Nov 26 00:48:06 crc kubenswrapper[4766]: E1126 00:48:06.930803 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16337a9e-8651-492e-9f39-21515275e3cb" containerName="heat-api" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.930809 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="16337a9e-8651-492e-9f39-21515275e3cb" containerName="heat-api" Nov 26 00:48:06 crc kubenswrapper[4766]: E1126 00:48:06.930819 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bba0b1b-4848-4452-a535-c0cf5d13b30b" containerName="installer" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.930825 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bba0b1b-4848-4452-a535-c0cf5d13b30b" containerName="installer" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.931007 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" containerName="ceilometer-notification-agent" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.931019 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" containerName="ceilometer-central-agent" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.931024 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.931035 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" containerName="proxy-httpd" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.931054 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="1424c325-de53-4623-8d9c-7b1ef990c76c" containerName="sg-core" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.931060 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4442abe-9e23-4f4e-880c-89ca9d8ef785" containerName="heat-cfnapi" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.931075 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="16337a9e-8651-492e-9f39-21515275e3cb" containerName="heat-api" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.931087 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bba0b1b-4848-4452-a535-c0cf5d13b30b" containerName="installer" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.932918 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.934838 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.935012 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 26 00:48:06 crc kubenswrapper[4766]: I1126 00:48:06.965292 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.022286 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"cluster-logging-operator-dockercfg-5hkff" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.023405 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.064808 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdd2abaa-1d30-4422-80c9-496e89a9aad9-run-httpd\") pod \"ceilometer-0\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.064979 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdd2abaa-1d30-4422-80c9-496e89a9aad9-log-httpd\") pod \"ceilometer-0\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.065227 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgxgw\" (UniqueName: \"kubernetes.io/projected/bdd2abaa-1d30-4422-80c9-496e89a9aad9-kube-api-access-jgxgw\") pod \"ceilometer-0\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.065281 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.065498 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.065544 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-scripts\") pod \"ceilometer-0\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.065803 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-config-data\") pod \"ceilometer-0\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.078340 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.113095 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-d7c28" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.136720 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.167011 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.167991 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgxgw\" (UniqueName: \"kubernetes.io/projected/bdd2abaa-1d30-4422-80c9-496e89a9aad9-kube-api-access-jgxgw\") pod \"ceilometer-0\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.168054 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.168128 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.168161 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-scripts\") pod \"ceilometer-0\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.168263 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-config-data\") pod \"ceilometer-0\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.168335 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdd2abaa-1d30-4422-80c9-496e89a9aad9-run-httpd\") pod \"ceilometer-0\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.168368 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdd2abaa-1d30-4422-80c9-496e89a9aad9-log-httpd\") pod \"ceilometer-0\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.168877 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdd2abaa-1d30-4422-80c9-496e89a9aad9-run-httpd\") pod \"ceilometer-0\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.168894 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdd2abaa-1d30-4422-80c9-496e89a9aad9-log-httpd\") pod \"ceilometer-0\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.172765 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.174294 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.174577 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-scripts\") pod \"ceilometer-0\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.175868 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-config-data\") pod \"ceilometer-0\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.175902 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.184684 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.185863 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgxgw\" (UniqueName: \"kubernetes.io/projected/bdd2abaa-1d30-4422-80c9-496e89a9aad9-kube-api-access-jgxgw\") pod \"ceilometer-0\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.197223 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.199088 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.211857 4766 generic.go:334] "Generic (PLEG): container finished" podID="cc87f631-ba4e-4943-b51d-bf1012cca754" containerID="cea7bdd56843f2b2b035ccfb43d7a95792ec34d2199a87ccd3f29a3a8286b383" exitCode=1 Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.211954 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" event={"ID":"cc87f631-ba4e-4943-b51d-bf1012cca754","Type":"ContainerDied","Data":"cea7bdd56843f2b2b035ccfb43d7a95792ec34d2199a87ccd3f29a3a8286b383"} Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.212008 4766 scope.go:117] "RemoveContainer" containerID="b6766bbab2729ba50956a4e98254bd688babe7e7ba968cdd1ea57ef3ea991f0e" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.212868 4766 scope.go:117] "RemoveContainer" containerID="cea7bdd56843f2b2b035ccfb43d7a95792ec34d2199a87ccd3f29a3a8286b383" Nov 26 00:48:07 crc kubenswrapper[4766]: E1126 00:48:07.213194 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=barbican-operator-controller-manager-7b64f4fb85-gbdhz_openstack-operators(cc87f631-ba4e-4943-b51d-bf1012cca754)\"" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" podUID="cc87f631-ba4e-4943-b51d-bf1012cca754" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.214311 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-vzktk" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.221351 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.234544 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" event={"ID":"00a32e6c-edcc-4625-ac35-26fed19f1638","Type":"ContainerStarted","Data":"e334b081ac2683145ab1dd86d34adf85f1f0aabd3f025419401ad076704959de"} Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.235348 4766 scope.go:117] "RemoveContainer" containerID="e334b081ac2683145ab1dd86d34adf85f1f0aabd3f025419401ad076704959de" Nov 26 00:48:07 crc kubenswrapper[4766]: E1126 00:48:07.235740 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-d77b94747-2t9w2_openstack-operators(00a32e6c-edcc-4625-ac35-26fed19f1638)\"" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" podUID="00a32e6c-edcc-4625-ac35-26fed19f1638" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.244311 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-5mqxk" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.248597 4766 generic.go:334] "Generic (PLEG): container finished" podID="04e53d09-de84-4682-b17b-61f9ea7bf152" containerID="b1b0c5aa8f40d0f2c07265a853eba57d2333036b22f55004315b3f36d17ce754" exitCode=1 Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.248730 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" event={"ID":"04e53d09-de84-4682-b17b-61f9ea7bf152","Type":"ContainerDied","Data":"b1b0c5aa8f40d0f2c07265a853eba57d2333036b22f55004315b3f36d17ce754"} Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.252377 4766 scope.go:117] "RemoveContainer" containerID="b1b0c5aa8f40d0f2c07265a853eba57d2333036b22f55004315b3f36d17ce754" Nov 26 00:48:07 crc kubenswrapper[4766]: E1126 00:48:07.252627 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=telemetry-operator-controller-manager-5ccf99cfdd-9cfft_openstack-operators(04e53d09-de84-4682-b17b-61f9ea7bf152)\"" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" podUID="04e53d09-de84-4682-b17b-61f9ea7bf152" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.262240 4766 generic.go:334] "Generic (PLEG): container finished" podID="91ee52b5-0e2e-4732-87bf-809e378d33d3" containerID="0566d4933fc09678c5b262e6d0a2f71deb7b2668a246a353a87132378e3d1324" exitCode=1 Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.262292 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw" event={"ID":"91ee52b5-0e2e-4732-87bf-809e378d33d3","Type":"ContainerDied","Data":"0566d4933fc09678c5b262e6d0a2f71deb7b2668a246a353a87132378e3d1324"} Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.263028 4766 scope.go:117] "RemoveContainer" containerID="0566d4933fc09678c5b262e6d0a2f71deb7b2668a246a353a87132378e3d1324" Nov 26 00:48:07 crc kubenswrapper[4766]: E1126 00:48:07.263329 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=operator pod=rabbitmq-cluster-operator-manager-668c99d594-mx9vw_openstack-operators(91ee52b5-0e2e-4732-87bf-809e378d33d3)\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw" podUID="91ee52b5-0e2e-4732-87bf-809e378d33d3" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.275838 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-4w6wk" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.281775 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.293114 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.310111 4766 scope.go:117] "RemoveContainer" containerID="089ff57a19d572ad607d9b8aefbeca178c8046089d9d0f1a2b360a9cd51d8e1b" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.345556 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.405459 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.455506 4766 scope.go:117] "RemoveContainer" containerID="9e0ffa5e614b56b5972b597f5d7cf1d69c9755a6edfc1db6ac05034cb4e3d62d" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.458954 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-http" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.511909 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.517106 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.518581 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-kbwvs" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.524189 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.535710 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.539234 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 26 00:48:07 crc kubenswrapper[4766]: E1126 00:48:07.588732 4766 log.go:32] "RunPodSandbox from runtime service failed" err=< Nov 26 00:48:07 crc kubenswrapper[4766]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_ceilometer-0_openstack_bdd2abaa-1d30-4422-80c9-496e89a9aad9_0(66e01506dcaed091a9d489f436062f593ac3f424c51425b9333b56dc2e20b9ef): error adding pod openstack_ceilometer-0 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"66e01506dcaed091a9d489f436062f593ac3f424c51425b9333b56dc2e20b9ef" Netns:"/var/run/netns/f8e6f1be-b3f2-4c31-8e63-2145256e48cb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=ceilometer-0;K8S_POD_INFRA_CONTAINER_ID=66e01506dcaed091a9d489f436062f593ac3f424c51425b9333b56dc2e20b9ef;K8S_POD_UID=bdd2abaa-1d30-4422-80c9-496e89a9aad9" Path:"" ERRORED: error configuring pod [openstack/ceilometer-0] networking: [openstack/ceilometer-0/bdd2abaa-1d30-4422-80c9-496e89a9aad9:ovn-kubernetes]: error adding container to network "ovn-kubernetes": CNI request failed with status 400: '[openstack/ceilometer-0 66e01506dcaed091a9d489f436062f593ac3f424c51425b9333b56dc2e20b9ef network default NAD default] [openstack/ceilometer-0 66e01506dcaed091a9d489f436062f593ac3f424c51425b9333b56dc2e20b9ef network default NAD default] pod deleted before sandbox ADD operation began Nov 26 00:48:07 crc kubenswrapper[4766]: ' Nov 26 00:48:07 crc kubenswrapper[4766]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 26 00:48:07 crc kubenswrapper[4766]: > Nov 26 00:48:07 crc kubenswrapper[4766]: E1126 00:48:07.588797 4766 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Nov 26 00:48:07 crc kubenswrapper[4766]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_ceilometer-0_openstack_bdd2abaa-1d30-4422-80c9-496e89a9aad9_0(66e01506dcaed091a9d489f436062f593ac3f424c51425b9333b56dc2e20b9ef): error adding pod openstack_ceilometer-0 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"66e01506dcaed091a9d489f436062f593ac3f424c51425b9333b56dc2e20b9ef" Netns:"/var/run/netns/f8e6f1be-b3f2-4c31-8e63-2145256e48cb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=ceilometer-0;K8S_POD_INFRA_CONTAINER_ID=66e01506dcaed091a9d489f436062f593ac3f424c51425b9333b56dc2e20b9ef;K8S_POD_UID=bdd2abaa-1d30-4422-80c9-496e89a9aad9" Path:"" ERRORED: error configuring pod [openstack/ceilometer-0] networking: [openstack/ceilometer-0/bdd2abaa-1d30-4422-80c9-496e89a9aad9:ovn-kubernetes]: error adding container to network "ovn-kubernetes": CNI request failed with status 400: '[openstack/ceilometer-0 66e01506dcaed091a9d489f436062f593ac3f424c51425b9333b56dc2e20b9ef network default NAD default] [openstack/ceilometer-0 66e01506dcaed091a9d489f436062f593ac3f424c51425b9333b56dc2e20b9ef network default NAD default] pod deleted before sandbox ADD operation began Nov 26 00:48:07 crc kubenswrapper[4766]: ' Nov 26 00:48:07 crc kubenswrapper[4766]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 26 00:48:07 crc kubenswrapper[4766]: > pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: E1126 00:48:07.588816 4766 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Nov 26 00:48:07 crc kubenswrapper[4766]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_ceilometer-0_openstack_bdd2abaa-1d30-4422-80c9-496e89a9aad9_0(66e01506dcaed091a9d489f436062f593ac3f424c51425b9333b56dc2e20b9ef): error adding pod openstack_ceilometer-0 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"66e01506dcaed091a9d489f436062f593ac3f424c51425b9333b56dc2e20b9ef" Netns:"/var/run/netns/f8e6f1be-b3f2-4c31-8e63-2145256e48cb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=ceilometer-0;K8S_POD_INFRA_CONTAINER_ID=66e01506dcaed091a9d489f436062f593ac3f424c51425b9333b56dc2e20b9ef;K8S_POD_UID=bdd2abaa-1d30-4422-80c9-496e89a9aad9" Path:"" ERRORED: error configuring pod [openstack/ceilometer-0] networking: [openstack/ceilometer-0/bdd2abaa-1d30-4422-80c9-496e89a9aad9:ovn-kubernetes]: error adding container to network "ovn-kubernetes": CNI request failed with status 400: '[openstack/ceilometer-0 66e01506dcaed091a9d489f436062f593ac3f424c51425b9333b56dc2e20b9ef network default NAD default] [openstack/ceilometer-0 66e01506dcaed091a9d489f436062f593ac3f424c51425b9333b56dc2e20b9ef network default NAD default] pod deleted before sandbox ADD operation began Nov 26 00:48:07 crc kubenswrapper[4766]: ' Nov 26 00:48:07 crc kubenswrapper[4766]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 26 00:48:07 crc kubenswrapper[4766]: > pod="openstack/ceilometer-0" Nov 26 00:48:07 crc kubenswrapper[4766]: E1126 00:48:07.588871 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"ceilometer-0_openstack(bdd2abaa-1d30-4422-80c9-496e89a9aad9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"ceilometer-0_openstack(bdd2abaa-1d30-4422-80c9-496e89a9aad9)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_ceilometer-0_openstack_bdd2abaa-1d30-4422-80c9-496e89a9aad9_0(66e01506dcaed091a9d489f436062f593ac3f424c51425b9333b56dc2e20b9ef): error adding pod openstack_ceilometer-0 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"66e01506dcaed091a9d489f436062f593ac3f424c51425b9333b56dc2e20b9ef\\\" Netns:\\\"/var/run/netns/f8e6f1be-b3f2-4c31-8e63-2145256e48cb\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=ceilometer-0;K8S_POD_INFRA_CONTAINER_ID=66e01506dcaed091a9d489f436062f593ac3f424c51425b9333b56dc2e20b9ef;K8S_POD_UID=bdd2abaa-1d30-4422-80c9-496e89a9aad9\\\" Path:\\\"\\\" ERRORED: error configuring pod [openstack/ceilometer-0] networking: [openstack/ceilometer-0/bdd2abaa-1d30-4422-80c9-496e89a9aad9:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": CNI request failed with status 400: '[openstack/ceilometer-0 66e01506dcaed091a9d489f436062f593ac3f424c51425b9333b56dc2e20b9ef network default NAD default] [openstack/ceilometer-0 66e01506dcaed091a9d489f436062f593ac3f424c51425b9333b56dc2e20b9ef network default NAD default] pod deleted before sandbox ADD operation began\\n'\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openstack/ceilometer-0" podUID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.601397 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.634480 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.689920 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.695438 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.695454 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-6wqr6" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.701080 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-whslv" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.730480 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-qv6sf" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.749270 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.753433 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.757692 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.765930 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.800350 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.822069 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.827815 4766 scope.go:117] "RemoveContainer" containerID="37c98bbe0d1a1d7dec9f2442d2c16ebe49c7485725817b532cbc38bae4365252" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.827887 4766 scope.go:117] "RemoveContainer" containerID="ccb6a5c90a0ad287ea4fde750c373ddb97fe844e0b2024e905b73d849ad7c7d6" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.844018 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.849434 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.849564 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.859466 4766 scope.go:117] "RemoveContainer" containerID="0c594df0d9660266fff1417a3f4f615f9c3a9b37b90eeef4b4d18d0d955a98a2" Nov 26 00:48:07 crc kubenswrapper[4766]: E1126 00:48:07.859709 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(71bb4a3aecc4ba5b26c4b7318770ce13)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.869778 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.891913 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.902262 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.903616 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.907206 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.973457 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-7b2rn" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.976691 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 26 00:48:07 crc kubenswrapper[4766]: I1126 00:48:07.982157 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.018970 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.031500 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.047171 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-gbglb" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.062929 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.064471 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.168573 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.242506 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.243526 4766 scope.go:117] "RemoveContainer" containerID="05fe629429cc17682b24434acacc6d08fc24ebc0d20e7008292b71d64a9c5529" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.275093 4766 generic.go:334] "Generic (PLEG): container finished" podID="751b8e60-46d4-4f5e-b62d-800f641ccd40" containerID="8137bb426d3c7cbb0adfb5a999511839a3f9b23baec1066df5c7c6b2c64af5e4" exitCode=1 Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.275219 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" event={"ID":"751b8e60-46d4-4f5e-b62d-800f641ccd40","Type":"ContainerDied","Data":"8137bb426d3c7cbb0adfb5a999511839a3f9b23baec1066df5c7c6b2c64af5e4"} Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.275251 4766 scope.go:117] "RemoveContainer" containerID="37c98bbe0d1a1d7dec9f2442d2c16ebe49c7485725817b532cbc38bae4365252" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.276175 4766 scope.go:117] "RemoveContainer" containerID="8137bb426d3c7cbb0adfb5a999511839a3f9b23baec1066df5c7c6b2c64af5e4" Nov 26 00:48:08 crc kubenswrapper[4766]: E1126 00:48:08.276582 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=watcher-operator-controller-manager-656dcb59d4-2hpg8_openstack-operators(751b8e60-46d4-4f5e-b62d-800f641ccd40)\"" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" podUID="751b8e60-46d4-4f5e-b62d-800f641ccd40" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.288928 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-dockercfg-mw87j" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.299258 4766 generic.go:334] "Generic (PLEG): container finished" podID="3d0d59b2-0486-40a2-ae60-b738a4f76dc4" containerID="786387060a4e3cb589ce85f666fff7f97910552f1b888209e9019cdeebedf9bd" exitCode=1 Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.299533 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" event={"ID":"3d0d59b2-0486-40a2-ae60-b738a4f76dc4","Type":"ContainerDied","Data":"786387060a4e3cb589ce85f666fff7f97910552f1b888209e9019cdeebedf9bd"} Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.300292 4766 scope.go:117] "RemoveContainer" containerID="786387060a4e3cb589ce85f666fff7f97910552f1b888209e9019cdeebedf9bd" Nov 26 00:48:08 crc kubenswrapper[4766]: E1126 00:48:08.300835 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-7b4567c7cf-fcr76_openstack-operators(3d0d59b2-0486-40a2-ae60-b738a4f76dc4)\"" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" podUID="3d0d59b2-0486-40a2-ae60-b738a4f76dc4" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.305207 4766 scope.go:117] "RemoveContainer" containerID="cea7bdd56843f2b2b035ccfb43d7a95792ec34d2199a87ccd3f29a3a8286b383" Nov 26 00:48:08 crc kubenswrapper[4766]: E1126 00:48:08.305606 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=barbican-operator-controller-manager-7b64f4fb85-gbdhz_openstack-operators(cc87f631-ba4e-4943-b51d-bf1012cca754)\"" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" podUID="cc87f631-ba4e-4943-b51d-bf1012cca754" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.309619 4766 generic.go:334] "Generic (PLEG): container finished" podID="00a32e6c-edcc-4625-ac35-26fed19f1638" containerID="e334b081ac2683145ab1dd86d34adf85f1f0aabd3f025419401ad076704959de" exitCode=1 Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.309776 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" event={"ID":"00a32e6c-edcc-4625-ac35-26fed19f1638","Type":"ContainerDied","Data":"e334b081ac2683145ab1dd86d34adf85f1f0aabd3f025419401ad076704959de"} Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.310236 4766 scope.go:117] "RemoveContainer" containerID="e334b081ac2683145ab1dd86d34adf85f1f0aabd3f025419401ad076704959de" Nov 26 00:48:08 crc kubenswrapper[4766]: E1126 00:48:08.311302 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-d77b94747-2t9w2_openstack-operators(00a32e6c-edcc-4625-ac35-26fed19f1638)\"" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" podUID="00a32e6c-edcc-4625-ac35-26fed19f1638" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.323531 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Nov 26 00:48:08 crc kubenswrapper[4766]: E1126 00:48:08.327759 4766 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.328510 4766 scope.go:117] "RemoveContainer" containerID="0c594df0d9660266fff1417a3f4f615f9c3a9b37b90eeef4b4d18d0d955a98a2" Nov 26 00:48:08 crc kubenswrapper[4766]: E1126 00:48:08.329133 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(71bb4a3aecc4ba5b26c4b7318770ce13)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.354421 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.357164 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.384361 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.403863 4766 scope.go:117] "RemoveContainer" containerID="ccb6a5c90a0ad287ea4fde750c373ddb97fe844e0b2024e905b73d849ad7c7d6" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.425836 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.456298 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.475457 4766 scope.go:117] "RemoveContainer" containerID="0c41858ad420d326d413541187c4331ae0d60df9b0ce9ce39bd4e1c69d9e0536" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.479317 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.502148 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.557351 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.572228 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.572686 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.702678 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.731099 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.751485 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-http" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.756080 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.756409 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.762265 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.804801 4766 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.827320 4766 scope.go:117] "RemoveContainer" containerID="624e9ac049eb714839dab1ba6bf0a2082221a591770ac0e439ad6f283f024f43" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.827713 4766 scope.go:117] "RemoveContainer" containerID="8ce23e89748eaf63d2ede32d0d0f586ce2050c357cf7ddf77eebed117cb8a3da" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.832072 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.845485 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.847763 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.847923 4766 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.851918 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-http" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.892352 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.901297 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.959567 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 26 00:48:08 crc kubenswrapper[4766]: I1126 00:48:08.999769 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-http" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.015719 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-ca-bundle" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.039488 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.043416 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.046529 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.074055 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.084573 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.135198 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-grpc" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.144730 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.164797 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.170342 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.175988 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.181370 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.185587 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.216052 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.233368 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.233574 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.257544 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.263573 4766 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.330015 4766 generic.go:334] "Generic (PLEG): container finished" podID="d00b9c60-0fb7-49c6-944a-fac598cc22a4" containerID="04f968501bccdff8ecc4863f1826684c21db8e3e31f21130322a744ed9db3121" exitCode=1 Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.330070 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" event={"ID":"d00b9c60-0fb7-49c6-944a-fac598cc22a4","Type":"ContainerDied","Data":"04f968501bccdff8ecc4863f1826684c21db8e3e31f21130322a744ed9db3121"} Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.330106 4766 scope.go:117] "RemoveContainer" containerID="05fe629429cc17682b24434acacc6d08fc24ebc0d20e7008292b71d64a9c5529" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.330435 4766 scope.go:117] "RemoveContainer" containerID="04f968501bccdff8ecc4863f1826684c21db8e3e31f21130322a744ed9db3121" Nov 26 00:48:09 crc kubenswrapper[4766]: E1126 00:48:09.330740 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=openstack-operator-controller-manager-86674bdb4b-vq98m_openstack-operators(d00b9c60-0fb7-49c6-944a-fac598cc22a4)\"" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" podUID="d00b9c60-0fb7-49c6-944a-fac598cc22a4" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.340932 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" event={"ID":"97419f32-8768-495f-a652-8a1e8701546f","Type":"ContainerStarted","Data":"d692e1c399f03132124c0dd23f70b7781241d904e3b432175f22c1e5fdbfabc1"} Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.341213 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.347866 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn" event={"ID":"a16c5563-750f-4e91-93bd-4044c656cb88","Type":"ContainerStarted","Data":"14ad7c1b71715230a35b8a507d81bbf7a6efa854bc168304a0ff90658cc12869"} Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.348863 4766 scope.go:117] "RemoveContainer" containerID="0c594df0d9660266fff1417a3f4f615f9c3a9b37b90eeef4b4d18d0d955a98a2" Nov 26 00:48:09 crc kubenswrapper[4766]: E1126 00:48:09.349142 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(71bb4a3aecc4ba5b26c4b7318770ce13)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.374758 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.397403 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-tbnmt" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.417621 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-dhtlj" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.418331 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.555852 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.556193 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-jgphj" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.559441 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.562310 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.593178 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-gdhck" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.601273 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.639156 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.660976 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-6xjwq" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.765487 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.808979 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.817705 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.833736 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.888280 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.920969 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.939987 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 26 00:48:09 crc kubenswrapper[4766]: I1126 00:48:09.959410 4766 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 26 00:48:10 crc kubenswrapper[4766]: I1126 00:48:10.036484 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 26 00:48:10 crc kubenswrapper[4766]: I1126 00:48:10.060499 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 26 00:48:10 crc kubenswrapper[4766]: I1126 00:48:10.178425 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 26 00:48:10 crc kubenswrapper[4766]: I1126 00:48:10.254511 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 26 00:48:10 crc kubenswrapper[4766]: I1126 00:48:10.266927 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 26 00:48:10 crc kubenswrapper[4766]: I1126 00:48:10.313998 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 26 00:48:10 crc kubenswrapper[4766]: I1126 00:48:10.320047 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 26 00:48:10 crc kubenswrapper[4766]: I1126 00:48:10.332965 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 26 00:48:10 crc kubenswrapper[4766]: I1126 00:48:10.428830 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-pbtwq" Nov 26 00:48:10 crc kubenswrapper[4766]: I1126 00:48:10.446977 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 26 00:48:10 crc kubenswrapper[4766]: I1126 00:48:10.510728 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards" Nov 26 00:48:10 crc kubenswrapper[4766]: I1126 00:48:10.532508 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 26 00:48:10 crc kubenswrapper[4766]: I1126 00:48:10.542850 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 26 00:48:10 crc kubenswrapper[4766]: I1126 00:48:10.643606 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 26 00:48:10 crc kubenswrapper[4766]: I1126 00:48:10.729754 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Nov 26 00:48:10 crc kubenswrapper[4766]: I1126 00:48:10.770151 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Nov 26 00:48:10 crc kubenswrapper[4766]: I1126 00:48:10.797498 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-lxqm8" Nov 26 00:48:10 crc kubenswrapper[4766]: I1126 00:48:10.870707 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 26 00:48:10 crc kubenswrapper[4766]: I1126 00:48:10.908628 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 26 00:48:10 crc kubenswrapper[4766]: I1126 00:48:10.917566 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 26 00:48:10 crc kubenswrapper[4766]: I1126 00:48:10.946243 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.166959 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.238165 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.274255 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.321673 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.374883 4766 generic.go:334] "Generic (PLEG): container finished" podID="954f9d95-f1ab-4e36-9b7a-d16fbe282e53" containerID="cace71eb37005730736ea0818605949369df8b7df51bfed5483cf0e70336e9ce" exitCode=1 Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.374926 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-nd9m2" event={"ID":"954f9d95-f1ab-4e36-9b7a-d16fbe282e53","Type":"ContainerDied","Data":"cace71eb37005730736ea0818605949369df8b7df51bfed5483cf0e70336e9ce"} Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.375538 4766 scope.go:117] "RemoveContainer" containerID="cace71eb37005730736ea0818605949369df8b7df51bfed5483cf0e70336e9ce" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.469164 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.470495 4766 scope.go:117] "RemoveContainer" containerID="00bc96df29ffd032af13f751f10f9ebd94973bdaab39c23fa02a669782573988" Nov 26 00:48:11 crc kubenswrapper[4766]: E1126 00:48:11.470839 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=cinder-operator-controller-manager-6b7f75547b-jgz6c_openstack-operators(31b1b73b-1d10-49ac-8549-b202c965b63e)\"" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" podUID="31b1b73b-1d10-49ac-8549-b202c965b63e" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.470932 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.476634 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.479957 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.480003 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.504491 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.505479 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.507610 4766 scope.go:117] "RemoveContainer" containerID="11900098c4ce70adf98b70c668f6fd3114393fbfb5ebae1324aa13e7994d8001" Nov 26 00:48:11 crc kubenswrapper[4766]: E1126 00:48:11.508010 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=designate-operator-controller-manager-955677c94-dqtcj_openstack-operators(0e6cfef3-86af-4cf0-ab06-b3eef687836b)\"" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" podUID="0e6cfef3-86af-4cf0-ab06-b3eef687836b" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.574731 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.613543 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.614356 4766 scope.go:117] "RemoveContainer" containerID="05a3f0845ceb0e6ddd27c7537b01703da19f1edc1685d2a382208a93ba7d72a2" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.673402 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.688926 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.689868 4766 scope.go:117] "RemoveContainer" containerID="077353144c2eb33b5eb96b9173cb8e009bf57db150620c2e8afe527ca503c06c" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.690214 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" Nov 26 00:48:11 crc kubenswrapper[4766]: E1126 00:48:11.690271 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=horizon-operator-controller-manager-5d494799bf-h4mrn_openstack-operators(5284ad70-a36c-4666-8202-84633d83f50b)\"" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" podUID="5284ad70-a36c-4666-8202-84633d83f50b" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.710787 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7g6b2" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.710886 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.762019 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.763190 4766 scope.go:117] "RemoveContainer" containerID="cea7bdd56843f2b2b035ccfb43d7a95792ec34d2199a87ccd3f29a3a8286b383" Nov 26 00:48:11 crc kubenswrapper[4766]: E1126 00:48:11.763498 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=barbican-operator-controller-manager-7b64f4fb85-gbdhz_openstack-operators(cc87f631-ba4e-4943-b51d-bf1012cca754)\"" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" podUID="cc87f631-ba4e-4943-b51d-bf1012cca754" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.839900 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.839949 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.840812 4766 scope.go:117] "RemoveContainer" containerID="e483976d00f30841115f915004cb491c322867e1677e1242b465493dd4116df1" Nov 26 00:48:11 crc kubenswrapper[4766]: E1126 00:48:11.841138 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=heat-operator-controller-manager-5b77f656f-tkj9b_openstack-operators(1ecbd1d2-cb7f-432f-bc75-903d0643ea29)\"" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" podUID="1ecbd1d2-cb7f-432f-bc75-903d0643ea29" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.855017 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.865161 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.887138 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.915756 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.997379 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" Nov 26 00:48:11 crc kubenswrapper[4766]: I1126 00:48:11.998322 4766 scope.go:117] "RemoveContainer" containerID="f852f3819cc2e23f970fe6d5410e8b916b08f77990d1cf12bdeb13de0d480f2d" Nov 26 00:48:11 crc kubenswrapper[4766]: E1126 00:48:11.998556 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=ironic-operator-controller-manager-67cb4dc6d4-qtnxq_openstack-operators(08bd3753-3b70-44f8-92ff-98bedd88f662)\"" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" podUID="08bd3753-3b70-44f8-92ff-98bedd88f662" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.029081 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.029131 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.029845 4766 scope.go:117] "RemoveContainer" containerID="786387060a4e3cb589ce85f666fff7f97910552f1b888209e9019cdeebedf9bd" Nov 26 00:48:12 crc kubenswrapper[4766]: E1126 00:48:12.030135 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-7b4567c7cf-fcr76_openstack-operators(3d0d59b2-0486-40a2-ae60-b738a4f76dc4)\"" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" podUID="3d0d59b2-0486-40a2-ae60-b738a4f76dc4" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.044495 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.060960 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.078953 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.099850 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.100139 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-client-http" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.111402 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.111450 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.112170 4766 scope.go:117] "RemoveContainer" containerID="2616d6e260dc8764705bb96f63f0717a0cfc0738acfaa392b4301101a7942fe2" Nov 26 00:48:12 crc kubenswrapper[4766]: E1126 00:48:12.112442 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=mariadb-operator-controller-manager-66f4dd4bc7-z865g_openstack-operators(749a25e6-23eb-4fdd-8e15-ac2874cdee7f)\"" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" podUID="749a25e6-23eb-4fdd-8e15-ac2874cdee7f" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.113355 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.145622 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.145672 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.146311 4766 scope.go:117] "RemoveContainer" containerID="e2387b88677d93c567a1d130a1054097ae8e786e58aed741f4a8aa02b969a6b3" Nov 26 00:48:12 crc kubenswrapper[4766]: E1126 00:48:12.146560 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=neutron-operator-controller-manager-6fdcddb789-wdzgw_openstack-operators(2bb23a3b-1f4a-409f-9d9c-0afe73a33352)\"" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" podUID="2bb23a3b-1f4a-409f-9d9c-0afe73a33352" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.150455 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.163230 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.163283 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.164134 4766 scope.go:117] "RemoveContainer" containerID="40b84f98ff2e7a66714b3171a6fb601ec8a52d1fcc8b4f251028c0ec0684b6d2" Nov 26 00:48:12 crc kubenswrapper[4766]: E1126 00:48:12.164480 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=nova-operator-controller-manager-79556f57fc-dm7xb_openstack-operators(8969652f-14c9-4618-ad2d-1b3fd1ad3627)\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" podUID="8969652f-14c9-4618-ad2d-1b3fd1ad3627" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.200936 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.201637 4766 scope.go:117] "RemoveContainer" containerID="e8fbaea76abae029d1a7e07d17d13ba1ef29a28831a8d0e6f6fc04e295ff065e" Nov 26 00:48:12 crc kubenswrapper[4766]: E1126 00:48:12.201940 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=octavia-operator-controller-manager-64cdc6ff96-bfw5f_openstack-operators(91d3e9c8-8bbc-43b9-947b-71c4bb474b21)\"" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" podUID="91d3e9c8-8bbc-43b9-947b-71c4bb474b21" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.214407 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.222572 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.222800 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.223562 4766 scope.go:117] "RemoveContainer" containerID="31e4e2562c7a94a89c0e4b9d27edc8f0f6845cbd15437a89c00febff1c4aa709" Nov 26 00:48:12 crc kubenswrapper[4766]: E1126 00:48:12.224635 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=ovn-operator-controller-manager-56897c768d-9zhql_openstack-operators(a2e7c098-bf9d-4d99-84bf-adfb49ee6b12)\"" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" podUID="a2e7c098-bf9d-4d99-84bf-adfb49ee6b12" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.333019 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.334911 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.354005 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.354817 4766 scope.go:117] "RemoveContainer" containerID="e334b081ac2683145ab1dd86d34adf85f1f0aabd3f025419401ad076704959de" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.354819 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" Nov 26 00:48:12 crc kubenswrapper[4766]: E1126 00:48:12.355078 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-d77b94747-2t9w2_openstack-operators(00a32e6c-edcc-4625-ac35-26fed19f1638)\"" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" podUID="00a32e6c-edcc-4625-ac35-26fed19f1638" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.386707 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" event={"ID":"45353fc4-9955-460b-9276-9792dfdaf8f4","Type":"ContainerStarted","Data":"9dbd74b164f2f9373abc946ecc78fcce61ae62e89e00ecaff5fb95c877bdcb6e"} Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.386999 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.388862 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-nd9m2" event={"ID":"954f9d95-f1ab-4e36-9b7a-d16fbe282e53","Type":"ContainerStarted","Data":"3790ebdd75208f009721c30cb0e2f16e9a7e162bcfcd8dc8ac787098d5943a51"} Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.389677 4766 scope.go:117] "RemoveContainer" containerID="00bc96df29ffd032af13f751f10f9ebd94973bdaab39c23fa02a669782573988" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.389920 4766 scope.go:117] "RemoveContainer" containerID="077353144c2eb33b5eb96b9173cb8e009bf57db150620c2e8afe527ca503c06c" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.390116 4766 scope.go:117] "RemoveContainer" containerID="e334b081ac2683145ab1dd86d34adf85f1f0aabd3f025419401ad076704959de" Nov 26 00:48:12 crc kubenswrapper[4766]: E1126 00:48:12.390226 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=horizon-operator-controller-manager-5d494799bf-h4mrn_openstack-operators(5284ad70-a36c-4666-8202-84633d83f50b)\"" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" podUID="5284ad70-a36c-4666-8202-84633d83f50b" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.390291 4766 scope.go:117] "RemoveContainer" containerID="11900098c4ce70adf98b70c668f6fd3114393fbfb5ebae1324aa13e7994d8001" Nov 26 00:48:12 crc kubenswrapper[4766]: E1126 00:48:12.390379 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=cinder-operator-controller-manager-6b7f75547b-jgz6c_openstack-operators(31b1b73b-1d10-49ac-8549-b202c965b63e)\"" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" podUID="31b1b73b-1d10-49ac-8549-b202c965b63e" Nov 26 00:48:12 crc kubenswrapper[4766]: E1126 00:48:12.390518 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-d77b94747-2t9w2_openstack-operators(00a32e6c-edcc-4625-ac35-26fed19f1638)\"" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" podUID="00a32e6c-edcc-4625-ac35-26fed19f1638" Nov 26 00:48:12 crc kubenswrapper[4766]: E1126 00:48:12.391381 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=designate-operator-controller-manager-955677c94-dqtcj_openstack-operators(0e6cfef3-86af-4cf0-ab06-b3eef687836b)\"" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" podUID="0e6cfef3-86af-4cf0-ab06-b3eef687836b" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.437903 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.492998 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"openshift-service-ca.crt" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.563590 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.563666 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.564516 4766 scope.go:117] "RemoveContainer" containerID="b1b0c5aa8f40d0f2c07265a853eba57d2333036b22f55004315b3f36d17ce754" Nov 26 00:48:12 crc kubenswrapper[4766]: E1126 00:48:12.564833 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=telemetry-operator-controller-manager-5ccf99cfdd-9cfft_openstack-operators(04e53d09-de84-4682-b17b-61f9ea7bf152)\"" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" podUID="04e53d09-de84-4682-b17b-61f9ea7bf152" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.589134 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.628793 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.628842 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.629531 4766 scope.go:117] "RemoveContainer" containerID="8137bb426d3c7cbb0adfb5a999511839a3f9b23baec1066df5c7c6b2c64af5e4" Nov 26 00:48:12 crc kubenswrapper[4766]: E1126 00:48:12.629800 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=watcher-operator-controller-manager-656dcb59d4-2hpg8_openstack-operators(751b8e60-46d4-4f5e-b62d-800f641ccd40)\"" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" podUID="751b8e60-46d4-4f5e-b62d-800f641ccd40" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.704136 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.739907 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.827123 4766 scope.go:117] "RemoveContainer" containerID="fc1da18b2031a9683dec239eb1c313b14139ca65098b96f7294b63e1da1f01ac" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.934739 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 26 00:48:12 crc kubenswrapper[4766]: I1126 00:48:12.980553 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 26 00:48:13 crc kubenswrapper[4766]: I1126 00:48:13.059141 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 26 00:48:13 crc kubenswrapper[4766]: I1126 00:48:13.141103 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 26 00:48:13 crc kubenswrapper[4766]: I1126 00:48:13.332766 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 26 00:48:13 crc kubenswrapper[4766]: I1126 00:48:13.405961 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"aa692ff7-2356-4ff4-b9fe-08884e4081cf","Type":"ContainerStarted","Data":"3a54ae3d324cd23367e4b743604acd325d6f1f79b6e146944b0f49878887b855"} Nov 26 00:48:13 crc kubenswrapper[4766]: I1126 00:48:13.406229 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 26 00:48:13 crc kubenswrapper[4766]: I1126 00:48:13.548815 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 26 00:48:13 crc kubenswrapper[4766]: I1126 00:48:13.550039 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 26 00:48:13 crc kubenswrapper[4766]: I1126 00:48:13.555915 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 26 00:48:13 crc kubenswrapper[4766]: I1126 00:48:13.659419 4766 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-7j6jd" Nov 26 00:48:13 crc kubenswrapper[4766]: I1126 00:48:13.670974 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 26 00:48:13 crc kubenswrapper[4766]: I1126 00:48:13.785427 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 26 00:48:13 crc kubenswrapper[4766]: I1126 00:48:13.840976 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 26 00:48:13 crc kubenswrapper[4766]: I1126 00:48:13.841156 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:48:13 crc kubenswrapper[4766]: I1126 00:48:13.841848 4766 scope.go:117] "RemoveContainer" containerID="9ed08fb921ec77ec6e1e8ed96b51841858e7437516717a99a7033a0ee465972b" Nov 26 00:48:13 crc kubenswrapper[4766]: E1126 00:48:13.842086 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=infra-operator-controller-manager-57548d458d-x6h4v_openstack-operators(c3cec209-fdef-44a6-90d0-2d06cfd7e5cc)\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" podUID="c3cec209-fdef-44a6-90d0-2d06cfd7e5cc" Nov 26 00:48:13 crc kubenswrapper[4766]: I1126 00:48:13.979939 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 26 00:48:13 crc kubenswrapper[4766]: I1126 00:48:13.985818 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:48:13 crc kubenswrapper[4766]: I1126 00:48:13.985947 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:48:13 crc kubenswrapper[4766]: I1126 00:48:13.986553 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:48:14 crc kubenswrapper[4766]: I1126 00:48:14.221011 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-c7k77" Nov 26 00:48:14 crc kubenswrapper[4766]: I1126 00:48:14.334310 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 26 00:48:14 crc kubenswrapper[4766]: I1126 00:48:14.451356 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:48:14 crc kubenswrapper[4766]: I1126 00:48:14.553040 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 26 00:48:14 crc kubenswrapper[4766]: I1126 00:48:14.919207 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-5mxk5" Nov 26 00:48:15 crc kubenswrapper[4766]: I1126 00:48:15.014591 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 26 00:48:15 crc kubenswrapper[4766]: I1126 00:48:15.218894 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-84mld" Nov 26 00:48:15 crc kubenswrapper[4766]: I1126 00:48:15.429976 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdd2abaa-1d30-4422-80c9-496e89a9aad9","Type":"ContainerStarted","Data":"772ef3e6ad9388bd5ffed5f4e3e8599d67b0cb08ccd470bc826d64dfa68e21b3"} Nov 26 00:48:15 crc kubenswrapper[4766]: I1126 00:48:15.430030 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdd2abaa-1d30-4422-80c9-496e89a9aad9","Type":"ContainerStarted","Data":"0a967a8f03be6a9be05cfb69b3c9a733cadd958dbc86857bda20484c0d3d6cb5"} Nov 26 00:48:15 crc kubenswrapper[4766]: I1126 00:48:15.588514 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-gd78k" Nov 26 00:48:16 crc kubenswrapper[4766]: I1126 00:48:16.453506 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdd2abaa-1d30-4422-80c9-496e89a9aad9","Type":"ContainerStarted","Data":"f1318ae2848a1d1d3ada1d25bf19bfc50609a50cc494c97f8ff30704a752b896"} Nov 26 00:48:17 crc kubenswrapper[4766]: I1126 00:48:17.465371 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdd2abaa-1d30-4422-80c9-496e89a9aad9","Type":"ContainerStarted","Data":"d03a7c6e516f1357e128e452e362980fce2050b94ca9696ba2c21ac109e2698f"} Nov 26 00:48:18 crc kubenswrapper[4766]: I1126 00:48:18.242125 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:48:18 crc kubenswrapper[4766]: I1126 00:48:18.242830 4766 scope.go:117] "RemoveContainer" containerID="04f968501bccdff8ecc4863f1826684c21db8e3e31f21130322a744ed9db3121" Nov 26 00:48:18 crc kubenswrapper[4766]: E1126 00:48:18.243053 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=openstack-operator-controller-manager-86674bdb4b-vq98m_openstack-operators(d00b9c60-0fb7-49c6-944a-fac598cc22a4)\"" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" podUID="d00b9c60-0fb7-49c6-944a-fac598cc22a4" Nov 26 00:48:18 crc kubenswrapper[4766]: I1126 00:48:18.827378 4766 scope.go:117] "RemoveContainer" containerID="6396f3caca85c89b1192e1772820e1b1f5623132fa5baf5b852754febe7243de" Nov 26 00:48:18 crc kubenswrapper[4766]: E1126 00:48:18.827985 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=metallb-operator-controller-manager-764c8cf56f-vrwpt_metallb-system(fd3a6275-9f47-4799-bc15-96a51b93f31b)\"" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" podUID="fd3a6275-9f47-4799-bc15-96a51b93f31b" Nov 26 00:48:19 crc kubenswrapper[4766]: I1126 00:48:19.490882 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdd2abaa-1d30-4422-80c9-496e89a9aad9","Type":"ContainerStarted","Data":"5016c3aa7d56fbe841e4e97c9618cfd6e543eefc0a230562e677ff7b237bebdf"} Nov 26 00:48:19 crc kubenswrapper[4766]: I1126 00:48:19.491497 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 26 00:48:19 crc kubenswrapper[4766]: I1126 00:48:19.516381 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=11.541447462 podStartE2EDuration="15.516357421s" podCreationTimestamp="2025-11-26 00:48:04 +0000 UTC" firstStartedPulling="2025-11-26 00:48:14.457539006 +0000 UTC m=+1475.306309436" lastFinishedPulling="2025-11-26 00:48:18.432448945 +0000 UTC m=+1479.281219395" observedRunningTime="2025-11-26 00:48:19.509272935 +0000 UTC m=+1480.358043365" watchObservedRunningTime="2025-11-26 00:48:19.516357421 +0000 UTC m=+1480.365127841" Nov 26 00:48:20 crc kubenswrapper[4766]: I1126 00:48:20.424238 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 26 00:48:20 crc kubenswrapper[4766]: I1126 00:48:20.826867 4766 scope.go:117] "RemoveContainer" containerID="0566d4933fc09678c5b262e6d0a2f71deb7b2668a246a353a87132378e3d1324" Nov 26 00:48:20 crc kubenswrapper[4766]: E1126 00:48:20.827424 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=operator pod=rabbitmq-cluster-operator-manager-668c99d594-mx9vw_openstack-operators(91ee52b5-0e2e-4732-87bf-809e378d33d3)\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw" podUID="91ee52b5-0e2e-4732-87bf-809e378d33d3" Nov 26 00:48:21 crc kubenswrapper[4766]: I1126 00:48:21.616301 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-k68q9" Nov 26 00:48:22 crc kubenswrapper[4766]: I1126 00:48:22.082227 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-rs4kn" Nov 26 00:48:22 crc kubenswrapper[4766]: I1126 00:48:22.245323 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-xlxgj" Nov 26 00:48:22 crc kubenswrapper[4766]: I1126 00:48:22.827160 4766 scope.go:117] "RemoveContainer" containerID="e483976d00f30841115f915004cb491c322867e1677e1242b465493dd4116df1" Nov 26 00:48:22 crc kubenswrapper[4766]: I1126 00:48:22.827325 4766 scope.go:117] "RemoveContainer" containerID="077353144c2eb33b5eb96b9173cb8e009bf57db150620c2e8afe527ca503c06c" Nov 26 00:48:22 crc kubenswrapper[4766]: E1126 00:48:22.827701 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=heat-operator-controller-manager-5b77f656f-tkj9b_openstack-operators(1ecbd1d2-cb7f-432f-bc75-903d0643ea29)\"" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" podUID="1ecbd1d2-cb7f-432f-bc75-903d0643ea29" Nov 26 00:48:22 crc kubenswrapper[4766]: E1126 00:48:22.828186 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=horizon-operator-controller-manager-5d494799bf-h4mrn_openstack-operators(5284ad70-a36c-4666-8202-84633d83f50b)\"" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" podUID="5284ad70-a36c-4666-8202-84633d83f50b" Nov 26 00:48:23 crc kubenswrapper[4766]: I1126 00:48:23.828796 4766 scope.go:117] "RemoveContainer" containerID="f852f3819cc2e23f970fe6d5410e8b916b08f77990d1cf12bdeb13de0d480f2d" Nov 26 00:48:23 crc kubenswrapper[4766]: I1126 00:48:23.828955 4766 scope.go:117] "RemoveContainer" containerID="8137bb426d3c7cbb0adfb5a999511839a3f9b23baec1066df5c7c6b2c64af5e4" Nov 26 00:48:23 crc kubenswrapper[4766]: I1126 00:48:23.829315 4766 scope.go:117] "RemoveContainer" containerID="40b84f98ff2e7a66714b3171a6fb601ec8a52d1fcc8b4f251028c0ec0684b6d2" Nov 26 00:48:23 crc kubenswrapper[4766]: E1126 00:48:23.829377 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=watcher-operator-controller-manager-656dcb59d4-2hpg8_openstack-operators(751b8e60-46d4-4f5e-b62d-800f641ccd40)\"" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" podUID="751b8e60-46d4-4f5e-b62d-800f641ccd40" Nov 26 00:48:23 crc kubenswrapper[4766]: I1126 00:48:23.829606 4766 scope.go:117] "RemoveContainer" containerID="e8fbaea76abae029d1a7e07d17d13ba1ef29a28831a8d0e6f6fc04e295ff065e" Nov 26 00:48:23 crc kubenswrapper[4766]: I1126 00:48:23.829806 4766 scope.go:117] "RemoveContainer" containerID="00bc96df29ffd032af13f751f10f9ebd94973bdaab39c23fa02a669782573988" Nov 26 00:48:23 crc kubenswrapper[4766]: E1126 00:48:23.829605 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=nova-operator-controller-manager-79556f57fc-dm7xb_openstack-operators(8969652f-14c9-4618-ad2d-1b3fd1ad3627)\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" podUID="8969652f-14c9-4618-ad2d-1b3fd1ad3627" Nov 26 00:48:23 crc kubenswrapper[4766]: I1126 00:48:23.830560 4766 scope.go:117] "RemoveContainer" containerID="b1b0c5aa8f40d0f2c07265a853eba57d2333036b22f55004315b3f36d17ce754" Nov 26 00:48:23 crc kubenswrapper[4766]: E1126 00:48:23.831508 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=telemetry-operator-controller-manager-5ccf99cfdd-9cfft_openstack-operators(04e53d09-de84-4682-b17b-61f9ea7bf152)\"" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" podUID="04e53d09-de84-4682-b17b-61f9ea7bf152" Nov 26 00:48:23 crc kubenswrapper[4766]: I1126 00:48:23.845728 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:48:23 crc kubenswrapper[4766]: I1126 00:48:23.846786 4766 scope.go:117] "RemoveContainer" containerID="9ed08fb921ec77ec6e1e8ed96b51841858e7437516717a99a7033a0ee465972b" Nov 26 00:48:23 crc kubenswrapper[4766]: E1126 00:48:23.847083 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=infra-operator-controller-manager-57548d458d-x6h4v_openstack-operators(c3cec209-fdef-44a6-90d0-2d06cfd7e5cc)\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" podUID="c3cec209-fdef-44a6-90d0-2d06cfd7e5cc" Nov 26 00:48:24 crc kubenswrapper[4766]: I1126 00:48:24.553526 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" event={"ID":"31b1b73b-1d10-49ac-8549-b202c965b63e","Type":"ContainerStarted","Data":"32d3162f13a3b85802e908791263eff3e3e391a9eb2eb97e2a4d3d04d6a4e5ff"} Nov 26 00:48:24 crc kubenswrapper[4766]: I1126 00:48:24.553832 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" Nov 26 00:48:24 crc kubenswrapper[4766]: I1126 00:48:24.555339 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" event={"ID":"08bd3753-3b70-44f8-92ff-98bedd88f662","Type":"ContainerStarted","Data":"16e4c47c06e12424649711938c8792259c3009758e750896d8fdfc956085566f"} Nov 26 00:48:24 crc kubenswrapper[4766]: I1126 00:48:24.555508 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" Nov 26 00:48:24 crc kubenswrapper[4766]: I1126 00:48:24.557726 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" event={"ID":"91d3e9c8-8bbc-43b9-947b-71c4bb474b21","Type":"ContainerStarted","Data":"45d54d64399991934e45bc3e8066d156f340df8c3d693b94b2bbfffbb0775cff"} Nov 26 00:48:24 crc kubenswrapper[4766]: I1126 00:48:24.557992 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" Nov 26 00:48:24 crc kubenswrapper[4766]: I1126 00:48:24.828281 4766 scope.go:117] "RemoveContainer" containerID="0c594df0d9660266fff1417a3f4f615f9c3a9b37b90eeef4b4d18d0d955a98a2" Nov 26 00:48:24 crc kubenswrapper[4766]: I1126 00:48:24.830484 4766 scope.go:117] "RemoveContainer" containerID="11900098c4ce70adf98b70c668f6fd3114393fbfb5ebae1324aa13e7994d8001" Nov 26 00:48:25 crc kubenswrapper[4766]: I1126 00:48:25.568672 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" event={"ID":"0e6cfef3-86af-4cf0-ab06-b3eef687836b","Type":"ContainerStarted","Data":"74829e56b901ad5d1a2597e90ac2f1f60f19c2d8f8a4c717792e2729dc346b68"} Nov 26 00:48:25 crc kubenswrapper[4766]: I1126 00:48:25.568884 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" Nov 26 00:48:25 crc kubenswrapper[4766]: I1126 00:48:25.570674 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_71bb4a3aecc4ba5b26c4b7318770ce13/kube-apiserver-check-endpoints/2.log" Nov 26 00:48:25 crc kubenswrapper[4766]: I1126 00:48:25.571911 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"76999bd621f3ac55eb49b649efa4c9cde2c326225c1dedcdf0551a2bf95fdfa5"} Nov 26 00:48:25 crc kubenswrapper[4766]: I1126 00:48:25.607198 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=18.607179572 podStartE2EDuration="18.607179572s" podCreationTimestamp="2025-11-26 00:48:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:48:25.595852329 +0000 UTC m=+1486.444622759" watchObservedRunningTime="2025-11-26 00:48:25.607179572 +0000 UTC m=+1486.455950002" Nov 26 00:48:25 crc kubenswrapper[4766]: I1126 00:48:25.827173 4766 scope.go:117] "RemoveContainer" containerID="786387060a4e3cb589ce85f666fff7f97910552f1b888209e9019cdeebedf9bd" Nov 26 00:48:25 crc kubenswrapper[4766]: I1126 00:48:25.827292 4766 scope.go:117] "RemoveContainer" containerID="e2387b88677d93c567a1d130a1054097ae8e786e58aed741f4a8aa02b969a6b3" Nov 26 00:48:25 crc kubenswrapper[4766]: E1126 00:48:25.827424 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-7b4567c7cf-fcr76_openstack-operators(3d0d59b2-0486-40a2-ae60-b738a4f76dc4)\"" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" podUID="3d0d59b2-0486-40a2-ae60-b738a4f76dc4" Nov 26 00:48:25 crc kubenswrapper[4766]: I1126 00:48:25.827634 4766 scope.go:117] "RemoveContainer" containerID="e334b081ac2683145ab1dd86d34adf85f1f0aabd3f025419401ad076704959de" Nov 26 00:48:25 crc kubenswrapper[4766]: E1126 00:48:25.828006 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-d77b94747-2t9w2_openstack-operators(00a32e6c-edcc-4625-ac35-26fed19f1638)\"" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" podUID="00a32e6c-edcc-4625-ac35-26fed19f1638" Nov 26 00:48:25 crc kubenswrapper[4766]: I1126 00:48:25.828686 4766 scope.go:117] "RemoveContainer" containerID="cea7bdd56843f2b2b035ccfb43d7a95792ec34d2199a87ccd3f29a3a8286b383" Nov 26 00:48:25 crc kubenswrapper[4766]: E1126 00:48:25.829838 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=barbican-operator-controller-manager-7b64f4fb85-gbdhz_openstack-operators(cc87f631-ba4e-4943-b51d-bf1012cca754)\"" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" podUID="cc87f631-ba4e-4943-b51d-bf1012cca754" Nov 26 00:48:26 crc kubenswrapper[4766]: I1126 00:48:26.588674 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" event={"ID":"2bb23a3b-1f4a-409f-9d9c-0afe73a33352","Type":"ContainerStarted","Data":"b892243455d00cc1c97ba48ee6babea3460d4917c8434bb7961af2d673d7d8fc"} Nov 26 00:48:26 crc kubenswrapper[4766]: I1126 00:48:26.589002 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" Nov 26 00:48:27 crc kubenswrapper[4766]: I1126 00:48:27.827444 4766 scope.go:117] "RemoveContainer" containerID="2616d6e260dc8764705bb96f63f0717a0cfc0738acfaa392b4301101a7942fe2" Nov 26 00:48:27 crc kubenswrapper[4766]: I1126 00:48:27.827913 4766 scope.go:117] "RemoveContainer" containerID="31e4e2562c7a94a89c0e4b9d27edc8f0f6845cbd15437a89c00febff1c4aa709" Nov 26 00:48:27 crc kubenswrapper[4766]: I1126 00:48:27.855548 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:48:28 crc kubenswrapper[4766]: I1126 00:48:28.242681 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:48:28 crc kubenswrapper[4766]: I1126 00:48:28.243362 4766 scope.go:117] "RemoveContainer" containerID="04f968501bccdff8ecc4863f1826684c21db8e3e31f21130322a744ed9db3121" Nov 26 00:48:28 crc kubenswrapper[4766]: E1126 00:48:28.243690 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=openstack-operator-controller-manager-86674bdb4b-vq98m_openstack-operators(d00b9c60-0fb7-49c6-944a-fac598cc22a4)\"" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" podUID="d00b9c60-0fb7-49c6-944a-fac598cc22a4" Nov 26 00:48:28 crc kubenswrapper[4766]: I1126 00:48:28.609349 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" event={"ID":"749a25e6-23eb-4fdd-8e15-ac2874cdee7f","Type":"ContainerStarted","Data":"babba42a7b9eeb768c2be5e0b75f87369e4963e7565ae0ff50d4086c81d94710"} Nov 26 00:48:28 crc kubenswrapper[4766]: I1126 00:48:28.609561 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" Nov 26 00:48:28 crc kubenswrapper[4766]: I1126 00:48:28.612725 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" event={"ID":"a2e7c098-bf9d-4d99-84bf-adfb49ee6b12","Type":"ContainerStarted","Data":"385a6049b90ad63e87b276858f390266ad24c899cec59a37328de21775a786f0"} Nov 26 00:48:28 crc kubenswrapper[4766]: I1126 00:48:28.612967 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" Nov 26 00:48:29 crc kubenswrapper[4766]: I1126 00:48:29.839022 4766 scope.go:117] "RemoveContainer" containerID="6396f3caca85c89b1192e1772820e1b1f5623132fa5baf5b852754febe7243de" Nov 26 00:48:30 crc kubenswrapper[4766]: I1126 00:48:30.638481 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" event={"ID":"fd3a6275-9f47-4799-bc15-96a51b93f31b","Type":"ContainerStarted","Data":"af431e8472f9b4435b777d8837873f2e18adba4ee02ffd8b3b08876f251024b3"} Nov 26 00:48:30 crc kubenswrapper[4766]: I1126 00:48:30.639072 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" Nov 26 00:48:30 crc kubenswrapper[4766]: I1126 00:48:30.847516 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 26 00:48:31 crc kubenswrapper[4766]: I1126 00:48:31.470269 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-jgz6c" Nov 26 00:48:31 crc kubenswrapper[4766]: I1126 00:48:31.485993 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=1.485974538 podStartE2EDuration="1.485974538s" podCreationTimestamp="2025-11-26 00:48:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:48:31.480890581 +0000 UTC m=+1492.329661001" watchObservedRunningTime="2025-11-26 00:48:31.485974538 +0000 UTC m=+1492.334744968" Nov 26 00:48:31 crc kubenswrapper[4766]: I1126 00:48:31.508486 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-955677c94-dqtcj" Nov 26 00:48:31 crc kubenswrapper[4766]: I1126 00:48:31.998998 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-qtnxq" Nov 26 00:48:32 crc kubenswrapper[4766]: I1126 00:48:32.148733 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-wdzgw" Nov 26 00:48:32 crc kubenswrapper[4766]: I1126 00:48:32.204231 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-bfw5f" Nov 26 00:48:34 crc kubenswrapper[4766]: I1126 00:48:34.829973 4766 scope.go:117] "RemoveContainer" containerID="077353144c2eb33b5eb96b9173cb8e009bf57db150620c2e8afe527ca503c06c" Nov 26 00:48:34 crc kubenswrapper[4766]: I1126 00:48:34.830513 4766 scope.go:117] "RemoveContainer" containerID="b1b0c5aa8f40d0f2c07265a853eba57d2333036b22f55004315b3f36d17ce754" Nov 26 00:48:35 crc kubenswrapper[4766]: I1126 00:48:35.690766 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" event={"ID":"04e53d09-de84-4682-b17b-61f9ea7bf152","Type":"ContainerStarted","Data":"39915bc29f7810b9244185662497c71f45ab3fc474998d07a559f783acc3b9ec"} Nov 26 00:48:35 crc kubenswrapper[4766]: I1126 00:48:35.691563 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" Nov 26 00:48:35 crc kubenswrapper[4766]: I1126 00:48:35.693761 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" event={"ID":"5284ad70-a36c-4666-8202-84633d83f50b","Type":"ContainerStarted","Data":"2218c175b2061baa24a11220c911d224e26c141cadd91408b3c9838fc964c76b"} Nov 26 00:48:35 crc kubenswrapper[4766]: I1126 00:48:35.693985 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" Nov 26 00:48:35 crc kubenswrapper[4766]: I1126 00:48:35.826851 4766 scope.go:117] "RemoveContainer" containerID="0566d4933fc09678c5b262e6d0a2f71deb7b2668a246a353a87132378e3d1324" Nov 26 00:48:35 crc kubenswrapper[4766]: I1126 00:48:35.827162 4766 scope.go:117] "RemoveContainer" containerID="40b84f98ff2e7a66714b3171a6fb601ec8a52d1fcc8b4f251028c0ec0684b6d2" Nov 26 00:48:36 crc kubenswrapper[4766]: I1126 00:48:36.706201 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" event={"ID":"8969652f-14c9-4618-ad2d-1b3fd1ad3627","Type":"ContainerStarted","Data":"ebbd42a7aa85d5b13435cc3d20d700442ca8ce562abd9dd4c1bb2fac2ce593c2"} Nov 26 00:48:36 crc kubenswrapper[4766]: I1126 00:48:36.706476 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" Nov 26 00:48:36 crc kubenswrapper[4766]: I1126 00:48:36.708259 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mx9vw" event={"ID":"91ee52b5-0e2e-4732-87bf-809e378d33d3","Type":"ContainerStarted","Data":"04f7909f86cd86b353c9b4cf66fe3e399f8503dcdb7d72cb45cf92b1281feafb"} Nov 26 00:48:36 crc kubenswrapper[4766]: I1126 00:48:36.826759 4766 scope.go:117] "RemoveContainer" containerID="9ed08fb921ec77ec6e1e8ed96b51841858e7437516717a99a7033a0ee465972b" Nov 26 00:48:36 crc kubenswrapper[4766]: I1126 00:48:36.827710 4766 scope.go:117] "RemoveContainer" containerID="e483976d00f30841115f915004cb491c322867e1677e1242b465493dd4116df1" Nov 26 00:48:37 crc kubenswrapper[4766]: I1126 00:48:37.295515 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 26 00:48:37 crc kubenswrapper[4766]: I1126 00:48:37.729684 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" event={"ID":"1ecbd1d2-cb7f-432f-bc75-903d0643ea29","Type":"ContainerStarted","Data":"0d0e99ab20d52744ac9154ded2e5b620eaad3222727ac7b2b32f7bdc6f35f403"} Nov 26 00:48:37 crc kubenswrapper[4766]: I1126 00:48:37.730215 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" Nov 26 00:48:37 crc kubenswrapper[4766]: I1126 00:48:37.736122 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" event={"ID":"c3cec209-fdef-44a6-90d0-2d06cfd7e5cc","Type":"ContainerStarted","Data":"b7900ecea8137f808bf6022d5c8be422cc02ba1c818bf7000a2ade673798754a"} Nov 26 00:48:37 crc kubenswrapper[4766]: I1126 00:48:37.736588 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:48:37 crc kubenswrapper[4766]: I1126 00:48:37.827235 4766 scope.go:117] "RemoveContainer" containerID="8137bb426d3c7cbb0adfb5a999511839a3f9b23baec1066df5c7c6b2c64af5e4" Nov 26 00:48:37 crc kubenswrapper[4766]: I1126 00:48:37.864789 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 00:48:38 crc kubenswrapper[4766]: I1126 00:48:38.753912 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" event={"ID":"751b8e60-46d4-4f5e-b62d-800f641ccd40","Type":"ContainerStarted","Data":"df925f4b039d9f92d9268cf522769dbbc63556d1f3d15e2b5f1b25c2ab3cba37"} Nov 26 00:48:38 crc kubenswrapper[4766]: I1126 00:48:38.754873 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" Nov 26 00:48:38 crc kubenswrapper[4766]: I1126 00:48:38.826709 4766 scope.go:117] "RemoveContainer" containerID="786387060a4e3cb589ce85f666fff7f97910552f1b888209e9019cdeebedf9bd" Nov 26 00:48:39 crc kubenswrapper[4766]: I1126 00:48:39.768699 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" event={"ID":"3d0d59b2-0486-40a2-ae60-b738a4f76dc4","Type":"ContainerStarted","Data":"009dd1adffbf2deae69ffe9eb4245e11a9c2fa20665ecc2698d75f98149e0b06"} Nov 26 00:48:39 crc kubenswrapper[4766]: I1126 00:48:39.769330 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" Nov 26 00:48:40 crc kubenswrapper[4766]: I1126 00:48:40.826805 4766 scope.go:117] "RemoveContainer" containerID="e334b081ac2683145ab1dd86d34adf85f1f0aabd3f025419401ad076704959de" Nov 26 00:48:40 crc kubenswrapper[4766]: I1126 00:48:40.828009 4766 scope.go:117] "RemoveContainer" containerID="cea7bdd56843f2b2b035ccfb43d7a95792ec34d2199a87ccd3f29a3a8286b383" Nov 26 00:48:40 crc kubenswrapper[4766]: I1126 00:48:40.828678 4766 scope.go:117] "RemoveContainer" containerID="04f968501bccdff8ecc4863f1826684c21db8e3e31f21130322a744ed9db3121" Nov 26 00:48:41 crc kubenswrapper[4766]: I1126 00:48:41.268207 4766 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 26 00:48:41 crc kubenswrapper[4766]: I1126 00:48:41.268737 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://27937a6f6f497f5f7c3f3a89a54f4857db004c27ee4f74df829476bd1a716a89" gracePeriod=5 Nov 26 00:48:41 crc kubenswrapper[4766]: I1126 00:48:41.479927 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:48:41 crc kubenswrapper[4766]: I1126 00:48:41.480226 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:48:41 crc kubenswrapper[4766]: I1126 00:48:41.691082 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-h4mrn" Nov 26 00:48:41 crc kubenswrapper[4766]: I1126 00:48:41.795155 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" event={"ID":"cc87f631-ba4e-4943-b51d-bf1012cca754","Type":"ContainerStarted","Data":"70579234d8113a31ca01e59981263d27b8cc6e0686c94cd6cb9af790ae4a5a22"} Nov 26 00:48:41 crc kubenswrapper[4766]: I1126 00:48:41.796522 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" Nov 26 00:48:41 crc kubenswrapper[4766]: I1126 00:48:41.799713 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" event={"ID":"00a32e6c-edcc-4625-ac35-26fed19f1638","Type":"ContainerStarted","Data":"5b97356a0dfa45f38cedd1b3a01d5310c959301ad58001ee16b94a48c451f5e6"} Nov 26 00:48:41 crc kubenswrapper[4766]: I1126 00:48:41.799985 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" Nov 26 00:48:41 crc kubenswrapper[4766]: I1126 00:48:41.801598 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" event={"ID":"d00b9c60-0fb7-49c6-944a-fac598cc22a4","Type":"ContainerStarted","Data":"fae999e2b1a8e59b9d29b4ec91e5bab78f63b045857450685dfd34e3592d122a"} Nov 26 00:48:41 crc kubenswrapper[4766]: I1126 00:48:41.802396 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:48:42 crc kubenswrapper[4766]: I1126 00:48:42.113205 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-z865g" Nov 26 00:48:42 crc kubenswrapper[4766]: I1126 00:48:42.169392 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-dm7xb" Nov 26 00:48:42 crc kubenswrapper[4766]: I1126 00:48:42.224447 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-9zhql" Nov 26 00:48:42 crc kubenswrapper[4766]: I1126 00:48:42.568196 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5ccf99cfdd-9cfft" Nov 26 00:48:43 crc kubenswrapper[4766]: I1126 00:48:43.850085 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-x6h4v" Nov 26 00:48:46 crc kubenswrapper[4766]: I1126 00:48:46.855610 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 26 00:48:46 crc kubenswrapper[4766]: I1126 00:48:46.856168 4766 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="27937a6f6f497f5f7c3f3a89a54f4857db004c27ee4f74df829476bd1a716a89" exitCode=137 Nov 26 00:48:46 crc kubenswrapper[4766]: I1126 00:48:46.856203 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="add7c42a346acd8d772af3a4ba488def09ece74ce7425d4686a52fae27c7b583" Nov 26 00:48:46 crc kubenswrapper[4766]: I1126 00:48:46.933600 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 26 00:48:46 crc kubenswrapper[4766]: I1126 00:48:46.933689 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.053002 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.053066 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.053094 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.053181 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.053216 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.053274 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.053764 4766 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.053797 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.053793 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.053809 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.061866 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.155666 4766 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.155691 4766 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.155702 4766 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.155710 4766 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.841799 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.842812 4766 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.857200 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.857237 4766 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="d3860325-d4f1-4fc3-a226-5bf3c8b39956" Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.869385 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.872550 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 26 00:48:47 crc kubenswrapper[4766]: I1126 00:48:47.872590 4766 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="d3860325-d4f1-4fc3-a226-5bf3c8b39956" Nov 26 00:48:48 crc kubenswrapper[4766]: I1126 00:48:48.249280 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-86674bdb4b-vq98m" Nov 26 00:48:51 crc kubenswrapper[4766]: I1126 00:48:51.763249 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-gbdhz" Nov 26 00:48:51 crc kubenswrapper[4766]: I1126 00:48:51.877081 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-tkj9b" Nov 26 00:48:52 crc kubenswrapper[4766]: I1126 00:48:52.030322 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-fcr76" Nov 26 00:48:52 crc kubenswrapper[4766]: I1126 00:48:52.356194 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2t9w2" Nov 26 00:48:52 crc kubenswrapper[4766]: I1126 00:48:52.632016 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-2hpg8" Nov 26 00:49:03 crc kubenswrapper[4766]: I1126 00:49:03.530522 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sswvf"] Nov 26 00:49:03 crc kubenswrapper[4766]: E1126 00:49:03.531407 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 26 00:49:03 crc kubenswrapper[4766]: I1126 00:49:03.531419 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 26 00:49:03 crc kubenswrapper[4766]: I1126 00:49:03.531618 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 26 00:49:03 crc kubenswrapper[4766]: I1126 00:49:03.533053 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sswvf" Nov 26 00:49:03 crc kubenswrapper[4766]: I1126 00:49:03.555704 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sswvf"] Nov 26 00:49:03 crc kubenswrapper[4766]: I1126 00:49:03.559024 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b171321-6be1-4d99-ba59-ec6e80499900-utilities\") pod \"certified-operators-sswvf\" (UID: \"6b171321-6be1-4d99-ba59-ec6e80499900\") " pod="openshift-marketplace/certified-operators-sswvf" Nov 26 00:49:03 crc kubenswrapper[4766]: I1126 00:49:03.559089 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v78w6\" (UniqueName: \"kubernetes.io/projected/6b171321-6be1-4d99-ba59-ec6e80499900-kube-api-access-v78w6\") pod \"certified-operators-sswvf\" (UID: \"6b171321-6be1-4d99-ba59-ec6e80499900\") " pod="openshift-marketplace/certified-operators-sswvf" Nov 26 00:49:03 crc kubenswrapper[4766]: I1126 00:49:03.559158 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b171321-6be1-4d99-ba59-ec6e80499900-catalog-content\") pod \"certified-operators-sswvf\" (UID: \"6b171321-6be1-4d99-ba59-ec6e80499900\") " pod="openshift-marketplace/certified-operators-sswvf" Nov 26 00:49:03 crc kubenswrapper[4766]: I1126 00:49:03.661160 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b171321-6be1-4d99-ba59-ec6e80499900-utilities\") pod \"certified-operators-sswvf\" (UID: \"6b171321-6be1-4d99-ba59-ec6e80499900\") " pod="openshift-marketplace/certified-operators-sswvf" Nov 26 00:49:03 crc kubenswrapper[4766]: I1126 00:49:03.661238 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v78w6\" (UniqueName: \"kubernetes.io/projected/6b171321-6be1-4d99-ba59-ec6e80499900-kube-api-access-v78w6\") pod \"certified-operators-sswvf\" (UID: \"6b171321-6be1-4d99-ba59-ec6e80499900\") " pod="openshift-marketplace/certified-operators-sswvf" Nov 26 00:49:03 crc kubenswrapper[4766]: I1126 00:49:03.661295 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b171321-6be1-4d99-ba59-ec6e80499900-catalog-content\") pod \"certified-operators-sswvf\" (UID: \"6b171321-6be1-4d99-ba59-ec6e80499900\") " pod="openshift-marketplace/certified-operators-sswvf" Nov 26 00:49:03 crc kubenswrapper[4766]: I1126 00:49:03.661757 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b171321-6be1-4d99-ba59-ec6e80499900-utilities\") pod \"certified-operators-sswvf\" (UID: \"6b171321-6be1-4d99-ba59-ec6e80499900\") " pod="openshift-marketplace/certified-operators-sswvf" Nov 26 00:49:03 crc kubenswrapper[4766]: I1126 00:49:03.661799 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b171321-6be1-4d99-ba59-ec6e80499900-catalog-content\") pod \"certified-operators-sswvf\" (UID: \"6b171321-6be1-4d99-ba59-ec6e80499900\") " pod="openshift-marketplace/certified-operators-sswvf" Nov 26 00:49:03 crc kubenswrapper[4766]: I1126 00:49:03.701732 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v78w6\" (UniqueName: \"kubernetes.io/projected/6b171321-6be1-4d99-ba59-ec6e80499900-kube-api-access-v78w6\") pod \"certified-operators-sswvf\" (UID: \"6b171321-6be1-4d99-ba59-ec6e80499900\") " pod="openshift-marketplace/certified-operators-sswvf" Nov 26 00:49:03 crc kubenswrapper[4766]: I1126 00:49:03.856218 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sswvf" Nov 26 00:49:04 crc kubenswrapper[4766]: W1126 00:49:04.366909 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b171321_6be1_4d99_ba59_ec6e80499900.slice/crio-384014c196579f784761066a0d07652f0244b28f24841eb8a210f9e7fbb92a9f WatchSource:0}: Error finding container 384014c196579f784761066a0d07652f0244b28f24841eb8a210f9e7fbb92a9f: Status 404 returned error can't find the container with id 384014c196579f784761066a0d07652f0244b28f24841eb8a210f9e7fbb92a9f Nov 26 00:49:04 crc kubenswrapper[4766]: I1126 00:49:04.385353 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sswvf"] Nov 26 00:49:04 crc kubenswrapper[4766]: I1126 00:49:04.415543 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-764c8cf56f-vrwpt" Nov 26 00:49:05 crc kubenswrapper[4766]: I1126 00:49:05.089421 4766 generic.go:334] "Generic (PLEG): container finished" podID="6b171321-6be1-4d99-ba59-ec6e80499900" containerID="5d7ab06e44b293e6afb2de2cdd4c0b2d4e2349124daa7a7eb90d39d26614a03c" exitCode=0 Nov 26 00:49:05 crc kubenswrapper[4766]: I1126 00:49:05.089467 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sswvf" event={"ID":"6b171321-6be1-4d99-ba59-ec6e80499900","Type":"ContainerDied","Data":"5d7ab06e44b293e6afb2de2cdd4c0b2d4e2349124daa7a7eb90d39d26614a03c"} Nov 26 00:49:05 crc kubenswrapper[4766]: I1126 00:49:05.089499 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sswvf" event={"ID":"6b171321-6be1-4d99-ba59-ec6e80499900","Type":"ContainerStarted","Data":"384014c196579f784761066a0d07652f0244b28f24841eb8a210f9e7fbb92a9f"} Nov 26 00:49:05 crc kubenswrapper[4766]: I1126 00:49:05.092137 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 00:49:06 crc kubenswrapper[4766]: I1126 00:49:06.104179 4766 generic.go:334] "Generic (PLEG): container finished" podID="6b171321-6be1-4d99-ba59-ec6e80499900" containerID="b39cf8696a92e8d2038ebd430f3d20761def0b2745b98cc2925cd613bcf8b7a1" exitCode=0 Nov 26 00:49:06 crc kubenswrapper[4766]: I1126 00:49:06.104228 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sswvf" event={"ID":"6b171321-6be1-4d99-ba59-ec6e80499900","Type":"ContainerDied","Data":"b39cf8696a92e8d2038ebd430f3d20761def0b2745b98cc2925cd613bcf8b7a1"} Nov 26 00:49:07 crc kubenswrapper[4766]: I1126 00:49:07.115787 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sswvf" event={"ID":"6b171321-6be1-4d99-ba59-ec6e80499900","Type":"ContainerStarted","Data":"152ff07a3de0e95758d71246f12cd1c9dd222b2ebf5cf8ef62f79997239a0aa6"} Nov 26 00:49:07 crc kubenswrapper[4766]: I1126 00:49:07.145023 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sswvf" podStartSLOduration=2.660156517 podStartE2EDuration="4.145005103s" podCreationTimestamp="2025-11-26 00:49:03 +0000 UTC" firstStartedPulling="2025-11-26 00:49:05.091880387 +0000 UTC m=+1525.940650827" lastFinishedPulling="2025-11-26 00:49:06.576728993 +0000 UTC m=+1527.425499413" observedRunningTime="2025-11-26 00:49:07.137743792 +0000 UTC m=+1527.986514222" watchObservedRunningTime="2025-11-26 00:49:07.145005103 +0000 UTC m=+1527.993775533" Nov 26 00:49:07 crc kubenswrapper[4766]: I1126 00:49:07.886838 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-c562-account-create-update-gpcbx"] Nov 26 00:49:07 crc kubenswrapper[4766]: I1126 00:49:07.888239 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-c562-account-create-update-gpcbx" Nov 26 00:49:07 crc kubenswrapper[4766]: I1126 00:49:07.890047 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Nov 26 00:49:07 crc kubenswrapper[4766]: I1126 00:49:07.900514 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-c562-account-create-update-gpcbx"] Nov 26 00:49:07 crc kubenswrapper[4766]: I1126 00:49:07.920632 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-gk5ww"] Nov 26 00:49:07 crc kubenswrapper[4766]: I1126 00:49:07.922390 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-gk5ww" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.016917 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-gk5ww"] Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.045665 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-vc4ps"] Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.047193 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vc4ps" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.051222 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.051628 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.064707 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vc4ps"] Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.068138 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvdmn\" (UniqueName: \"kubernetes.io/projected/8f161def-0ff3-4f5d-8f9f-083800b5e1f2-kube-api-access-rvdmn\") pod \"aodh-db-create-gk5ww\" (UID: \"8f161def-0ff3-4f5d-8f9f-083800b5e1f2\") " pod="openstack/aodh-db-create-gk5ww" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.068211 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9nnh\" (UniqueName: \"kubernetes.io/projected/53093344-eff4-438d-8782-9d131efa8427-kube-api-access-d9nnh\") pod \"aodh-c562-account-create-update-gpcbx\" (UID: \"53093344-eff4-438d-8782-9d131efa8427\") " pod="openstack/aodh-c562-account-create-update-gpcbx" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.068283 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53093344-eff4-438d-8782-9d131efa8427-operator-scripts\") pod \"aodh-c562-account-create-update-gpcbx\" (UID: \"53093344-eff4-438d-8782-9d131efa8427\") " pod="openstack/aodh-c562-account-create-update-gpcbx" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.068332 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f161def-0ff3-4f5d-8f9f-083800b5e1f2-operator-scripts\") pod \"aodh-db-create-gk5ww\" (UID: \"8f161def-0ff3-4f5d-8f9f-083800b5e1f2\") " pod="openstack/aodh-db-create-gk5ww" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.130470 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.132784 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.139041 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.146447 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.171123 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvdmn\" (UniqueName: \"kubernetes.io/projected/8f161def-0ff3-4f5d-8f9f-083800b5e1f2-kube-api-access-rvdmn\") pod \"aodh-db-create-gk5ww\" (UID: \"8f161def-0ff3-4f5d-8f9f-083800b5e1f2\") " pod="openstack/aodh-db-create-gk5ww" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.172073 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-scripts\") pod \"nova-cell0-cell-mapping-vc4ps\" (UID: \"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c\") " pod="openstack/nova-cell0-cell-mapping-vc4ps" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.174479 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9nnh\" (UniqueName: \"kubernetes.io/projected/53093344-eff4-438d-8782-9d131efa8427-kube-api-access-d9nnh\") pod \"aodh-c562-account-create-update-gpcbx\" (UID: \"53093344-eff4-438d-8782-9d131efa8427\") " pod="openstack/aodh-c562-account-create-update-gpcbx" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.174618 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pslsg\" (UniqueName: \"kubernetes.io/projected/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-kube-api-access-pslsg\") pod \"nova-cell0-cell-mapping-vc4ps\" (UID: \"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c\") " pod="openstack/nova-cell0-cell-mapping-vc4ps" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.174919 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-config-data\") pod \"nova-cell0-cell-mapping-vc4ps\" (UID: \"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c\") " pod="openstack/nova-cell0-cell-mapping-vc4ps" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.179860 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53093344-eff4-438d-8782-9d131efa8427-operator-scripts\") pod \"aodh-c562-account-create-update-gpcbx\" (UID: \"53093344-eff4-438d-8782-9d131efa8427\") " pod="openstack/aodh-c562-account-create-update-gpcbx" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.180051 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vc4ps\" (UID: \"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c\") " pod="openstack/nova-cell0-cell-mapping-vc4ps" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.180159 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f161def-0ff3-4f5d-8f9f-083800b5e1f2-operator-scripts\") pod \"aodh-db-create-gk5ww\" (UID: \"8f161def-0ff3-4f5d-8f9f-083800b5e1f2\") " pod="openstack/aodh-db-create-gk5ww" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.180553 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53093344-eff4-438d-8782-9d131efa8427-operator-scripts\") pod \"aodh-c562-account-create-update-gpcbx\" (UID: \"53093344-eff4-438d-8782-9d131efa8427\") " pod="openstack/aodh-c562-account-create-update-gpcbx" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.180839 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f161def-0ff3-4f5d-8f9f-083800b5e1f2-operator-scripts\") pod \"aodh-db-create-gk5ww\" (UID: \"8f161def-0ff3-4f5d-8f9f-083800b5e1f2\") " pod="openstack/aodh-db-create-gk5ww" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.207978 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.209802 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.213240 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.230261 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9nnh\" (UniqueName: \"kubernetes.io/projected/53093344-eff4-438d-8782-9d131efa8427-kube-api-access-d9nnh\") pod \"aodh-c562-account-create-update-gpcbx\" (UID: \"53093344-eff4-438d-8782-9d131efa8427\") " pod="openstack/aodh-c562-account-create-update-gpcbx" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.234036 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.236434 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvdmn\" (UniqueName: \"kubernetes.io/projected/8f161def-0ff3-4f5d-8f9f-083800b5e1f2-kube-api-access-rvdmn\") pod \"aodh-db-create-gk5ww\" (UID: \"8f161def-0ff3-4f5d-8f9f-083800b5e1f2\") " pod="openstack/aodh-db-create-gk5ww" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.243208 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.244529 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.248943 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.265129 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-gk5ww" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.283624 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.284986 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfd3ab8a-e243-40a4-b8fc-b9c094339c78-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"cfd3ab8a-e243-40a4-b8fc-b9c094339c78\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.285036 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjdr7\" (UniqueName: \"kubernetes.io/projected/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-kube-api-access-gjdr7\") pod \"nova-metadata-0\" (UID: \"6557fcf7-bd46-4552-bc2b-339ff9c9efbe\") " pod="openstack/nova-metadata-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.285067 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab-config-data\") pod \"nova-scheduler-0\" (UID: \"8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.285090 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vc4ps\" (UID: \"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c\") " pod="openstack/nova-cell0-cell-mapping-vc4ps" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.285111 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l742j\" (UniqueName: \"kubernetes.io/projected/cfd3ab8a-e243-40a4-b8fc-b9c094339c78-kube-api-access-l742j\") pod \"nova-cell1-novncproxy-0\" (UID: \"cfd3ab8a-e243-40a4-b8fc-b9c094339c78\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.285268 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zczv\" (UniqueName: \"kubernetes.io/projected/8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab-kube-api-access-6zczv\") pod \"nova-scheduler-0\" (UID: \"8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.285328 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfd3ab8a-e243-40a4-b8fc-b9c094339c78-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"cfd3ab8a-e243-40a4-b8fc-b9c094339c78\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.285534 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6557fcf7-bd46-4552-bc2b-339ff9c9efbe\") " pod="openstack/nova-metadata-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.285771 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.285946 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-logs\") pod \"nova-metadata-0\" (UID: \"6557fcf7-bd46-4552-bc2b-339ff9c9efbe\") " pod="openstack/nova-metadata-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.286082 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-scripts\") pod \"nova-cell0-cell-mapping-vc4ps\" (UID: \"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c\") " pod="openstack/nova-cell0-cell-mapping-vc4ps" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.286197 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pslsg\" (UniqueName: \"kubernetes.io/projected/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-kube-api-access-pslsg\") pod \"nova-cell0-cell-mapping-vc4ps\" (UID: \"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c\") " pod="openstack/nova-cell0-cell-mapping-vc4ps" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.286331 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-config-data\") pod \"nova-cell0-cell-mapping-vc4ps\" (UID: \"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c\") " pod="openstack/nova-cell0-cell-mapping-vc4ps" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.286518 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-config-data\") pod \"nova-metadata-0\" (UID: \"6557fcf7-bd46-4552-bc2b-339ff9c9efbe\") " pod="openstack/nova-metadata-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.344887 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-config-data\") pod \"nova-cell0-cell-mapping-vc4ps\" (UID: \"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c\") " pod="openstack/nova-cell0-cell-mapping-vc4ps" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.346171 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-scripts\") pod \"nova-cell0-cell-mapping-vc4ps\" (UID: \"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c\") " pod="openstack/nova-cell0-cell-mapping-vc4ps" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.374569 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pslsg\" (UniqueName: \"kubernetes.io/projected/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-kube-api-access-pslsg\") pod \"nova-cell0-cell-mapping-vc4ps\" (UID: \"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c\") " pod="openstack/nova-cell0-cell-mapping-vc4ps" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.376635 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vc4ps\" (UID: \"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c\") " pod="openstack/nova-cell0-cell-mapping-vc4ps" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.388744 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.390619 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.392590 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.394769 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6557fcf7-bd46-4552-bc2b-339ff9c9efbe\") " pod="openstack/nova-metadata-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.394828 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66040b18-6e3b-43dc-8d90-a3eeab8f31df-logs\") pod \"nova-api-0\" (UID: \"66040b18-6e3b-43dc-8d90-a3eeab8f31df\") " pod="openstack/nova-api-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.394849 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.394871 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-logs\") pod \"nova-metadata-0\" (UID: \"6557fcf7-bd46-4552-bc2b-339ff9c9efbe\") " pod="openstack/nova-metadata-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.394894 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66040b18-6e3b-43dc-8d90-a3eeab8f31df-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"66040b18-6e3b-43dc-8d90-a3eeab8f31df\") " pod="openstack/nova-api-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.394934 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xdmn\" (UniqueName: \"kubernetes.io/projected/66040b18-6e3b-43dc-8d90-a3eeab8f31df-kube-api-access-4xdmn\") pod \"nova-api-0\" (UID: \"66040b18-6e3b-43dc-8d90-a3eeab8f31df\") " pod="openstack/nova-api-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.394966 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-config-data\") pod \"nova-metadata-0\" (UID: \"6557fcf7-bd46-4552-bc2b-339ff9c9efbe\") " pod="openstack/nova-metadata-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.394987 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfd3ab8a-e243-40a4-b8fc-b9c094339c78-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"cfd3ab8a-e243-40a4-b8fc-b9c094339c78\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.395024 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjdr7\" (UniqueName: \"kubernetes.io/projected/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-kube-api-access-gjdr7\") pod \"nova-metadata-0\" (UID: \"6557fcf7-bd46-4552-bc2b-339ff9c9efbe\") " pod="openstack/nova-metadata-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.395052 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab-config-data\") pod \"nova-scheduler-0\" (UID: \"8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.395070 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l742j\" (UniqueName: \"kubernetes.io/projected/cfd3ab8a-e243-40a4-b8fc-b9c094339c78-kube-api-access-l742j\") pod \"nova-cell1-novncproxy-0\" (UID: \"cfd3ab8a-e243-40a4-b8fc-b9c094339c78\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.395111 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zczv\" (UniqueName: \"kubernetes.io/projected/8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab-kube-api-access-6zczv\") pod \"nova-scheduler-0\" (UID: \"8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.395131 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66040b18-6e3b-43dc-8d90-a3eeab8f31df-config-data\") pod \"nova-api-0\" (UID: \"66040b18-6e3b-43dc-8d90-a3eeab8f31df\") " pod="openstack/nova-api-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.395152 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfd3ab8a-e243-40a4-b8fc-b9c094339c78-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"cfd3ab8a-e243-40a4-b8fc-b9c094339c78\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.410388 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.416310 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-logs\") pod \"nova-metadata-0\" (UID: \"6557fcf7-bd46-4552-bc2b-339ff9c9efbe\") " pod="openstack/nova-metadata-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.429688 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.433467 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfd3ab8a-e243-40a4-b8fc-b9c094339c78-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"cfd3ab8a-e243-40a4-b8fc-b9c094339c78\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.438476 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfd3ab8a-e243-40a4-b8fc-b9c094339c78-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"cfd3ab8a-e243-40a4-b8fc-b9c094339c78\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.438599 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6557fcf7-bd46-4552-bc2b-339ff9c9efbe\") " pod="openstack/nova-metadata-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.439334 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-config-data\") pod \"nova-metadata-0\" (UID: \"6557fcf7-bd46-4552-bc2b-339ff9c9efbe\") " pod="openstack/nova-metadata-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.445291 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab-config-data\") pod \"nova-scheduler-0\" (UID: \"8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.450138 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l742j\" (UniqueName: \"kubernetes.io/projected/cfd3ab8a-e243-40a4-b8fc-b9c094339c78-kube-api-access-l742j\") pod \"nova-cell1-novncproxy-0\" (UID: \"cfd3ab8a-e243-40a4-b8fc-b9c094339c78\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.478515 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.495199 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zczv\" (UniqueName: \"kubernetes.io/projected/8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab-kube-api-access-6zczv\") pod \"nova-scheduler-0\" (UID: \"8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.496347 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66040b18-6e3b-43dc-8d90-a3eeab8f31df-logs\") pod \"nova-api-0\" (UID: \"66040b18-6e3b-43dc-8d90-a3eeab8f31df\") " pod="openstack/nova-api-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.496388 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66040b18-6e3b-43dc-8d90-a3eeab8f31df-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"66040b18-6e3b-43dc-8d90-a3eeab8f31df\") " pod="openstack/nova-api-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.496417 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xdmn\" (UniqueName: \"kubernetes.io/projected/66040b18-6e3b-43dc-8d90-a3eeab8f31df-kube-api-access-4xdmn\") pod \"nova-api-0\" (UID: \"66040b18-6e3b-43dc-8d90-a3eeab8f31df\") " pod="openstack/nova-api-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.496516 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66040b18-6e3b-43dc-8d90-a3eeab8f31df-config-data\") pod \"nova-api-0\" (UID: \"66040b18-6e3b-43dc-8d90-a3eeab8f31df\") " pod="openstack/nova-api-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.501300 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjdr7\" (UniqueName: \"kubernetes.io/projected/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-kube-api-access-gjdr7\") pod \"nova-metadata-0\" (UID: \"6557fcf7-bd46-4552-bc2b-339ff9c9efbe\") " pod="openstack/nova-metadata-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.501741 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66040b18-6e3b-43dc-8d90-a3eeab8f31df-logs\") pod \"nova-api-0\" (UID: \"66040b18-6e3b-43dc-8d90-a3eeab8f31df\") " pod="openstack/nova-api-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.504192 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66040b18-6e3b-43dc-8d90-a3eeab8f31df-config-data\") pod \"nova-api-0\" (UID: \"66040b18-6e3b-43dc-8d90-a3eeab8f31df\") " pod="openstack/nova-api-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.508100 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-c562-account-create-update-gpcbx" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.513220 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66040b18-6e3b-43dc-8d90-a3eeab8f31df-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"66040b18-6e3b-43dc-8d90-a3eeab8f31df\") " pod="openstack/nova-api-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.526010 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.545163 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xdmn\" (UniqueName: \"kubernetes.io/projected/66040b18-6e3b-43dc-8d90-a3eeab8f31df-kube-api-access-4xdmn\") pod \"nova-api-0\" (UID: \"66040b18-6e3b-43dc-8d90-a3eeab8f31df\") " pod="openstack/nova-api-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.549587 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.576598 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.675329 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vc4ps" Nov 26 00:49:08 crc kubenswrapper[4766]: I1126 00:49:08.979331 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-gk5ww"] Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.160015 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-gk5ww" event={"ID":"8f161def-0ff3-4f5d-8f9f-083800b5e1f2","Type":"ContainerStarted","Data":"d9b76e4ef942537d4ddef9a947a26b42b2d705083aa17e4000b56291ead3c62e"} Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.203712 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.269066 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.351151 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-wxhtm"] Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.352547 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-wxhtm" Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.361504 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-wxhtm"] Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.363919 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.366977 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.439769 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12f40773-d3bf-46ac-96a0-8f114083229f-config-data\") pod \"nova-cell1-conductor-db-sync-wxhtm\" (UID: \"12f40773-d3bf-46ac-96a0-8f114083229f\") " pod="openstack/nova-cell1-conductor-db-sync-wxhtm" Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.440003 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12f40773-d3bf-46ac-96a0-8f114083229f-scripts\") pod \"nova-cell1-conductor-db-sync-wxhtm\" (UID: \"12f40773-d3bf-46ac-96a0-8f114083229f\") " pod="openstack/nova-cell1-conductor-db-sync-wxhtm" Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.440120 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp252\" (UniqueName: \"kubernetes.io/projected/12f40773-d3bf-46ac-96a0-8f114083229f-kube-api-access-cp252\") pod \"nova-cell1-conductor-db-sync-wxhtm\" (UID: \"12f40773-d3bf-46ac-96a0-8f114083229f\") " pod="openstack/nova-cell1-conductor-db-sync-wxhtm" Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.440249 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12f40773-d3bf-46ac-96a0-8f114083229f-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-wxhtm\" (UID: \"12f40773-d3bf-46ac-96a0-8f114083229f\") " pod="openstack/nova-cell1-conductor-db-sync-wxhtm" Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.541615 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12f40773-d3bf-46ac-96a0-8f114083229f-scripts\") pod \"nova-cell1-conductor-db-sync-wxhtm\" (UID: \"12f40773-d3bf-46ac-96a0-8f114083229f\") " pod="openstack/nova-cell1-conductor-db-sync-wxhtm" Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.541724 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp252\" (UniqueName: \"kubernetes.io/projected/12f40773-d3bf-46ac-96a0-8f114083229f-kube-api-access-cp252\") pod \"nova-cell1-conductor-db-sync-wxhtm\" (UID: \"12f40773-d3bf-46ac-96a0-8f114083229f\") " pod="openstack/nova-cell1-conductor-db-sync-wxhtm" Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.541774 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12f40773-d3bf-46ac-96a0-8f114083229f-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-wxhtm\" (UID: \"12f40773-d3bf-46ac-96a0-8f114083229f\") " pod="openstack/nova-cell1-conductor-db-sync-wxhtm" Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.541831 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12f40773-d3bf-46ac-96a0-8f114083229f-config-data\") pod \"nova-cell1-conductor-db-sync-wxhtm\" (UID: \"12f40773-d3bf-46ac-96a0-8f114083229f\") " pod="openstack/nova-cell1-conductor-db-sync-wxhtm" Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.547395 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12f40773-d3bf-46ac-96a0-8f114083229f-scripts\") pod \"nova-cell1-conductor-db-sync-wxhtm\" (UID: \"12f40773-d3bf-46ac-96a0-8f114083229f\") " pod="openstack/nova-cell1-conductor-db-sync-wxhtm" Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.547395 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12f40773-d3bf-46ac-96a0-8f114083229f-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-wxhtm\" (UID: \"12f40773-d3bf-46ac-96a0-8f114083229f\") " pod="openstack/nova-cell1-conductor-db-sync-wxhtm" Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.548032 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12f40773-d3bf-46ac-96a0-8f114083229f-config-data\") pod \"nova-cell1-conductor-db-sync-wxhtm\" (UID: \"12f40773-d3bf-46ac-96a0-8f114083229f\") " pod="openstack/nova-cell1-conductor-db-sync-wxhtm" Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.565863 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp252\" (UniqueName: \"kubernetes.io/projected/12f40773-d3bf-46ac-96a0-8f114083229f-kube-api-access-cp252\") pod \"nova-cell1-conductor-db-sync-wxhtm\" (UID: \"12f40773-d3bf-46ac-96a0-8f114083229f\") " pod="openstack/nova-cell1-conductor-db-sync-wxhtm" Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.672674 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.683211 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-wxhtm" Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.690668 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-c562-account-create-update-gpcbx"] Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.709452 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 00:49:09 crc kubenswrapper[4766]: W1126 00:49:09.734809 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod03e87e7f_6aa5_45f9_8e0e_acb7c0087f1c.slice/crio-fd2bcd3d7b5783e0595e8037a91a8485c1a7360e38e74d79c5d369aa93cd35bc WatchSource:0}: Error finding container fd2bcd3d7b5783e0595e8037a91a8485c1a7360e38e74d79c5d369aa93cd35bc: Status 404 returned error can't find the container with id fd2bcd3d7b5783e0595e8037a91a8485c1a7360e38e74d79c5d369aa93cd35bc Nov 26 00:49:09 crc kubenswrapper[4766]: I1126 00:49:09.735590 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vc4ps"] Nov 26 00:49:09 crc kubenswrapper[4766]: E1126 00:49:09.782266 4766 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f161def_0ff3_4f5d_8f9f_083800b5e1f2.slice/crio-conmon-01b645a0911c62a1c44fec9ad6c73b934f98107042782e07bb3bc616789e8a50.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f161def_0ff3_4f5d_8f9f_083800b5e1f2.slice/crio-01b645a0911c62a1c44fec9ad6c73b934f98107042782e07bb3bc616789e8a50.scope\": RecentStats: unable to find data in memory cache]" Nov 26 00:49:10 crc kubenswrapper[4766]: I1126 00:49:10.191423 4766 generic.go:334] "Generic (PLEG): container finished" podID="8f161def-0ff3-4f5d-8f9f-083800b5e1f2" containerID="01b645a0911c62a1c44fec9ad6c73b934f98107042782e07bb3bc616789e8a50" exitCode=0 Nov 26 00:49:10 crc kubenswrapper[4766]: I1126 00:49:10.191593 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-gk5ww" event={"ID":"8f161def-0ff3-4f5d-8f9f-083800b5e1f2","Type":"ContainerDied","Data":"01b645a0911c62a1c44fec9ad6c73b934f98107042782e07bb3bc616789e8a50"} Nov 26 00:49:10 crc kubenswrapper[4766]: I1126 00:49:10.197212 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-c562-account-create-update-gpcbx" event={"ID":"53093344-eff4-438d-8782-9d131efa8427","Type":"ContainerStarted","Data":"91430a52bc08672d7a76f2968ea638d9102ce29b7f28f477f8635795a2f46779"} Nov 26 00:49:10 crc kubenswrapper[4766]: I1126 00:49:10.197266 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-c562-account-create-update-gpcbx" event={"ID":"53093344-eff4-438d-8782-9d131efa8427","Type":"ContainerStarted","Data":"97641cffb38710baeabe89eded132625f92bf04b2c0692d7db83bc1f47ff1696"} Nov 26 00:49:10 crc kubenswrapper[4766]: I1126 00:49:10.205439 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab","Type":"ContainerStarted","Data":"689a90d00fef65123c6f346a7fdeaae541d711e63c450a3a218d0737bf367dc3"} Nov 26 00:49:10 crc kubenswrapper[4766]: I1126 00:49:10.218560 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6557fcf7-bd46-4552-bc2b-339ff9c9efbe","Type":"ContainerStarted","Data":"ff56ced604a636bed74357e9c1cd70b40b131d4d56f331d3ee2e1b1df06267da"} Nov 26 00:49:10 crc kubenswrapper[4766]: I1126 00:49:10.245342 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vc4ps" event={"ID":"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c","Type":"ContainerStarted","Data":"b1c01ee5e01c4926e7120c25299f00cb50055c6b3458163d33ae4cdbad953a9e"} Nov 26 00:49:10 crc kubenswrapper[4766]: I1126 00:49:10.245390 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vc4ps" event={"ID":"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c","Type":"ContainerStarted","Data":"fd2bcd3d7b5783e0595e8037a91a8485c1a7360e38e74d79c5d369aa93cd35bc"} Nov 26 00:49:10 crc kubenswrapper[4766]: I1126 00:49:10.245490 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-c562-account-create-update-gpcbx" podStartSLOduration=3.245465164 podStartE2EDuration="3.245465164s" podCreationTimestamp="2025-11-26 00:49:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:49:10.229806534 +0000 UTC m=+1531.078576964" watchObservedRunningTime="2025-11-26 00:49:10.245465164 +0000 UTC m=+1531.094235594" Nov 26 00:49:10 crc kubenswrapper[4766]: I1126 00:49:10.267330 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-vc4ps" podStartSLOduration=3.267313248 podStartE2EDuration="3.267313248s" podCreationTimestamp="2025-11-26 00:49:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:49:10.264783445 +0000 UTC m=+1531.113553875" watchObservedRunningTime="2025-11-26 00:49:10.267313248 +0000 UTC m=+1531.116083678" Nov 26 00:49:10 crc kubenswrapper[4766]: I1126 00:49:10.278026 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"66040b18-6e3b-43dc-8d90-a3eeab8f31df","Type":"ContainerStarted","Data":"57e358d9aab099e9ca0ab3973018cece1b324a254e3cd8f33a3d2b3de9899eef"} Nov 26 00:49:10 crc kubenswrapper[4766]: I1126 00:49:10.283764 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cfd3ab8a-e243-40a4-b8fc-b9c094339c78","Type":"ContainerStarted","Data":"94ab299f474cc0da858abfaab622396a22fb3bca09edcf9e3f83028b839af4ef"} Nov 26 00:49:10 crc kubenswrapper[4766]: I1126 00:49:10.362769 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-wxhtm"] Nov 26 00:49:11 crc kubenswrapper[4766]: I1126 00:49:11.307700 4766 generic.go:334] "Generic (PLEG): container finished" podID="53093344-eff4-438d-8782-9d131efa8427" containerID="91430a52bc08672d7a76f2968ea638d9102ce29b7f28f477f8635795a2f46779" exitCode=0 Nov 26 00:49:11 crc kubenswrapper[4766]: I1126 00:49:11.307976 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-c562-account-create-update-gpcbx" event={"ID":"53093344-eff4-438d-8782-9d131efa8427","Type":"ContainerDied","Data":"91430a52bc08672d7a76f2968ea638d9102ce29b7f28f477f8635795a2f46779"} Nov 26 00:49:11 crc kubenswrapper[4766]: I1126 00:49:11.314519 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-wxhtm" event={"ID":"12f40773-d3bf-46ac-96a0-8f114083229f","Type":"ContainerStarted","Data":"451e952ceb6097241b3aa93cc5e036cd73babfd8d613c4fc3109e5138908acde"} Nov 26 00:49:11 crc kubenswrapper[4766]: I1126 00:49:11.314610 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-wxhtm" event={"ID":"12f40773-d3bf-46ac-96a0-8f114083229f","Type":"ContainerStarted","Data":"07d02e15f360579d09049fea658de37783fefae75e9d7af416216d9364557cf4"} Nov 26 00:49:11 crc kubenswrapper[4766]: I1126 00:49:11.346475 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-wxhtm" podStartSLOduration=2.346453497 podStartE2EDuration="2.346453497s" podCreationTimestamp="2025-11-26 00:49:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:49:11.341074653 +0000 UTC m=+1532.189845093" watchObservedRunningTime="2025-11-26 00:49:11.346453497 +0000 UTC m=+1532.195223937" Nov 26 00:49:11 crc kubenswrapper[4766]: I1126 00:49:11.480144 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:49:11 crc kubenswrapper[4766]: I1126 00:49:11.480480 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:49:11 crc kubenswrapper[4766]: I1126 00:49:11.480529 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:49:11 crc kubenswrapper[4766]: I1126 00:49:11.481433 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 00:49:11 crc kubenswrapper[4766]: I1126 00:49:11.481487 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" gracePeriod=600 Nov 26 00:49:11 crc kubenswrapper[4766]: I1126 00:49:11.985368 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-gk5ww" Nov 26 00:49:12 crc kubenswrapper[4766]: I1126 00:49:12.103390 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvdmn\" (UniqueName: \"kubernetes.io/projected/8f161def-0ff3-4f5d-8f9f-083800b5e1f2-kube-api-access-rvdmn\") pod \"8f161def-0ff3-4f5d-8f9f-083800b5e1f2\" (UID: \"8f161def-0ff3-4f5d-8f9f-083800b5e1f2\") " Nov 26 00:49:12 crc kubenswrapper[4766]: I1126 00:49:12.103623 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f161def-0ff3-4f5d-8f9f-083800b5e1f2-operator-scripts\") pod \"8f161def-0ff3-4f5d-8f9f-083800b5e1f2\" (UID: \"8f161def-0ff3-4f5d-8f9f-083800b5e1f2\") " Nov 26 00:49:12 crc kubenswrapper[4766]: I1126 00:49:12.104416 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f161def-0ff3-4f5d-8f9f-083800b5e1f2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8f161def-0ff3-4f5d-8f9f-083800b5e1f2" (UID: "8f161def-0ff3-4f5d-8f9f-083800b5e1f2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:49:12 crc kubenswrapper[4766]: I1126 00:49:12.115343 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f161def-0ff3-4f5d-8f9f-083800b5e1f2-kube-api-access-rvdmn" (OuterVolumeSpecName: "kube-api-access-rvdmn") pod "8f161def-0ff3-4f5d-8f9f-083800b5e1f2" (UID: "8f161def-0ff3-4f5d-8f9f-083800b5e1f2"). InnerVolumeSpecName "kube-api-access-rvdmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:12 crc kubenswrapper[4766]: E1126 00:49:12.168762 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:49:12 crc kubenswrapper[4766]: I1126 00:49:12.209783 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvdmn\" (UniqueName: \"kubernetes.io/projected/8f161def-0ff3-4f5d-8f9f-083800b5e1f2-kube-api-access-rvdmn\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:12 crc kubenswrapper[4766]: I1126 00:49:12.209813 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f161def-0ff3-4f5d-8f9f-083800b5e1f2-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:12 crc kubenswrapper[4766]: I1126 00:49:12.343155 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-gk5ww" Nov 26 00:49:12 crc kubenswrapper[4766]: I1126 00:49:12.343201 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-gk5ww" event={"ID":"8f161def-0ff3-4f5d-8f9f-083800b5e1f2","Type":"ContainerDied","Data":"d9b76e4ef942537d4ddef9a947a26b42b2d705083aa17e4000b56291ead3c62e"} Nov 26 00:49:12 crc kubenswrapper[4766]: I1126 00:49:12.343257 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9b76e4ef942537d4ddef9a947a26b42b2d705083aa17e4000b56291ead3c62e" Nov 26 00:49:12 crc kubenswrapper[4766]: I1126 00:49:12.352802 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" exitCode=0 Nov 26 00:49:12 crc kubenswrapper[4766]: I1126 00:49:12.352959 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc"} Nov 26 00:49:12 crc kubenswrapper[4766]: I1126 00:49:12.353004 4766 scope.go:117] "RemoveContainer" containerID="8a2702c35239a61bcf6c33a8721b25d955ee16f1fd71838bddfceaf1a1a62d99" Nov 26 00:49:12 crc kubenswrapper[4766]: I1126 00:49:12.354211 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:49:12 crc kubenswrapper[4766]: E1126 00:49:12.354466 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.093694 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t7c6w"] Nov 26 00:49:13 crc kubenswrapper[4766]: E1126 00:49:13.094423 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f161def-0ff3-4f5d-8f9f-083800b5e1f2" containerName="mariadb-database-create" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.094440 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f161def-0ff3-4f5d-8f9f-083800b5e1f2" containerName="mariadb-database-create" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.094644 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f161def-0ff3-4f5d-8f9f-083800b5e1f2" containerName="mariadb-database-create" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.096387 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t7c6w" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.123590 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t7c6w"] Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.241432 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64e2c3e5-3cf6-4310-bc0b-325d569af577-utilities\") pod \"certified-operators-t7c6w\" (UID: \"64e2c3e5-3cf6-4310-bc0b-325d569af577\") " pod="openshift-marketplace/certified-operators-t7c6w" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.241475 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64e2c3e5-3cf6-4310-bc0b-325d569af577-catalog-content\") pod \"certified-operators-t7c6w\" (UID: \"64e2c3e5-3cf6-4310-bc0b-325d569af577\") " pod="openshift-marketplace/certified-operators-t7c6w" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.241603 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf7c2\" (UniqueName: \"kubernetes.io/projected/64e2c3e5-3cf6-4310-bc0b-325d569af577-kube-api-access-cf7c2\") pod \"certified-operators-t7c6w\" (UID: \"64e2c3e5-3cf6-4310-bc0b-325d569af577\") " pod="openshift-marketplace/certified-operators-t7c6w" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.343319 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64e2c3e5-3cf6-4310-bc0b-325d569af577-utilities\") pod \"certified-operators-t7c6w\" (UID: \"64e2c3e5-3cf6-4310-bc0b-325d569af577\") " pod="openshift-marketplace/certified-operators-t7c6w" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.343367 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64e2c3e5-3cf6-4310-bc0b-325d569af577-catalog-content\") pod \"certified-operators-t7c6w\" (UID: \"64e2c3e5-3cf6-4310-bc0b-325d569af577\") " pod="openshift-marketplace/certified-operators-t7c6w" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.343478 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf7c2\" (UniqueName: \"kubernetes.io/projected/64e2c3e5-3cf6-4310-bc0b-325d569af577-kube-api-access-cf7c2\") pod \"certified-operators-t7c6w\" (UID: \"64e2c3e5-3cf6-4310-bc0b-325d569af577\") " pod="openshift-marketplace/certified-operators-t7c6w" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.344171 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64e2c3e5-3cf6-4310-bc0b-325d569af577-utilities\") pod \"certified-operators-t7c6w\" (UID: \"64e2c3e5-3cf6-4310-bc0b-325d569af577\") " pod="openshift-marketplace/certified-operators-t7c6w" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.344379 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64e2c3e5-3cf6-4310-bc0b-325d569af577-catalog-content\") pod \"certified-operators-t7c6w\" (UID: \"64e2c3e5-3cf6-4310-bc0b-325d569af577\") " pod="openshift-marketplace/certified-operators-t7c6w" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.361430 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf7c2\" (UniqueName: \"kubernetes.io/projected/64e2c3e5-3cf6-4310-bc0b-325d569af577-kube-api-access-cf7c2\") pod \"certified-operators-t7c6w\" (UID: \"64e2c3e5-3cf6-4310-bc0b-325d569af577\") " pod="openshift-marketplace/certified-operators-t7c6w" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.375045 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-c562-account-create-update-gpcbx" event={"ID":"53093344-eff4-438d-8782-9d131efa8427","Type":"ContainerDied","Data":"97641cffb38710baeabe89eded132625f92bf04b2c0692d7db83bc1f47ff1696"} Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.375087 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97641cffb38710baeabe89eded132625f92bf04b2c0692d7db83bc1f47ff1696" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.427304 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t7c6w" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.463296 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-c562-account-create-update-gpcbx" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.549538 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53093344-eff4-438d-8782-9d131efa8427-operator-scripts\") pod \"53093344-eff4-438d-8782-9d131efa8427\" (UID: \"53093344-eff4-438d-8782-9d131efa8427\") " Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.549671 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9nnh\" (UniqueName: \"kubernetes.io/projected/53093344-eff4-438d-8782-9d131efa8427-kube-api-access-d9nnh\") pod \"53093344-eff4-438d-8782-9d131efa8427\" (UID: \"53093344-eff4-438d-8782-9d131efa8427\") " Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.554434 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53093344-eff4-438d-8782-9d131efa8427-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "53093344-eff4-438d-8782-9d131efa8427" (UID: "53093344-eff4-438d-8782-9d131efa8427"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.555422 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53093344-eff4-438d-8782-9d131efa8427-kube-api-access-d9nnh" (OuterVolumeSpecName: "kube-api-access-d9nnh") pod "53093344-eff4-438d-8782-9d131efa8427" (UID: "53093344-eff4-438d-8782-9d131efa8427"). InnerVolumeSpecName "kube-api-access-d9nnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.652185 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53093344-eff4-438d-8782-9d131efa8427-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.652445 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9nnh\" (UniqueName: \"kubernetes.io/projected/53093344-eff4-438d-8782-9d131efa8427-kube-api-access-d9nnh\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.857125 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sswvf" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.857206 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sswvf" Nov 26 00:49:13 crc kubenswrapper[4766]: I1126 00:49:13.920430 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sswvf" Nov 26 00:49:14 crc kubenswrapper[4766]: I1126 00:49:14.387516 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-c562-account-create-update-gpcbx" Nov 26 00:49:14 crc kubenswrapper[4766]: I1126 00:49:14.437220 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sswvf" Nov 26 00:49:15 crc kubenswrapper[4766]: I1126 00:49:15.413691 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t7c6w"] Nov 26 00:49:15 crc kubenswrapper[4766]: I1126 00:49:15.440585 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab","Type":"ContainerStarted","Data":"00f3b60244d89847b3a53fd77058301dc3ec73e6fc1ebe02e2044b53df8adf65"} Nov 26 00:49:15 crc kubenswrapper[4766]: I1126 00:49:15.453458 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6557fcf7-bd46-4552-bc2b-339ff9c9efbe","Type":"ContainerStarted","Data":"a045755472e90bfaa78d93793ed4e685edd333b3ab9ab8f6ebc16a200ca28ee7"} Nov 26 00:49:15 crc kubenswrapper[4766]: I1126 00:49:15.459730 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.266326894 podStartE2EDuration="7.459708533s" podCreationTimestamp="2025-11-26 00:49:08 +0000 UTC" firstStartedPulling="2025-11-26 00:49:09.700983238 +0000 UTC m=+1530.549753668" lastFinishedPulling="2025-11-26 00:49:14.894364877 +0000 UTC m=+1535.743135307" observedRunningTime="2025-11-26 00:49:15.458607066 +0000 UTC m=+1536.307377506" watchObservedRunningTime="2025-11-26 00:49:15.459708533 +0000 UTC m=+1536.308478973" Nov 26 00:49:15 crc kubenswrapper[4766]: I1126 00:49:15.467938 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"66040b18-6e3b-43dc-8d90-a3eeab8f31df","Type":"ContainerStarted","Data":"252626f0723e2a175f72d821ec6908dff31e1409f3ca28fd691e29d2869b2be6"} Nov 26 00:49:15 crc kubenswrapper[4766]: I1126 00:49:15.482719 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cfd3ab8a-e243-40a4-b8fc-b9c094339c78","Type":"ContainerStarted","Data":"2a91d5d53a9515d269adb9138b4124002864c8aa09d66aa9e808ff56cf6b1c34"} Nov 26 00:49:15 crc kubenswrapper[4766]: I1126 00:49:15.517088 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.830402103 podStartE2EDuration="7.517065983s" podCreationTimestamp="2025-11-26 00:49:08 +0000 UTC" firstStartedPulling="2025-11-26 00:49:09.214975638 +0000 UTC m=+1530.063746068" lastFinishedPulling="2025-11-26 00:49:14.901639518 +0000 UTC m=+1535.750409948" observedRunningTime="2025-11-26 00:49:15.512892288 +0000 UTC m=+1536.361662728" watchObservedRunningTime="2025-11-26 00:49:15.517065983 +0000 UTC m=+1536.365836413" Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.491198 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ljrdw"] Nov 26 00:49:16 crc kubenswrapper[4766]: E1126 00:49:16.492106 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53093344-eff4-438d-8782-9d131efa8427" containerName="mariadb-account-create-update" Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.492123 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="53093344-eff4-438d-8782-9d131efa8427" containerName="mariadb-account-create-update" Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.492395 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="53093344-eff4-438d-8782-9d131efa8427" containerName="mariadb-account-create-update" Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.522966 4766 generic.go:334] "Generic (PLEG): container finished" podID="64e2c3e5-3cf6-4310-bc0b-325d569af577" containerID="07e60ff7bb673aad0dcef7603fd849a25b95bda239d813b1124eeb14b684401d" exitCode=0 Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.537980 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7c6w" event={"ID":"64e2c3e5-3cf6-4310-bc0b-325d569af577","Type":"ContainerDied","Data":"07e60ff7bb673aad0dcef7603fd849a25b95bda239d813b1124eeb14b684401d"} Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.543106 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7c6w" event={"ID":"64e2c3e5-3cf6-4310-bc0b-325d569af577","Type":"ContainerStarted","Data":"b3efcc07eebad05c660c127eb6eb054d59cd79880cdb62cefb9709a63e13fdce"} Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.540532 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ljrdw" Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.554700 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6557fcf7-bd46-4552-bc2b-339ff9c9efbe","Type":"ContainerStarted","Data":"fee5153714804d61e61c312f61aacf0b71754de605ae56a39bcb1765403a2d88"} Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.560448 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ljrdw"] Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.566190 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"66040b18-6e3b-43dc-8d90-a3eeab8f31df","Type":"ContainerStarted","Data":"063c2b5642c0debff8bdd6d1fb44f28d827eb58ca69c0244076b2830713167a0"} Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.623035 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed38267b-18fd-4735-a51c-a09d735b756a-utilities\") pod \"redhat-marketplace-ljrdw\" (UID: \"ed38267b-18fd-4735-a51c-a09d735b756a\") " pod="openshift-marketplace/redhat-marketplace-ljrdw" Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.623142 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed38267b-18fd-4735-a51c-a09d735b756a-catalog-content\") pod \"redhat-marketplace-ljrdw\" (UID: \"ed38267b-18fd-4735-a51c-a09d735b756a\") " pod="openshift-marketplace/redhat-marketplace-ljrdw" Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.623248 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-695zw\" (UniqueName: \"kubernetes.io/projected/ed38267b-18fd-4735-a51c-a09d735b756a-kube-api-access-695zw\") pod \"redhat-marketplace-ljrdw\" (UID: \"ed38267b-18fd-4735-a51c-a09d735b756a\") " pod="openshift-marketplace/redhat-marketplace-ljrdw" Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.635686 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.025242342 podStartE2EDuration="8.635668963s" podCreationTimestamp="2025-11-26 00:49:08 +0000 UTC" firstStartedPulling="2025-11-26 00:49:09.288311125 +0000 UTC m=+1530.137081555" lastFinishedPulling="2025-11-26 00:49:14.898737746 +0000 UTC m=+1535.747508176" observedRunningTime="2025-11-26 00:49:16.598729803 +0000 UTC m=+1537.447500233" watchObservedRunningTime="2025-11-26 00:49:16.635668963 +0000 UTC m=+1537.484439393" Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.650519 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.43370411 podStartE2EDuration="8.650502213s" podCreationTimestamp="2025-11-26 00:49:08 +0000 UTC" firstStartedPulling="2025-11-26 00:49:09.68261373 +0000 UTC m=+1530.531384160" lastFinishedPulling="2025-11-26 00:49:14.899411833 +0000 UTC m=+1535.748182263" observedRunningTime="2025-11-26 00:49:16.617230354 +0000 UTC m=+1537.466000774" watchObservedRunningTime="2025-11-26 00:49:16.650502213 +0000 UTC m=+1537.499272643" Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.724918 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed38267b-18fd-4735-a51c-a09d735b756a-utilities\") pod \"redhat-marketplace-ljrdw\" (UID: \"ed38267b-18fd-4735-a51c-a09d735b756a\") " pod="openshift-marketplace/redhat-marketplace-ljrdw" Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.724994 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed38267b-18fd-4735-a51c-a09d735b756a-catalog-content\") pod \"redhat-marketplace-ljrdw\" (UID: \"ed38267b-18fd-4735-a51c-a09d735b756a\") " pod="openshift-marketplace/redhat-marketplace-ljrdw" Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.725078 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-695zw\" (UniqueName: \"kubernetes.io/projected/ed38267b-18fd-4735-a51c-a09d735b756a-kube-api-access-695zw\") pod \"redhat-marketplace-ljrdw\" (UID: \"ed38267b-18fd-4735-a51c-a09d735b756a\") " pod="openshift-marketplace/redhat-marketplace-ljrdw" Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.725443 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed38267b-18fd-4735-a51c-a09d735b756a-catalog-content\") pod \"redhat-marketplace-ljrdw\" (UID: \"ed38267b-18fd-4735-a51c-a09d735b756a\") " pod="openshift-marketplace/redhat-marketplace-ljrdw" Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.725756 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed38267b-18fd-4735-a51c-a09d735b756a-utilities\") pod \"redhat-marketplace-ljrdw\" (UID: \"ed38267b-18fd-4735-a51c-a09d735b756a\") " pod="openshift-marketplace/redhat-marketplace-ljrdw" Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.748874 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-695zw\" (UniqueName: \"kubernetes.io/projected/ed38267b-18fd-4735-a51c-a09d735b756a-kube-api-access-695zw\") pod \"redhat-marketplace-ljrdw\" (UID: \"ed38267b-18fd-4735-a51c-a09d735b756a\") " pod="openshift-marketplace/redhat-marketplace-ljrdw" Nov 26 00:49:16 crc kubenswrapper[4766]: I1126 00:49:16.877916 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ljrdw" Nov 26 00:49:17 crc kubenswrapper[4766]: I1126 00:49:17.513100 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ljrdw"] Nov 26 00:49:17 crc kubenswrapper[4766]: W1126 00:49:17.521623 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded38267b_18fd_4735_a51c_a09d735b756a.slice/crio-198720d9532ee700314c562036153a48578986b18764731e1bd82031317ab36d WatchSource:0}: Error finding container 198720d9532ee700314c562036153a48578986b18764731e1bd82031317ab36d: Status 404 returned error can't find the container with id 198720d9532ee700314c562036153a48578986b18764731e1bd82031317ab36d Nov 26 00:49:17 crc kubenswrapper[4766]: I1126 00:49:17.585307 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ljrdw" event={"ID":"ed38267b-18fd-4735-a51c-a09d735b756a","Type":"ContainerStarted","Data":"198720d9532ee700314c562036153a48578986b18764731e1bd82031317ab36d"} Nov 26 00:49:17 crc kubenswrapper[4766]: I1126 00:49:17.597767 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7c6w" event={"ID":"64e2c3e5-3cf6-4310-bc0b-325d569af577","Type":"ContainerStarted","Data":"af9cae77ddaefbb2f49db9dd147e090e5b53033753a7c7d5d83d725142424bcb"} Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.430054 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-mfr52"] Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.431689 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-mfr52" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.433925 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-9jvhv" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.434790 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.435100 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.435276 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.450836 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-mfr52"] Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.460317 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l486f\" (UniqueName: \"kubernetes.io/projected/638a59e9-68c3-43cf-ae3b-4510e7df42ec-kube-api-access-l486f\") pod \"aodh-db-sync-mfr52\" (UID: \"638a59e9-68c3-43cf-ae3b-4510e7df42ec\") " pod="openstack/aodh-db-sync-mfr52" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.460447 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/638a59e9-68c3-43cf-ae3b-4510e7df42ec-config-data\") pod \"aodh-db-sync-mfr52\" (UID: \"638a59e9-68c3-43cf-ae3b-4510e7df42ec\") " pod="openstack/aodh-db-sync-mfr52" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.460482 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/638a59e9-68c3-43cf-ae3b-4510e7df42ec-scripts\") pod \"aodh-db-sync-mfr52\" (UID: \"638a59e9-68c3-43cf-ae3b-4510e7df42ec\") " pod="openstack/aodh-db-sync-mfr52" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.460622 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/638a59e9-68c3-43cf-ae3b-4510e7df42ec-combined-ca-bundle\") pod \"aodh-db-sync-mfr52\" (UID: \"638a59e9-68c3-43cf-ae3b-4510e7df42ec\") " pod="openstack/aodh-db-sync-mfr52" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.479992 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.480039 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.496722 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mbqtt"] Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.499084 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mbqtt" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.522515 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mbqtt"] Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.527077 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.527113 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.527126 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.527140 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.550844 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.550884 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.562772 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed786021-3dca-48e5-b286-f1a06dcbe58e-catalog-content\") pod \"redhat-operators-mbqtt\" (UID: \"ed786021-3dca-48e5-b286-f1a06dcbe58e\") " pod="openshift-marketplace/redhat-operators-mbqtt" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.562838 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l486f\" (UniqueName: \"kubernetes.io/projected/638a59e9-68c3-43cf-ae3b-4510e7df42ec-kube-api-access-l486f\") pod \"aodh-db-sync-mfr52\" (UID: \"638a59e9-68c3-43cf-ae3b-4510e7df42ec\") " pod="openstack/aodh-db-sync-mfr52" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.562952 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/638a59e9-68c3-43cf-ae3b-4510e7df42ec-config-data\") pod \"aodh-db-sync-mfr52\" (UID: \"638a59e9-68c3-43cf-ae3b-4510e7df42ec\") " pod="openstack/aodh-db-sync-mfr52" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.562986 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/638a59e9-68c3-43cf-ae3b-4510e7df42ec-scripts\") pod \"aodh-db-sync-mfr52\" (UID: \"638a59e9-68c3-43cf-ae3b-4510e7df42ec\") " pod="openstack/aodh-db-sync-mfr52" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.563008 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed786021-3dca-48e5-b286-f1a06dcbe58e-utilities\") pod \"redhat-operators-mbqtt\" (UID: \"ed786021-3dca-48e5-b286-f1a06dcbe58e\") " pod="openshift-marketplace/redhat-operators-mbqtt" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.563035 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vt5g\" (UniqueName: \"kubernetes.io/projected/ed786021-3dca-48e5-b286-f1a06dcbe58e-kube-api-access-7vt5g\") pod \"redhat-operators-mbqtt\" (UID: \"ed786021-3dca-48e5-b286-f1a06dcbe58e\") " pod="openshift-marketplace/redhat-operators-mbqtt" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.563101 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/638a59e9-68c3-43cf-ae3b-4510e7df42ec-combined-ca-bundle\") pod \"aodh-db-sync-mfr52\" (UID: \"638a59e9-68c3-43cf-ae3b-4510e7df42ec\") " pod="openstack/aodh-db-sync-mfr52" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.573712 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/638a59e9-68c3-43cf-ae3b-4510e7df42ec-combined-ca-bundle\") pod \"aodh-db-sync-mfr52\" (UID: \"638a59e9-68c3-43cf-ae3b-4510e7df42ec\") " pod="openstack/aodh-db-sync-mfr52" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.577346 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/638a59e9-68c3-43cf-ae3b-4510e7df42ec-scripts\") pod \"aodh-db-sync-mfr52\" (UID: \"638a59e9-68c3-43cf-ae3b-4510e7df42ec\") " pod="openstack/aodh-db-sync-mfr52" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.579466 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.579504 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.583216 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/638a59e9-68c3-43cf-ae3b-4510e7df42ec-config-data\") pod \"aodh-db-sync-mfr52\" (UID: \"638a59e9-68c3-43cf-ae3b-4510e7df42ec\") " pod="openstack/aodh-db-sync-mfr52" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.592189 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l486f\" (UniqueName: \"kubernetes.io/projected/638a59e9-68c3-43cf-ae3b-4510e7df42ec-kube-api-access-l486f\") pod \"aodh-db-sync-mfr52\" (UID: \"638a59e9-68c3-43cf-ae3b-4510e7df42ec\") " pod="openstack/aodh-db-sync-mfr52" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.625760 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.649891 4766 generic.go:334] "Generic (PLEG): container finished" podID="64e2c3e5-3cf6-4310-bc0b-325d569af577" containerID="af9cae77ddaefbb2f49db9dd147e090e5b53033753a7c7d5d83d725142424bcb" exitCode=0 Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.650000 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7c6w" event={"ID":"64e2c3e5-3cf6-4310-bc0b-325d569af577","Type":"ContainerDied","Data":"af9cae77ddaefbb2f49db9dd147e090e5b53033753a7c7d5d83d725142424bcb"} Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.660852 4766 generic.go:334] "Generic (PLEG): container finished" podID="ed38267b-18fd-4735-a51c-a09d735b756a" containerID="d9cd1a627d0d5e8681c265f4866275225a4a95c3fc9fe5d42d7090d131edb692" exitCode=0 Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.661052 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ljrdw" event={"ID":"ed38267b-18fd-4735-a51c-a09d735b756a","Type":"ContainerDied","Data":"d9cd1a627d0d5e8681c265f4866275225a4a95c3fc9fe5d42d7090d131edb692"} Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.673319 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed786021-3dca-48e5-b286-f1a06dcbe58e-utilities\") pod \"redhat-operators-mbqtt\" (UID: \"ed786021-3dca-48e5-b286-f1a06dcbe58e\") " pod="openshift-marketplace/redhat-operators-mbqtt" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.673360 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vt5g\" (UniqueName: \"kubernetes.io/projected/ed786021-3dca-48e5-b286-f1a06dcbe58e-kube-api-access-7vt5g\") pod \"redhat-operators-mbqtt\" (UID: \"ed786021-3dca-48e5-b286-f1a06dcbe58e\") " pod="openshift-marketplace/redhat-operators-mbqtt" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.673509 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed786021-3dca-48e5-b286-f1a06dcbe58e-catalog-content\") pod \"redhat-operators-mbqtt\" (UID: \"ed786021-3dca-48e5-b286-f1a06dcbe58e\") " pod="openshift-marketplace/redhat-operators-mbqtt" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.673964 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed786021-3dca-48e5-b286-f1a06dcbe58e-utilities\") pod \"redhat-operators-mbqtt\" (UID: \"ed786021-3dca-48e5-b286-f1a06dcbe58e\") " pod="openshift-marketplace/redhat-operators-mbqtt" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.674187 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed786021-3dca-48e5-b286-f1a06dcbe58e-catalog-content\") pod \"redhat-operators-mbqtt\" (UID: \"ed786021-3dca-48e5-b286-f1a06dcbe58e\") " pod="openshift-marketplace/redhat-operators-mbqtt" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.730863 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vt5g\" (UniqueName: \"kubernetes.io/projected/ed786021-3dca-48e5-b286-f1a06dcbe58e-kube-api-access-7vt5g\") pod \"redhat-operators-mbqtt\" (UID: \"ed786021-3dca-48e5-b286-f1a06dcbe58e\") " pod="openshift-marketplace/redhat-operators-mbqtt" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.737465 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.757053 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-mfr52" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.806846 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.831470 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mbqtt" Nov 26 00:49:18 crc kubenswrapper[4766]: I1126 00:49:18.919952 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:19 crc kubenswrapper[4766]: I1126 00:49:19.585035 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mbqtt"] Nov 26 00:49:19 crc kubenswrapper[4766]: I1126 00:49:19.612867 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6557fcf7-bd46-4552-bc2b-339ff9c9efbe" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.0.224:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 26 00:49:19 crc kubenswrapper[4766]: I1126 00:49:19.613465 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6557fcf7-bd46-4552-bc2b-339ff9c9efbe" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.0.224:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 26 00:49:19 crc kubenswrapper[4766]: I1126 00:49:19.692999 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mbqtt" event={"ID":"ed786021-3dca-48e5-b286-f1a06dcbe58e","Type":"ContainerStarted","Data":"c1bda5bbfb1bb6ee9188e20ec7d5e5a7f1a551c686d571397074f7b30801ddcd"} Nov 26 00:49:19 crc kubenswrapper[4766]: I1126 00:49:19.698007 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="66040b18-6e3b-43dc-8d90-a3eeab8f31df" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.226:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 26 00:49:19 crc kubenswrapper[4766]: I1126 00:49:19.698808 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="66040b18-6e3b-43dc-8d90-a3eeab8f31df" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.226:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 26 00:49:19 crc kubenswrapper[4766]: I1126 00:49:19.699239 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7c6w" event={"ID":"64e2c3e5-3cf6-4310-bc0b-325d569af577","Type":"ContainerStarted","Data":"a57ca2c45ee39ed364cb91e5ea745c245dcfc3c7b4c1bd7d7893717ba99a5ea0"} Nov 26 00:49:19 crc kubenswrapper[4766]: I1126 00:49:19.728266 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t7c6w" podStartSLOduration=4.12718653 podStartE2EDuration="6.728249709s" podCreationTimestamp="2025-11-26 00:49:13 +0000 UTC" firstStartedPulling="2025-11-26 00:49:16.528097863 +0000 UTC m=+1537.376868293" lastFinishedPulling="2025-11-26 00:49:19.129161042 +0000 UTC m=+1539.977931472" observedRunningTime="2025-11-26 00:49:19.727559202 +0000 UTC m=+1540.576329632" watchObservedRunningTime="2025-11-26 00:49:19.728249709 +0000 UTC m=+1540.577020139" Nov 26 00:49:19 crc kubenswrapper[4766]: I1126 00:49:19.794757 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-mfr52"] Nov 26 00:49:20 crc kubenswrapper[4766]: I1126 00:49:20.759499 4766 generic.go:334] "Generic (PLEG): container finished" podID="ed786021-3dca-48e5-b286-f1a06dcbe58e" containerID="efcf5a32b9c10dadb05b1e9595a2e97f628ee328999c92b6bebd546967f968f8" exitCode=0 Nov 26 00:49:20 crc kubenswrapper[4766]: I1126 00:49:20.759755 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mbqtt" event={"ID":"ed786021-3dca-48e5-b286-f1a06dcbe58e","Type":"ContainerDied","Data":"efcf5a32b9c10dadb05b1e9595a2e97f628ee328999c92b6bebd546967f968f8"} Nov 26 00:49:20 crc kubenswrapper[4766]: I1126 00:49:20.772455 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-mfr52" event={"ID":"638a59e9-68c3-43cf-ae3b-4510e7df42ec","Type":"ContainerStarted","Data":"7fb6d2a387f1b4b8391caf5d7fe480d89f001fb0cb3fb6782f708142812f530f"} Nov 26 00:49:20 crc kubenswrapper[4766]: I1126 00:49:20.774359 4766 generic.go:334] "Generic (PLEG): container finished" podID="ed38267b-18fd-4735-a51c-a09d735b756a" containerID="0aca943ab636156e086d709729b9f4b92a4f5a345b5a80bfa04e114a69d02009" exitCode=0 Nov 26 00:49:20 crc kubenswrapper[4766]: I1126 00:49:20.775877 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ljrdw" event={"ID":"ed38267b-18fd-4735-a51c-a09d735b756a","Type":"ContainerDied","Data":"0aca943ab636156e086d709729b9f4b92a4f5a345b5a80bfa04e114a69d02009"} Nov 26 00:49:22 crc kubenswrapper[4766]: I1126 00:49:22.290565 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fr5xr"] Nov 26 00:49:22 crc kubenswrapper[4766]: I1126 00:49:22.293463 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fr5xr" Nov 26 00:49:22 crc kubenswrapper[4766]: I1126 00:49:22.349795 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fr5xr"] Nov 26 00:49:22 crc kubenswrapper[4766]: I1126 00:49:22.406021 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/580ecfc1-6334-44d3-a833-a68f98364860-utilities\") pod \"certified-operators-fr5xr\" (UID: \"580ecfc1-6334-44d3-a833-a68f98364860\") " pod="openshift-marketplace/certified-operators-fr5xr" Nov 26 00:49:22 crc kubenswrapper[4766]: I1126 00:49:22.406099 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/580ecfc1-6334-44d3-a833-a68f98364860-catalog-content\") pod \"certified-operators-fr5xr\" (UID: \"580ecfc1-6334-44d3-a833-a68f98364860\") " pod="openshift-marketplace/certified-operators-fr5xr" Nov 26 00:49:22 crc kubenswrapper[4766]: I1126 00:49:22.406206 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkmdw\" (UniqueName: \"kubernetes.io/projected/580ecfc1-6334-44d3-a833-a68f98364860-kube-api-access-nkmdw\") pod \"certified-operators-fr5xr\" (UID: \"580ecfc1-6334-44d3-a833-a68f98364860\") " pod="openshift-marketplace/certified-operators-fr5xr" Nov 26 00:49:22 crc kubenswrapper[4766]: I1126 00:49:22.507858 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/580ecfc1-6334-44d3-a833-a68f98364860-catalog-content\") pod \"certified-operators-fr5xr\" (UID: \"580ecfc1-6334-44d3-a833-a68f98364860\") " pod="openshift-marketplace/certified-operators-fr5xr" Nov 26 00:49:22 crc kubenswrapper[4766]: I1126 00:49:22.508280 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/580ecfc1-6334-44d3-a833-a68f98364860-catalog-content\") pod \"certified-operators-fr5xr\" (UID: \"580ecfc1-6334-44d3-a833-a68f98364860\") " pod="openshift-marketplace/certified-operators-fr5xr" Nov 26 00:49:22 crc kubenswrapper[4766]: I1126 00:49:22.508480 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkmdw\" (UniqueName: \"kubernetes.io/projected/580ecfc1-6334-44d3-a833-a68f98364860-kube-api-access-nkmdw\") pod \"certified-operators-fr5xr\" (UID: \"580ecfc1-6334-44d3-a833-a68f98364860\") " pod="openshift-marketplace/certified-operators-fr5xr" Nov 26 00:49:22 crc kubenswrapper[4766]: I1126 00:49:22.508912 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/580ecfc1-6334-44d3-a833-a68f98364860-utilities\") pod \"certified-operators-fr5xr\" (UID: \"580ecfc1-6334-44d3-a833-a68f98364860\") " pod="openshift-marketplace/certified-operators-fr5xr" Nov 26 00:49:22 crc kubenswrapper[4766]: I1126 00:49:22.509265 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/580ecfc1-6334-44d3-a833-a68f98364860-utilities\") pod \"certified-operators-fr5xr\" (UID: \"580ecfc1-6334-44d3-a833-a68f98364860\") " pod="openshift-marketplace/certified-operators-fr5xr" Nov 26 00:49:22 crc kubenswrapper[4766]: I1126 00:49:22.548582 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkmdw\" (UniqueName: \"kubernetes.io/projected/580ecfc1-6334-44d3-a833-a68f98364860-kube-api-access-nkmdw\") pod \"certified-operators-fr5xr\" (UID: \"580ecfc1-6334-44d3-a833-a68f98364860\") " pod="openshift-marketplace/certified-operators-fr5xr" Nov 26 00:49:22 crc kubenswrapper[4766]: I1126 00:49:22.643051 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fr5xr" Nov 26 00:49:23 crc kubenswrapper[4766]: I1126 00:49:23.427886 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t7c6w" Nov 26 00:49:23 crc kubenswrapper[4766]: I1126 00:49:23.429315 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t7c6w" Nov 26 00:49:23 crc kubenswrapper[4766]: I1126 00:49:23.492482 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t7c6w" Nov 26 00:49:23 crc kubenswrapper[4766]: I1126 00:49:23.567812 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fr5xr"] Nov 26 00:49:23 crc kubenswrapper[4766]: W1126 00:49:23.582250 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod580ecfc1_6334_44d3_a833_a68f98364860.slice/crio-c856bccb94c734e57f276e5e6e355d052046c97689085e1930b4faf0f57e8104 WatchSource:0}: Error finding container c856bccb94c734e57f276e5e6e355d052046c97689085e1930b4faf0f57e8104: Status 404 returned error can't find the container with id c856bccb94c734e57f276e5e6e355d052046c97689085e1930b4faf0f57e8104 Nov 26 00:49:23 crc kubenswrapper[4766]: I1126 00:49:23.849210 4766 generic.go:334] "Generic (PLEG): container finished" podID="12f40773-d3bf-46ac-96a0-8f114083229f" containerID="451e952ceb6097241b3aa93cc5e036cd73babfd8d613c4fc3109e5138908acde" exitCode=0 Nov 26 00:49:23 crc kubenswrapper[4766]: I1126 00:49:23.849274 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-wxhtm" event={"ID":"12f40773-d3bf-46ac-96a0-8f114083229f","Type":"ContainerDied","Data":"451e952ceb6097241b3aa93cc5e036cd73babfd8d613c4fc3109e5138908acde"} Nov 26 00:49:23 crc kubenswrapper[4766]: I1126 00:49:23.864630 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ljrdw" event={"ID":"ed38267b-18fd-4735-a51c-a09d735b756a","Type":"ContainerStarted","Data":"297475e998d0cc5924523d72b30842ab3c6c55e3f765a8e00dcfdc69a8da1c67"} Nov 26 00:49:23 crc kubenswrapper[4766]: I1126 00:49:23.869979 4766 generic.go:334] "Generic (PLEG): container finished" podID="03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c" containerID="b1c01ee5e01c4926e7120c25299f00cb50055c6b3458163d33ae4cdbad953a9e" exitCode=0 Nov 26 00:49:23 crc kubenswrapper[4766]: I1126 00:49:23.870055 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vc4ps" event={"ID":"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c","Type":"ContainerDied","Data":"b1c01ee5e01c4926e7120c25299f00cb50055c6b3458163d33ae4cdbad953a9e"} Nov 26 00:49:23 crc kubenswrapper[4766]: I1126 00:49:23.882979 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fr5xr" event={"ID":"580ecfc1-6334-44d3-a833-a68f98364860","Type":"ContainerStarted","Data":"c856bccb94c734e57f276e5e6e355d052046c97689085e1930b4faf0f57e8104"} Nov 26 00:49:23 crc kubenswrapper[4766]: I1126 00:49:23.888204 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ljrdw" podStartSLOduration=4.115712433 podStartE2EDuration="7.888185048s" podCreationTimestamp="2025-11-26 00:49:16 +0000 UTC" firstStartedPulling="2025-11-26 00:49:18.673108219 +0000 UTC m=+1539.521878649" lastFinishedPulling="2025-11-26 00:49:22.445580834 +0000 UTC m=+1543.294351264" observedRunningTime="2025-11-26 00:49:23.883122772 +0000 UTC m=+1544.731893202" watchObservedRunningTime="2025-11-26 00:49:23.888185048 +0000 UTC m=+1544.736955478" Nov 26 00:49:23 crc kubenswrapper[4766]: I1126 00:49:23.891397 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mbqtt" event={"ID":"ed786021-3dca-48e5-b286-f1a06dcbe58e","Type":"ContainerStarted","Data":"d2023a97f41b4d03c09189daceb3da6384094f0ab8c9cc4b5d6dfb213e0b6f8d"} Nov 26 00:49:24 crc kubenswrapper[4766]: I1126 00:49:24.924029 4766 generic.go:334] "Generic (PLEG): container finished" podID="580ecfc1-6334-44d3-a833-a68f98364860" containerID="33bc57a0146c86e26237df37029ee1ba4ecb162095aa4abec2eaa16321318b07" exitCode=0 Nov 26 00:49:24 crc kubenswrapper[4766]: I1126 00:49:24.926006 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fr5xr" event={"ID":"580ecfc1-6334-44d3-a833-a68f98364860","Type":"ContainerDied","Data":"33bc57a0146c86e26237df37029ee1ba4ecb162095aa4abec2eaa16321318b07"} Nov 26 00:49:25 crc kubenswrapper[4766]: I1126 00:49:25.029532 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 00:49:25 crc kubenswrapper[4766]: I1126 00:49:25.030080 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6557fcf7-bd46-4552-bc2b-339ff9c9efbe" containerName="nova-metadata-log" containerID="cri-o://a045755472e90bfaa78d93793ed4e685edd333b3ab9ab8f6ebc16a200ca28ee7" gracePeriod=30 Nov 26 00:49:25 crc kubenswrapper[4766]: I1126 00:49:25.030607 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6557fcf7-bd46-4552-bc2b-339ff9c9efbe" containerName="nova-metadata-metadata" containerID="cri-o://fee5153714804d61e61c312f61aacf0b71754de605ae56a39bcb1765403a2d88" gracePeriod=30 Nov 26 00:49:25 crc kubenswrapper[4766]: I1126 00:49:25.044299 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 26 00:49:25 crc kubenswrapper[4766]: I1126 00:49:25.044577 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="cfd3ab8a-e243-40a4-b8fc-b9c094339c78" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://2a91d5d53a9515d269adb9138b4124002864c8aa09d66aa9e808ff56cf6b1c34" gracePeriod=30 Nov 26 00:49:25 crc kubenswrapper[4766]: I1126 00:49:25.051201 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t7c6w" Nov 26 00:49:26 crc kubenswrapper[4766]: I1126 00:49:25.832411 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:49:26 crc kubenswrapper[4766]: E1126 00:49:25.834446 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:49:26 crc kubenswrapper[4766]: I1126 00:49:25.938188 4766 generic.go:334] "Generic (PLEG): container finished" podID="cfd3ab8a-e243-40a4-b8fc-b9c094339c78" containerID="2a91d5d53a9515d269adb9138b4124002864c8aa09d66aa9e808ff56cf6b1c34" exitCode=0 Nov 26 00:49:26 crc kubenswrapper[4766]: I1126 00:49:25.938253 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cfd3ab8a-e243-40a4-b8fc-b9c094339c78","Type":"ContainerDied","Data":"2a91d5d53a9515d269adb9138b4124002864c8aa09d66aa9e808ff56cf6b1c34"} Nov 26 00:49:26 crc kubenswrapper[4766]: I1126 00:49:25.940359 4766 generic.go:334] "Generic (PLEG): container finished" podID="6557fcf7-bd46-4552-bc2b-339ff9c9efbe" containerID="a045755472e90bfaa78d93793ed4e685edd333b3ab9ab8f6ebc16a200ca28ee7" exitCode=143 Nov 26 00:49:26 crc kubenswrapper[4766]: I1126 00:49:25.941521 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6557fcf7-bd46-4552-bc2b-339ff9c9efbe","Type":"ContainerDied","Data":"a045755472e90bfaa78d93793ed4e685edd333b3ab9ab8f6ebc16a200ca28ee7"} Nov 26 00:49:26 crc kubenswrapper[4766]: I1126 00:49:26.878287 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ljrdw" Nov 26 00:49:26 crc kubenswrapper[4766]: I1126 00:49:26.878536 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ljrdw" Nov 26 00:49:26 crc kubenswrapper[4766]: I1126 00:49:26.966735 4766 generic.go:334] "Generic (PLEG): container finished" podID="ed786021-3dca-48e5-b286-f1a06dcbe58e" containerID="d2023a97f41b4d03c09189daceb3da6384094f0ab8c9cc4b5d6dfb213e0b6f8d" exitCode=0 Nov 26 00:49:26 crc kubenswrapper[4766]: I1126 00:49:26.966779 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mbqtt" event={"ID":"ed786021-3dca-48e5-b286-f1a06dcbe58e","Type":"ContainerDied","Data":"d2023a97f41b4d03c09189daceb3da6384094f0ab8c9cc4b5d6dfb213e0b6f8d"} Nov 26 00:49:27 crc kubenswrapper[4766]: I1126 00:49:27.940103 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-ljrdw" podUID="ed38267b-18fd-4735-a51c-a09d735b756a" containerName="registry-server" probeResult="failure" output=< Nov 26 00:49:27 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 00:49:27 crc kubenswrapper[4766]: > Nov 26 00:49:28 crc kubenswrapper[4766]: I1126 00:49:28.480489 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="cfd3ab8a-e243-40a4-b8fc-b9c094339c78" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.0.223:6080/vnc_lite.html\": dial tcp 10.217.0.223:6080: connect: connection refused" Nov 26 00:49:28 crc kubenswrapper[4766]: I1126 00:49:28.691356 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 26 00:49:28 crc kubenswrapper[4766]: I1126 00:49:28.693180 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 26 00:49:28 crc kubenswrapper[4766]: I1126 00:49:28.696161 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 26 00:49:28 crc kubenswrapper[4766]: I1126 00:49:28.696338 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 26 00:49:28 crc kubenswrapper[4766]: I1126 00:49:28.991774 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 26 00:49:28 crc kubenswrapper[4766]: I1126 00:49:28.992175 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" containerName="kube-state-metrics" containerID="cri-o://3a54ae3d324cd23367e4b743604acd325d6f1f79b6e146944b0f49878887b855" gracePeriod=30 Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:28.997737 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vc4ps" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.022824 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-wxhtm" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.066454 4766 generic.go:334] "Generic (PLEG): container finished" podID="6557fcf7-bd46-4552-bc2b-339ff9c9efbe" containerID="fee5153714804d61e61c312f61aacf0b71754de605ae56a39bcb1765403a2d88" exitCode=0 Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.066577 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6557fcf7-bd46-4552-bc2b-339ff9c9efbe","Type":"ContainerDied","Data":"fee5153714804d61e61c312f61aacf0b71754de605ae56a39bcb1765403a2d88"} Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.085028 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vc4ps" event={"ID":"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c","Type":"ContainerDied","Data":"fd2bcd3d7b5783e0595e8037a91a8485c1a7360e38e74d79c5d369aa93cd35bc"} Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.085072 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd2bcd3d7b5783e0595e8037a91a8485c1a7360e38e74d79c5d369aa93cd35bc" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.085127 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vc4ps" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.090996 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-wxhtm" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.091525 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-wxhtm" event={"ID":"12f40773-d3bf-46ac-96a0-8f114083229f","Type":"ContainerDied","Data":"07d02e15f360579d09049fea658de37783fefae75e9d7af416216d9364557cf4"} Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.091554 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07d02e15f360579d09049fea658de37783fefae75e9d7af416216d9364557cf4" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.091573 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.099854 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12f40773-d3bf-46ac-96a0-8f114083229f-config-data\") pod \"12f40773-d3bf-46ac-96a0-8f114083229f\" (UID: \"12f40773-d3bf-46ac-96a0-8f114083229f\") " Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.100006 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12f40773-d3bf-46ac-96a0-8f114083229f-combined-ca-bundle\") pod \"12f40773-d3bf-46ac-96a0-8f114083229f\" (UID: \"12f40773-d3bf-46ac-96a0-8f114083229f\") " Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.100061 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-scripts\") pod \"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c\" (UID: \"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c\") " Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.100134 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pslsg\" (UniqueName: \"kubernetes.io/projected/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-kube-api-access-pslsg\") pod \"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c\" (UID: \"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c\") " Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.100163 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-config-data\") pod \"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c\" (UID: \"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c\") " Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.100241 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12f40773-d3bf-46ac-96a0-8f114083229f-scripts\") pod \"12f40773-d3bf-46ac-96a0-8f114083229f\" (UID: \"12f40773-d3bf-46ac-96a0-8f114083229f\") " Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.100268 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cp252\" (UniqueName: \"kubernetes.io/projected/12f40773-d3bf-46ac-96a0-8f114083229f-kube-api-access-cp252\") pod \"12f40773-d3bf-46ac-96a0-8f114083229f\" (UID: \"12f40773-d3bf-46ac-96a0-8f114083229f\") " Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.100290 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-combined-ca-bundle\") pod \"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c\" (UID: \"03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c\") " Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.107602 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-kube-api-access-pslsg" (OuterVolumeSpecName: "kube-api-access-pslsg") pod "03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c" (UID: "03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c"). InnerVolumeSpecName "kube-api-access-pslsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.110101 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12f40773-d3bf-46ac-96a0-8f114083229f-scripts" (OuterVolumeSpecName: "scripts") pod "12f40773-d3bf-46ac-96a0-8f114083229f" (UID: "12f40773-d3bf-46ac-96a0-8f114083229f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.112938 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12f40773-d3bf-46ac-96a0-8f114083229f-kube-api-access-cp252" (OuterVolumeSpecName: "kube-api-access-cp252") pod "12f40773-d3bf-46ac-96a0-8f114083229f" (UID: "12f40773-d3bf-46ac-96a0-8f114083229f"). InnerVolumeSpecName "kube-api-access-cp252". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.113046 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.119151 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-scripts" (OuterVolumeSpecName: "scripts") pod "03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c" (UID: "03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.157308 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-config-data" (OuterVolumeSpecName: "config-data") pod "03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c" (UID: "03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.161019 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c" (UID: "03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.205119 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.205144 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pslsg\" (UniqueName: \"kubernetes.io/projected/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-kube-api-access-pslsg\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.205154 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.205162 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12f40773-d3bf-46ac-96a0-8f114083229f-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.205170 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cp252\" (UniqueName: \"kubernetes.io/projected/12f40773-d3bf-46ac-96a0-8f114083229f-kube-api-access-cp252\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.205178 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.224542 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12f40773-d3bf-46ac-96a0-8f114083229f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12f40773-d3bf-46ac-96a0-8f114083229f" (UID: "12f40773-d3bf-46ac-96a0-8f114083229f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.234336 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12f40773-d3bf-46ac-96a0-8f114083229f-config-data" (OuterVolumeSpecName: "config-data") pod "12f40773-d3bf-46ac-96a0-8f114083229f" (UID: "12f40773-d3bf-46ac-96a0-8f114083229f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.264743 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.264936 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mysqld-exporter-0" podUID="e35fc291-f4f0-41fe-93d3-dc2683811a39" containerName="mysqld-exporter" containerID="cri-o://b52dd15229e1e324f3c932df79226b0271e73d92dd9491f1bcb45846272b430c" gracePeriod=30 Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.308127 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12f40773-d3bf-46ac-96a0-8f114083229f-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.308160 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12f40773-d3bf-46ac-96a0-8f114083229f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.389565 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.515414 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-logs\") pod \"6557fcf7-bd46-4552-bc2b-339ff9c9efbe\" (UID: \"6557fcf7-bd46-4552-bc2b-339ff9c9efbe\") " Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.515710 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-config-data\") pod \"6557fcf7-bd46-4552-bc2b-339ff9c9efbe\" (UID: \"6557fcf7-bd46-4552-bc2b-339ff9c9efbe\") " Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.515904 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjdr7\" (UniqueName: \"kubernetes.io/projected/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-kube-api-access-gjdr7\") pod \"6557fcf7-bd46-4552-bc2b-339ff9c9efbe\" (UID: \"6557fcf7-bd46-4552-bc2b-339ff9c9efbe\") " Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.515963 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-combined-ca-bundle\") pod \"6557fcf7-bd46-4552-bc2b-339ff9c9efbe\" (UID: \"6557fcf7-bd46-4552-bc2b-339ff9c9efbe\") " Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.517333 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-logs" (OuterVolumeSpecName: "logs") pod "6557fcf7-bd46-4552-bc2b-339ff9c9efbe" (UID: "6557fcf7-bd46-4552-bc2b-339ff9c9efbe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.533538 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-kube-api-access-gjdr7" (OuterVolumeSpecName: "kube-api-access-gjdr7") pod "6557fcf7-bd46-4552-bc2b-339ff9c9efbe" (UID: "6557fcf7-bd46-4552-bc2b-339ff9c9efbe"). InnerVolumeSpecName "kube-api-access-gjdr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.619744 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjdr7\" (UniqueName: \"kubernetes.io/projected/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-kube-api-access-gjdr7\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.619781 4766 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-logs\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.655147 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6557fcf7-bd46-4552-bc2b-339ff9c9efbe" (UID: "6557fcf7-bd46-4552-bc2b-339ff9c9efbe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.722624 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.744809 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-config-data" (OuterVolumeSpecName: "config-data") pod "6557fcf7-bd46-4552-bc2b-339ff9c9efbe" (UID: "6557fcf7-bd46-4552-bc2b-339ff9c9efbe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.785564 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.823966 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfd3ab8a-e243-40a4-b8fc-b9c094339c78-config-data\") pod \"cfd3ab8a-e243-40a4-b8fc-b9c094339c78\" (UID: \"cfd3ab8a-e243-40a4-b8fc-b9c094339c78\") " Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.824492 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfd3ab8a-e243-40a4-b8fc-b9c094339c78-combined-ca-bundle\") pod \"cfd3ab8a-e243-40a4-b8fc-b9c094339c78\" (UID: \"cfd3ab8a-e243-40a4-b8fc-b9c094339c78\") " Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.824642 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l742j\" (UniqueName: \"kubernetes.io/projected/cfd3ab8a-e243-40a4-b8fc-b9c094339c78-kube-api-access-l742j\") pod \"cfd3ab8a-e243-40a4-b8fc-b9c094339c78\" (UID: \"cfd3ab8a-e243-40a4-b8fc-b9c094339c78\") " Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.825121 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6557fcf7-bd46-4552-bc2b-339ff9c9efbe-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.876695 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfd3ab8a-e243-40a4-b8fc-b9c094339c78-kube-api-access-l742j" (OuterVolumeSpecName: "kube-api-access-l742j") pod "cfd3ab8a-e243-40a4-b8fc-b9c094339c78" (UID: "cfd3ab8a-e243-40a4-b8fc-b9c094339c78"). InnerVolumeSpecName "kube-api-access-l742j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.929500 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l742j\" (UniqueName: \"kubernetes.io/projected/cfd3ab8a-e243-40a4-b8fc-b9c094339c78-kube-api-access-l742j\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.955786 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfd3ab8a-e243-40a4-b8fc-b9c094339c78-config-data" (OuterVolumeSpecName: "config-data") pod "cfd3ab8a-e243-40a4-b8fc-b9c094339c78" (UID: "cfd3ab8a-e243-40a4-b8fc-b9c094339c78"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:29 crc kubenswrapper[4766]: I1126 00:49:29.955885 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfd3ab8a-e243-40a4-b8fc-b9c094339c78-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cfd3ab8a-e243-40a4-b8fc-b9c094339c78" (UID: "cfd3ab8a-e243-40a4-b8fc-b9c094339c78"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.032610 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfd3ab8a-e243-40a4-b8fc-b9c094339c78-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.032659 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfd3ab8a-e243-40a4-b8fc-b9c094339c78-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.130351 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cfd3ab8a-e243-40a4-b8fc-b9c094339c78","Type":"ContainerDied","Data":"94ab299f474cc0da858abfaab622396a22fb3bca09edcf9e3f83028b839af4ef"} Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.130398 4766 scope.go:117] "RemoveContainer" containerID="2a91d5d53a9515d269adb9138b4124002864c8aa09d66aa9e808ff56cf6b1c34" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.130513 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.143944 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 26 00:49:30 crc kubenswrapper[4766]: E1126 00:49:30.144436 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfd3ab8a-e243-40a4-b8fc-b9c094339c78" containerName="nova-cell1-novncproxy-novncproxy" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.144453 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfd3ab8a-e243-40a4-b8fc-b9c094339c78" containerName="nova-cell1-novncproxy-novncproxy" Nov 26 00:49:30 crc kubenswrapper[4766]: E1126 00:49:30.144471 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6557fcf7-bd46-4552-bc2b-339ff9c9efbe" containerName="nova-metadata-metadata" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.144478 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6557fcf7-bd46-4552-bc2b-339ff9c9efbe" containerName="nova-metadata-metadata" Nov 26 00:49:30 crc kubenswrapper[4766]: E1126 00:49:30.144504 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12f40773-d3bf-46ac-96a0-8f114083229f" containerName="nova-cell1-conductor-db-sync" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.144510 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="12f40773-d3bf-46ac-96a0-8f114083229f" containerName="nova-cell1-conductor-db-sync" Nov 26 00:49:30 crc kubenswrapper[4766]: E1126 00:49:30.144535 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c" containerName="nova-manage" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.144541 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c" containerName="nova-manage" Nov 26 00:49:30 crc kubenswrapper[4766]: E1126 00:49:30.144550 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6557fcf7-bd46-4552-bc2b-339ff9c9efbe" containerName="nova-metadata-log" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.144556 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6557fcf7-bd46-4552-bc2b-339ff9c9efbe" containerName="nova-metadata-log" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.144770 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="12f40773-d3bf-46ac-96a0-8f114083229f" containerName="nova-cell1-conductor-db-sync" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.144798 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c" containerName="nova-manage" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.144818 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="6557fcf7-bd46-4552-bc2b-339ff9c9efbe" containerName="nova-metadata-log" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.144830 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="6557fcf7-bd46-4552-bc2b-339ff9c9efbe" containerName="nova-metadata-metadata" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.144842 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfd3ab8a-e243-40a4-b8fc-b9c094339c78" containerName="nova-cell1-novncproxy-novncproxy" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.145576 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.147623 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.147948 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.148919 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.162983 4766 generic.go:334] "Generic (PLEG): container finished" podID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" containerID="3a54ae3d324cd23367e4b743604acd325d6f1f79b6e146944b0f49878887b855" exitCode=2 Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.163116 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"aa692ff7-2356-4ff4-b9fe-08884e4081cf","Type":"ContainerDied","Data":"3a54ae3d324cd23367e4b743604acd325d6f1f79b6e146944b0f49878887b855"} Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.163145 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"aa692ff7-2356-4ff4-b9fe-08884e4081cf","Type":"ContainerDied","Data":"4d43c11d7971dbe73139e7fa058242d5076d9ac157ea01e58c08708156ac63f7"} Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.178761 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.182619 4766 scope.go:117] "RemoveContainer" containerID="3a54ae3d324cd23367e4b743604acd325d6f1f79b6e146944b0f49878887b855" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.190300 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6557fcf7-bd46-4552-bc2b-339ff9c9efbe","Type":"ContainerDied","Data":"ff56ced604a636bed74357e9c1cd70b40b131d4d56f331d3ee2e1b1df06267da"} Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.190444 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.236105 4766 generic.go:334] "Generic (PLEG): container finished" podID="580ecfc1-6334-44d3-a833-a68f98364860" containerID="3020c9120c42026824f47a3298746387e75244633107d3b7564271252dfab3ae" exitCode=0 Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.236218 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fr5xr" event={"ID":"580ecfc1-6334-44d3-a833-a68f98364860","Type":"ContainerDied","Data":"3020c9120c42026824f47a3298746387e75244633107d3b7564271252dfab3ae"} Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.236276 4766 scope.go:117] "RemoveContainer" containerID="fc1da18b2031a9683dec239eb1c313b14139ca65098b96f7294b63e1da1f01ac" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.241869 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e35fc291-f4f0-41fe-93d3-dc2683811a39-config-data\") pod \"e35fc291-f4f0-41fe-93d3-dc2683811a39\" (UID: \"e35fc291-f4f0-41fe-93d3-dc2683811a39\") " Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.242017 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tw5zf\" (UniqueName: \"kubernetes.io/projected/e35fc291-f4f0-41fe-93d3-dc2683811a39-kube-api-access-tw5zf\") pod \"e35fc291-f4f0-41fe-93d3-dc2683811a39\" (UID: \"e35fc291-f4f0-41fe-93d3-dc2683811a39\") " Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.242088 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sjwx\" (UniqueName: \"kubernetes.io/projected/aa692ff7-2356-4ff4-b9fe-08884e4081cf-kube-api-access-6sjwx\") pod \"aa692ff7-2356-4ff4-b9fe-08884e4081cf\" (UID: \"aa692ff7-2356-4ff4-b9fe-08884e4081cf\") " Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.242219 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e35fc291-f4f0-41fe-93d3-dc2683811a39-combined-ca-bundle\") pod \"e35fc291-f4f0-41fe-93d3-dc2683811a39\" (UID: \"e35fc291-f4f0-41fe-93d3-dc2683811a39\") " Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.254205 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e35fc291-f4f0-41fe-93d3-dc2683811a39-kube-api-access-tw5zf" (OuterVolumeSpecName: "kube-api-access-tw5zf") pod "e35fc291-f4f0-41fe-93d3-dc2683811a39" (UID: "e35fc291-f4f0-41fe-93d3-dc2683811a39"). InnerVolumeSpecName "kube-api-access-tw5zf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.269959 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.270574 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mbqtt" event={"ID":"ed786021-3dca-48e5-b286-f1a06dcbe58e","Type":"ContainerStarted","Data":"f1b33ec75dc749ccd1d290210016ab9fe7c2d1feac99e45044b55369a17d1428"} Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.283853 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa692ff7-2356-4ff4-b9fe-08884e4081cf-kube-api-access-6sjwx" (OuterVolumeSpecName: "kube-api-access-6sjwx") pod "aa692ff7-2356-4ff4-b9fe-08884e4081cf" (UID: "aa692ff7-2356-4ff4-b9fe-08884e4081cf"). InnerVolumeSpecName "kube-api-access-6sjwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.284252 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2ec2417-bd6f-4166-b42c-407424769274-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f2ec2417-bd6f-4166-b42c-407424769274\") " pod="openstack/nova-cell1-conductor-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.284392 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2ec2417-bd6f-4166-b42c-407424769274-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f2ec2417-bd6f-4166-b42c-407424769274\") " pod="openstack/nova-cell1-conductor-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.284611 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj9sh\" (UniqueName: \"kubernetes.io/projected/f2ec2417-bd6f-4166-b42c-407424769274-kube-api-access-xj9sh\") pod \"nova-cell1-conductor-0\" (UID: \"f2ec2417-bd6f-4166-b42c-407424769274\") " pod="openstack/nova-cell1-conductor-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.285691 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tw5zf\" (UniqueName: \"kubernetes.io/projected/e35fc291-f4f0-41fe-93d3-dc2683811a39-kube-api-access-tw5zf\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.285712 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sjwx\" (UniqueName: \"kubernetes.io/projected/aa692ff7-2356-4ff4-b9fe-08884e4081cf-kube-api-access-6sjwx\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.325907 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.348410 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-mfr52" event={"ID":"638a59e9-68c3-43cf-ae3b-4510e7df42ec","Type":"ContainerStarted","Data":"89528903de6bdb75b16ed3af567fb4d84f8c699933dfe36d9f49023cb06484b2"} Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.376381 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 26 00:49:30 crc kubenswrapper[4766]: E1126 00:49:30.377474 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" containerName="kube-state-metrics" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.377493 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" containerName="kube-state-metrics" Nov 26 00:49:30 crc kubenswrapper[4766]: E1126 00:49:30.377528 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" containerName="kube-state-metrics" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.377536 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" containerName="kube-state-metrics" Nov 26 00:49:30 crc kubenswrapper[4766]: E1126 00:49:30.377562 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e35fc291-f4f0-41fe-93d3-dc2683811a39" containerName="mysqld-exporter" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.377568 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e35fc291-f4f0-41fe-93d3-dc2683811a39" containerName="mysqld-exporter" Nov 26 00:49:30 crc kubenswrapper[4766]: E1126 00:49:30.377600 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" containerName="kube-state-metrics" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.377606 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" containerName="kube-state-metrics" Nov 26 00:49:30 crc kubenswrapper[4766]: E1126 00:49:30.377624 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" containerName="kube-state-metrics" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.377630 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" containerName="kube-state-metrics" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.378085 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" containerName="kube-state-metrics" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.378109 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" containerName="kube-state-metrics" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.378161 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="e35fc291-f4f0-41fe-93d3-dc2683811a39" containerName="mysqld-exporter" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.379286 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.381879 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.387216 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.389407 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2ec2417-bd6f-4166-b42c-407424769274-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f2ec2417-bd6f-4166-b42c-407424769274\") " pod="openstack/nova-cell1-conductor-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.389469 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2ec2417-bd6f-4166-b42c-407424769274-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f2ec2417-bd6f-4166-b42c-407424769274\") " pod="openstack/nova-cell1-conductor-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.389577 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj9sh\" (UniqueName: \"kubernetes.io/projected/f2ec2417-bd6f-4166-b42c-407424769274-kube-api-access-xj9sh\") pod \"nova-cell1-conductor-0\" (UID: \"f2ec2417-bd6f-4166-b42c-407424769274\") " pod="openstack/nova-cell1-conductor-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.404224 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.411765 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2ec2417-bd6f-4166-b42c-407424769274-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f2ec2417-bd6f-4166-b42c-407424769274\") " pod="openstack/nova-cell1-conductor-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.412517 4766 generic.go:334] "Generic (PLEG): container finished" podID="e35fc291-f4f0-41fe-93d3-dc2683811a39" containerID="b52dd15229e1e324f3c932df79226b0271e73d92dd9491f1bcb45846272b430c" exitCode=2 Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.412582 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"e35fc291-f4f0-41fe-93d3-dc2683811a39","Type":"ContainerDied","Data":"b52dd15229e1e324f3c932df79226b0271e73d92dd9491f1bcb45846272b430c"} Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.412620 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.412623 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"e35fc291-f4f0-41fe-93d3-dc2683811a39","Type":"ContainerDied","Data":"68dcaf6bfe9d46079cdcd180975134d8c3c33d493ab36a127d3a072ba1933397"} Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.418691 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2ec2417-bd6f-4166-b42c-407424769274-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f2ec2417-bd6f-4166-b42c-407424769274\") " pod="openstack/nova-cell1-conductor-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.422947 4766 scope.go:117] "RemoveContainer" containerID="3a54ae3d324cd23367e4b743604acd325d6f1f79b6e146944b0f49878887b855" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.429538 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e35fc291-f4f0-41fe-93d3-dc2683811a39-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e35fc291-f4f0-41fe-93d3-dc2683811a39" (UID: "e35fc291-f4f0-41fe-93d3-dc2683811a39"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.429760 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e35fc291-f4f0-41fe-93d3-dc2683811a39-config-data" (OuterVolumeSpecName: "config-data") pod "e35fc291-f4f0-41fe-93d3-dc2683811a39" (UID: "e35fc291-f4f0-41fe-93d3-dc2683811a39"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:30 crc kubenswrapper[4766]: E1126 00:49:30.429772 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a54ae3d324cd23367e4b743604acd325d6f1f79b6e146944b0f49878887b855\": container with ID starting with 3a54ae3d324cd23367e4b743604acd325d6f1f79b6e146944b0f49878887b855 not found: ID does not exist" containerID="3a54ae3d324cd23367e4b743604acd325d6f1f79b6e146944b0f49878887b855" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.429850 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a54ae3d324cd23367e4b743604acd325d6f1f79b6e146944b0f49878887b855"} err="failed to get container status \"3a54ae3d324cd23367e4b743604acd325d6f1f79b6e146944b0f49878887b855\": rpc error: code = NotFound desc = could not find container \"3a54ae3d324cd23367e4b743604acd325d6f1f79b6e146944b0f49878887b855\": container with ID starting with 3a54ae3d324cd23367e4b743604acd325d6f1f79b6e146944b0f49878887b855 not found: ID does not exist" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.429873 4766 scope.go:117] "RemoveContainer" containerID="fc1da18b2031a9683dec239eb1c313b14139ca65098b96f7294b63e1da1f01ac" Nov 26 00:49:30 crc kubenswrapper[4766]: E1126 00:49:30.432388 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc1da18b2031a9683dec239eb1c313b14139ca65098b96f7294b63e1da1f01ac\": container with ID starting with fc1da18b2031a9683dec239eb1c313b14139ca65098b96f7294b63e1da1f01ac not found: ID does not exist" containerID="fc1da18b2031a9683dec239eb1c313b14139ca65098b96f7294b63e1da1f01ac" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.432431 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc1da18b2031a9683dec239eb1c313b14139ca65098b96f7294b63e1da1f01ac"} err="failed to get container status \"fc1da18b2031a9683dec239eb1c313b14139ca65098b96f7294b63e1da1f01ac\": rpc error: code = NotFound desc = could not find container \"fc1da18b2031a9683dec239eb1c313b14139ca65098b96f7294b63e1da1f01ac\": container with ID starting with fc1da18b2031a9683dec239eb1c313b14139ca65098b96f7294b63e1da1f01ac not found: ID does not exist" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.432458 4766 scope.go:117] "RemoveContainer" containerID="fee5153714804d61e61c312f61aacf0b71754de605ae56a39bcb1765403a2d88" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.469430 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj9sh\" (UniqueName: \"kubernetes.io/projected/f2ec2417-bd6f-4166-b42c-407424769274-kube-api-access-xj9sh\") pod \"nova-cell1-conductor-0\" (UID: \"f2ec2417-bd6f-4166-b42c-407424769274\") " pod="openstack/nova-cell1-conductor-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.488115 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.499018 4766 scope.go:117] "RemoveContainer" containerID="a045755472e90bfaa78d93793ed4e685edd333b3ab9ab8f6ebc16a200ca28ee7" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.500433 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cf2f787-c351-49ce-a347-dc0975ea8bc4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cf2f787-c351-49ce-a347-dc0975ea8bc4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.500508 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljzmf\" (UniqueName: \"kubernetes.io/projected/4cf2f787-c351-49ce-a347-dc0975ea8bc4-kube-api-access-ljzmf\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cf2f787-c351-49ce-a347-dc0975ea8bc4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.500573 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cf2f787-c351-49ce-a347-dc0975ea8bc4-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cf2f787-c351-49ce-a347-dc0975ea8bc4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.500675 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cf2f787-c351-49ce-a347-dc0975ea8bc4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cf2f787-c351-49ce-a347-dc0975ea8bc4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.500778 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cf2f787-c351-49ce-a347-dc0975ea8bc4-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cf2f787-c351-49ce-a347-dc0975ea8bc4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.500935 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e35fc291-f4f0-41fe-93d3-dc2683811a39-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.500951 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e35fc291-f4f0-41fe-93d3-dc2683811a39-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.520196 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.611332 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cf2f787-c351-49ce-a347-dc0975ea8bc4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cf2f787-c351-49ce-a347-dc0975ea8bc4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.611420 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cf2f787-c351-49ce-a347-dc0975ea8bc4-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cf2f787-c351-49ce-a347-dc0975ea8bc4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.611507 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cf2f787-c351-49ce-a347-dc0975ea8bc4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cf2f787-c351-49ce-a347-dc0975ea8bc4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.611533 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljzmf\" (UniqueName: \"kubernetes.io/projected/4cf2f787-c351-49ce-a347-dc0975ea8bc4-kube-api-access-ljzmf\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cf2f787-c351-49ce-a347-dc0975ea8bc4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.611568 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cf2f787-c351-49ce-a347-dc0975ea8bc4-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cf2f787-c351-49ce-a347-dc0975ea8bc4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.647058 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cf2f787-c351-49ce-a347-dc0975ea8bc4-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cf2f787-c351-49ce-a347-dc0975ea8bc4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.647492 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cf2f787-c351-49ce-a347-dc0975ea8bc4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cf2f787-c351-49ce-a347-dc0975ea8bc4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.653220 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cf2f787-c351-49ce-a347-dc0975ea8bc4-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cf2f787-c351-49ce-a347-dc0975ea8bc4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.653284 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.657190 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljzmf\" (UniqueName: \"kubernetes.io/projected/4cf2f787-c351-49ce-a347-dc0975ea8bc4-kube-api-access-ljzmf\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cf2f787-c351-49ce-a347-dc0975ea8bc4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.660182 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cf2f787-c351-49ce-a347-dc0975ea8bc4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cf2f787-c351-49ce-a347-dc0975ea8bc4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.692836 4766 scope.go:117] "RemoveContainer" containerID="b52dd15229e1e324f3c932df79226b0271e73d92dd9491f1bcb45846272b430c" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.712833 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.736787 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.740102 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mbqtt" podStartSLOduration=4.543068521 podStartE2EDuration="12.74008671s" podCreationTimestamp="2025-11-26 00:49:18 +0000 UTC" firstStartedPulling="2025-11-26 00:49:20.7658169 +0000 UTC m=+1541.614587320" lastFinishedPulling="2025-11-26 00:49:28.962835079 +0000 UTC m=+1549.811605509" observedRunningTime="2025-11-26 00:49:30.376180974 +0000 UTC m=+1551.224951404" watchObservedRunningTime="2025-11-26 00:49:30.74008671 +0000 UTC m=+1551.588857140" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.748767 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" containerName="kube-state-metrics" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.748821 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" containerName="kube-state-metrics" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.750221 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.756699 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.757416 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.803162 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.819322 4766 scope.go:117] "RemoveContainer" containerID="b52dd15229e1e324f3c932df79226b0271e73d92dd9491f1bcb45846272b430c" Nov 26 00:49:30 crc kubenswrapper[4766]: E1126 00:49:30.823024 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b52dd15229e1e324f3c932df79226b0271e73d92dd9491f1bcb45846272b430c\": container with ID starting with b52dd15229e1e324f3c932df79226b0271e73d92dd9491f1bcb45846272b430c not found: ID does not exist" containerID="b52dd15229e1e324f3c932df79226b0271e73d92dd9491f1bcb45846272b430c" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.823065 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b52dd15229e1e324f3c932df79226b0271e73d92dd9491f1bcb45846272b430c"} err="failed to get container status \"b52dd15229e1e324f3c932df79226b0271e73d92dd9491f1bcb45846272b430c\": rpc error: code = NotFound desc = could not find container \"b52dd15229e1e324f3c932df79226b0271e73d92dd9491f1bcb45846272b430c\": container with ID starting with b52dd15229e1e324f3c932df79226b0271e73d92dd9491f1bcb45846272b430c not found: ID does not exist" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.841451 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.919550 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.919807 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab" containerName="nova-scheduler-scheduler" containerID="cri-o://00f3b60244d89847b3a53fd77058301dc3ec73e6fc1ebe02e2044b53df8adf65" gracePeriod=30 Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.934001 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69d94442-932b-417d-8857-f17acfc62174-logs\") pod \"nova-metadata-0\" (UID: \"69d94442-932b-417d-8857-f17acfc62174\") " pod="openstack/nova-metadata-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.934105 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69d94442-932b-417d-8857-f17acfc62174-config-data\") pod \"nova-metadata-0\" (UID: \"69d94442-932b-417d-8857-f17acfc62174\") " pod="openstack/nova-metadata-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.934136 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69d94442-932b-417d-8857-f17acfc62174-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"69d94442-932b-417d-8857-f17acfc62174\") " pod="openstack/nova-metadata-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.934174 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbxj9\" (UniqueName: \"kubernetes.io/projected/69d94442-932b-417d-8857-f17acfc62174-kube-api-access-cbxj9\") pod \"nova-metadata-0\" (UID: \"69d94442-932b-417d-8857-f17acfc62174\") " pod="openstack/nova-metadata-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.934247 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/69d94442-932b-417d-8857-f17acfc62174-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"69d94442-932b-417d-8857-f17acfc62174\") " pod="openstack/nova-metadata-0" Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.957914 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 26 00:49:30 crc kubenswrapper[4766]: I1126 00:49:30.988782 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-mfr52" podStartSLOduration=3.769434516 podStartE2EDuration="12.988756796s" podCreationTimestamp="2025-11-26 00:49:18 +0000 UTC" firstStartedPulling="2025-11-26 00:49:19.800644902 +0000 UTC m=+1540.649415332" lastFinishedPulling="2025-11-26 00:49:29.019967182 +0000 UTC m=+1549.868737612" observedRunningTime="2025-11-26 00:49:30.587053827 +0000 UTC m=+1551.435824257" watchObservedRunningTime="2025-11-26 00:49:30.988756796 +0000 UTC m=+1551.837527236" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.025116 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 00:49:31 crc kubenswrapper[4766]: E1126 00:49:31.026303 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-cbxj9 logs nova-metadata-tls-certs], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/nova-metadata-0" podUID="69d94442-932b-417d-8857-f17acfc62174" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.036909 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/69d94442-932b-417d-8857-f17acfc62174-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"69d94442-932b-417d-8857-f17acfc62174\") " pod="openstack/nova-metadata-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.037093 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69d94442-932b-417d-8857-f17acfc62174-logs\") pod \"nova-metadata-0\" (UID: \"69d94442-932b-417d-8857-f17acfc62174\") " pod="openstack/nova-metadata-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.037184 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69d94442-932b-417d-8857-f17acfc62174-config-data\") pod \"nova-metadata-0\" (UID: \"69d94442-932b-417d-8857-f17acfc62174\") " pod="openstack/nova-metadata-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.037221 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69d94442-932b-417d-8857-f17acfc62174-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"69d94442-932b-417d-8857-f17acfc62174\") " pod="openstack/nova-metadata-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.037254 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbxj9\" (UniqueName: \"kubernetes.io/projected/69d94442-932b-417d-8857-f17acfc62174-kube-api-access-cbxj9\") pod \"nova-metadata-0\" (UID: \"69d94442-932b-417d-8857-f17acfc62174\") " pod="openstack/nova-metadata-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.045469 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69d94442-932b-417d-8857-f17acfc62174-logs\") pod \"nova-metadata-0\" (UID: \"69d94442-932b-417d-8857-f17acfc62174\") " pod="openstack/nova-metadata-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.057496 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69d94442-932b-417d-8857-f17acfc62174-config-data\") pod \"nova-metadata-0\" (UID: \"69d94442-932b-417d-8857-f17acfc62174\") " pod="openstack/nova-metadata-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.063261 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69d94442-932b-417d-8857-f17acfc62174-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"69d94442-932b-417d-8857-f17acfc62174\") " pod="openstack/nova-metadata-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.065190 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/69d94442-932b-417d-8857-f17acfc62174-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"69d94442-932b-417d-8857-f17acfc62174\") " pod="openstack/nova-metadata-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.081840 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbxj9\" (UniqueName: \"kubernetes.io/projected/69d94442-932b-417d-8857-f17acfc62174-kube-api-access-cbxj9\") pod \"nova-metadata-0\" (UID: \"69d94442-932b-417d-8857-f17acfc62174\") " pod="openstack/nova-metadata-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.129823 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84fb747687-wsb8c"] Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.131739 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.142781 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84fb747687-wsb8c"] Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.156394 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.176909 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.199638 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.201324 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.214113 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-mysqld-exporter-svc" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.214404 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.217751 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.248332 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-dns-svc\") pod \"dnsmasq-dns-84fb747687-wsb8c\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.248432 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-dns-swift-storage-0\") pod \"dnsmasq-dns-84fb747687-wsb8c\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.248469 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-ovsdbserver-nb\") pod \"dnsmasq-dns-84fb747687-wsb8c\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.249072 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-ovsdbserver-sb\") pod \"dnsmasq-dns-84fb747687-wsb8c\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.249155 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb4g9\" (UniqueName: \"kubernetes.io/projected/87b59634-29ce-4283-a280-ebd27426348a-kube-api-access-tb4g9\") pod \"dnsmasq-dns-84fb747687-wsb8c\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.249287 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-config\") pod \"dnsmasq-dns-84fb747687-wsb8c\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.254147 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84fb747687-wsb8c"] Nov 26 00:49:31 crc kubenswrapper[4766]: E1126 00:49:31.255566 4766 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode35fc291_f4f0_41fe_93d3_dc2683811a39.slice/crio-68dcaf6bfe9d46079cdcd180975134d8c3c33d493ab36a127d3a072ba1933397\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode35fc291_f4f0_41fe_93d3_dc2683811a39.slice\": RecentStats: unable to find data in memory cache]" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.285718 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-ff4gf"] Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.289149 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.336170 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-ff4gf"] Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.356466 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-config\") pod \"dnsmasq-dns-84fb747687-wsb8c\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.356561 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p7hj\" (UniqueName: \"kubernetes.io/projected/51bb050a-0106-4f7c-acf2-8c6c19f93d39-kube-api-access-9p7hj\") pod \"mysqld-exporter-0\" (UID: \"51bb050a-0106-4f7c-acf2-8c6c19f93d39\") " pod="openstack/mysqld-exporter-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.356636 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-dns-svc\") pod \"dnsmasq-dns-84fb747687-wsb8c\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.356701 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51bb050a-0106-4f7c-acf2-8c6c19f93d39-config-data\") pod \"mysqld-exporter-0\" (UID: \"51bb050a-0106-4f7c-acf2-8c6c19f93d39\") " pod="openstack/mysqld-exporter-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.356796 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-dns-swift-storage-0\") pod \"dnsmasq-dns-84fb747687-wsb8c\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.356856 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-ovsdbserver-nb\") pod \"dnsmasq-dns-84fb747687-wsb8c\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.357007 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-ovsdbserver-sb\") pod \"dnsmasq-dns-84fb747687-wsb8c\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.357079 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/51bb050a-0106-4f7c-acf2-8c6c19f93d39-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"51bb050a-0106-4f7c-acf2-8c6c19f93d39\") " pod="openstack/mysqld-exporter-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.357128 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb4g9\" (UniqueName: \"kubernetes.io/projected/87b59634-29ce-4283-a280-ebd27426348a-kube-api-access-tb4g9\") pod \"dnsmasq-dns-84fb747687-wsb8c\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.357187 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51bb050a-0106-4f7c-acf2-8c6c19f93d39-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"51bb050a-0106-4f7c-acf2-8c6c19f93d39\") " pod="openstack/mysqld-exporter-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.363904 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-config\") pod \"dnsmasq-dns-84fb747687-wsb8c\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.365743 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-ovsdbserver-nb\") pod \"dnsmasq-dns-84fb747687-wsb8c\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.367135 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-dns-swift-storage-0\") pod \"dnsmasq-dns-84fb747687-wsb8c\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.373181 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-dns-svc\") pod \"dnsmasq-dns-84fb747687-wsb8c\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.377406 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-ovsdbserver-sb\") pod \"dnsmasq-dns-84fb747687-wsb8c\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.399263 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb4g9\" (UniqueName: \"kubernetes.io/projected/87b59634-29ce-4283-a280-ebd27426348a-kube-api-access-tb4g9\") pod \"dnsmasq-dns-84fb747687-wsb8c\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:31 crc kubenswrapper[4766]: E1126 00:49:31.405264 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-tb4g9], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-84fb747687-wsb8c" podUID="87b59634-29ce-4283-a280-ebd27426348a" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.459965 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p7hj\" (UniqueName: \"kubernetes.io/projected/51bb050a-0106-4f7c-acf2-8c6c19f93d39-kube-api-access-9p7hj\") pod \"mysqld-exporter-0\" (UID: \"51bb050a-0106-4f7c-acf2-8c6c19f93d39\") " pod="openstack/mysqld-exporter-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.460042 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqp58\" (UniqueName: \"kubernetes.io/projected/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-kube-api-access-dqp58\") pod \"dnsmasq-dns-f84f9ccf-ff4gf\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.460089 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51bb050a-0106-4f7c-acf2-8c6c19f93d39-config-data\") pod \"mysqld-exporter-0\" (UID: \"51bb050a-0106-4f7c-acf2-8c6c19f93d39\") " pod="openstack/mysqld-exporter-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.460173 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-dns-swift-storage-0\") pod \"dnsmasq-dns-f84f9ccf-ff4gf\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.460262 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-ovsdbserver-sb\") pod \"dnsmasq-dns-f84f9ccf-ff4gf\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.460283 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-ovsdbserver-nb\") pod \"dnsmasq-dns-f84f9ccf-ff4gf\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.460328 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-dns-svc\") pod \"dnsmasq-dns-f84f9ccf-ff4gf\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.460354 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-config\") pod \"dnsmasq-dns-f84f9ccf-ff4gf\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.460376 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/51bb050a-0106-4f7c-acf2-8c6c19f93d39-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"51bb050a-0106-4f7c-acf2-8c6c19f93d39\") " pod="openstack/mysqld-exporter-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.460427 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51bb050a-0106-4f7c-acf2-8c6c19f93d39-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"51bb050a-0106-4f7c-acf2-8c6c19f93d39\") " pod="openstack/mysqld-exporter-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.469468 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51bb050a-0106-4f7c-acf2-8c6c19f93d39-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"51bb050a-0106-4f7c-acf2-8c6c19f93d39\") " pod="openstack/mysqld-exporter-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.473739 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51bb050a-0106-4f7c-acf2-8c6c19f93d39-config-data\") pod \"mysqld-exporter-0\" (UID: \"51bb050a-0106-4f7c-acf2-8c6c19f93d39\") " pod="openstack/mysqld-exporter-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.479165 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/51bb050a-0106-4f7c-acf2-8c6c19f93d39-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"51bb050a-0106-4f7c-acf2-8c6c19f93d39\") " pod="openstack/mysqld-exporter-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.495397 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p7hj\" (UniqueName: \"kubernetes.io/projected/51bb050a-0106-4f7c-acf2-8c6c19f93d39-kube-api-access-9p7hj\") pod \"mysqld-exporter-0\" (UID: \"51bb050a-0106-4f7c-acf2-8c6c19f93d39\") " pod="openstack/mysqld-exporter-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.527159 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.541036 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.545769 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.561908 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqp58\" (UniqueName: \"kubernetes.io/projected/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-kube-api-access-dqp58\") pod \"dnsmasq-dns-f84f9ccf-ff4gf\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.562022 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-dns-swift-storage-0\") pod \"dnsmasq-dns-f84f9ccf-ff4gf\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.562109 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-ovsdbserver-sb\") pod \"dnsmasq-dns-f84f9ccf-ff4gf\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.562134 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-ovsdbserver-nb\") pod \"dnsmasq-dns-f84f9ccf-ff4gf\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.562193 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-dns-svc\") pod \"dnsmasq-dns-f84f9ccf-ff4gf\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.562227 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-config\") pod \"dnsmasq-dns-f84f9ccf-ff4gf\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.571146 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-dns-swift-storage-0\") pod \"dnsmasq-dns-f84f9ccf-ff4gf\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.576664 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-dns-svc\") pod \"dnsmasq-dns-f84f9ccf-ff4gf\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.578552 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-ovsdbserver-nb\") pod \"dnsmasq-dns-f84f9ccf-ff4gf\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.583562 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-ovsdbserver-sb\") pod \"dnsmasq-dns-f84f9ccf-ff4gf\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.590741 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.591751 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-config\") pod \"dnsmasq-dns-f84f9ccf-ff4gf\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.611115 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqp58\" (UniqueName: \"kubernetes.io/projected/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-kube-api-access-dqp58\") pod \"dnsmasq-dns-f84f9ccf-ff4gf\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.613216 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.632190 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.660123 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.663052 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-dns-swift-storage-0\") pod \"87b59634-29ce-4283-a280-ebd27426348a\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.663177 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-dns-svc\") pod \"87b59634-29ce-4283-a280-ebd27426348a\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.663198 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-config\") pod \"87b59634-29ce-4283-a280-ebd27426348a\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.663225 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-ovsdbserver-sb\") pod \"87b59634-29ce-4283-a280-ebd27426348a\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.663330 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tb4g9\" (UniqueName: \"kubernetes.io/projected/87b59634-29ce-4283-a280-ebd27426348a-kube-api-access-tb4g9\") pod \"87b59634-29ce-4283-a280-ebd27426348a\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.663435 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-ovsdbserver-nb\") pod \"87b59634-29ce-4283-a280-ebd27426348a\" (UID: \"87b59634-29ce-4283-a280-ebd27426348a\") " Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.664825 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "87b59634-29ce-4283-a280-ebd27426348a" (UID: "87b59634-29ce-4283-a280-ebd27426348a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.665135 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "87b59634-29ce-4283-a280-ebd27426348a" (UID: "87b59634-29ce-4283-a280-ebd27426348a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.665586 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-config" (OuterVolumeSpecName: "config") pod "87b59634-29ce-4283-a280-ebd27426348a" (UID: "87b59634-29ce-4283-a280-ebd27426348a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.665975 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "87b59634-29ce-4283-a280-ebd27426348a" (UID: "87b59634-29ce-4283-a280-ebd27426348a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.667712 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "87b59634-29ce-4283-a280-ebd27426348a" (UID: "87b59634-29ce-4283-a280-ebd27426348a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.678060 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87b59634-29ce-4283-a280-ebd27426348a-kube-api-access-tb4g9" (OuterVolumeSpecName: "kube-api-access-tb4g9") pod "87b59634-29ce-4283-a280-ebd27426348a" (UID: "87b59634-29ce-4283-a280-ebd27426348a"). InnerVolumeSpecName "kube-api-access-tb4g9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.683711 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.685276 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.693581 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.693780 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.721342 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.765373 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.766439 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69d94442-932b-417d-8857-f17acfc62174-config-data\") pod \"69d94442-932b-417d-8857-f17acfc62174\" (UID: \"69d94442-932b-417d-8857-f17acfc62174\") " Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.766774 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69d94442-932b-417d-8857-f17acfc62174-logs\") pod \"69d94442-932b-417d-8857-f17acfc62174\" (UID: \"69d94442-932b-417d-8857-f17acfc62174\") " Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.766876 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbxj9\" (UniqueName: \"kubernetes.io/projected/69d94442-932b-417d-8857-f17acfc62174-kube-api-access-cbxj9\") pod \"69d94442-932b-417d-8857-f17acfc62174\" (UID: \"69d94442-932b-417d-8857-f17acfc62174\") " Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.766918 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/69d94442-932b-417d-8857-f17acfc62174-nova-metadata-tls-certs\") pod \"69d94442-932b-417d-8857-f17acfc62174\" (UID: \"69d94442-932b-417d-8857-f17acfc62174\") " Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.766969 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69d94442-932b-417d-8857-f17acfc62174-combined-ca-bundle\") pod \"69d94442-932b-417d-8857-f17acfc62174\" (UID: \"69d94442-932b-417d-8857-f17acfc62174\") " Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.767345 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69d94442-932b-417d-8857-f17acfc62174-logs" (OuterVolumeSpecName: "logs") pod "69d94442-932b-417d-8857-f17acfc62174" (UID: "69d94442-932b-417d-8857-f17acfc62174"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.773894 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv9gx\" (UniqueName: \"kubernetes.io/projected/a700ac64-3ee9-4e7b-8da8-1cb7537fee78-kube-api-access-wv9gx\") pod \"kube-state-metrics-0\" (UID: \"a700ac64-3ee9-4e7b-8da8-1cb7537fee78\") " pod="openstack/kube-state-metrics-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.773946 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a700ac64-3ee9-4e7b-8da8-1cb7537fee78-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a700ac64-3ee9-4e7b-8da8-1cb7537fee78\") " pod="openstack/kube-state-metrics-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.774131 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a700ac64-3ee9-4e7b-8da8-1cb7537fee78-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a700ac64-3ee9-4e7b-8da8-1cb7537fee78\") " pod="openstack/kube-state-metrics-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.774248 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a700ac64-3ee9-4e7b-8da8-1cb7537fee78-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a700ac64-3ee9-4e7b-8da8-1cb7537fee78\") " pod="openstack/kube-state-metrics-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.774425 4766 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.774444 4766 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.774456 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.774466 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.774478 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tb4g9\" (UniqueName: \"kubernetes.io/projected/87b59634-29ce-4283-a280-ebd27426348a-kube-api-access-tb4g9\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.774488 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87b59634-29ce-4283-a280-ebd27426348a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.774497 4766 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69d94442-932b-417d-8857-f17acfc62174-logs\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.778924 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69d94442-932b-417d-8857-f17acfc62174-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "69d94442-932b-417d-8857-f17acfc62174" (UID: "69d94442-932b-417d-8857-f17acfc62174"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.779145 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69d94442-932b-417d-8857-f17acfc62174-config-data" (OuterVolumeSpecName: "config-data") pod "69d94442-932b-417d-8857-f17acfc62174" (UID: "69d94442-932b-417d-8857-f17acfc62174"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.780494 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69d94442-932b-417d-8857-f17acfc62174-kube-api-access-cbxj9" (OuterVolumeSpecName: "kube-api-access-cbxj9") pod "69d94442-932b-417d-8857-f17acfc62174" (UID: "69d94442-932b-417d-8857-f17acfc62174"). InnerVolumeSpecName "kube-api-access-cbxj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.810292 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.826299 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69d94442-932b-417d-8857-f17acfc62174-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69d94442-932b-417d-8857-f17acfc62174" (UID: "69d94442-932b-417d-8857-f17acfc62174"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.879036 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv9gx\" (UniqueName: \"kubernetes.io/projected/a700ac64-3ee9-4e7b-8da8-1cb7537fee78-kube-api-access-wv9gx\") pod \"kube-state-metrics-0\" (UID: \"a700ac64-3ee9-4e7b-8da8-1cb7537fee78\") " pod="openstack/kube-state-metrics-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.879077 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a700ac64-3ee9-4e7b-8da8-1cb7537fee78-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a700ac64-3ee9-4e7b-8da8-1cb7537fee78\") " pod="openstack/kube-state-metrics-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.879155 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a700ac64-3ee9-4e7b-8da8-1cb7537fee78-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a700ac64-3ee9-4e7b-8da8-1cb7537fee78\") " pod="openstack/kube-state-metrics-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.879219 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a700ac64-3ee9-4e7b-8da8-1cb7537fee78-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a700ac64-3ee9-4e7b-8da8-1cb7537fee78\") " pod="openstack/kube-state-metrics-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.879344 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69d94442-932b-417d-8857-f17acfc62174-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.879356 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbxj9\" (UniqueName: \"kubernetes.io/projected/69d94442-932b-417d-8857-f17acfc62174-kube-api-access-cbxj9\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.879368 4766 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/69d94442-932b-417d-8857-f17acfc62174-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.879377 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69d94442-932b-417d-8857-f17acfc62174-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.885346 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a700ac64-3ee9-4e7b-8da8-1cb7537fee78-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a700ac64-3ee9-4e7b-8da8-1cb7537fee78\") " pod="openstack/kube-state-metrics-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.888822 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6557fcf7-bd46-4552-bc2b-339ff9c9efbe" path="/var/lib/kubelet/pods/6557fcf7-bd46-4552-bc2b-339ff9c9efbe/volumes" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.890037 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a700ac64-3ee9-4e7b-8da8-1cb7537fee78-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a700ac64-3ee9-4e7b-8da8-1cb7537fee78\") " pod="openstack/kube-state-metrics-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.900491 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa692ff7-2356-4ff4-b9fe-08884e4081cf" path="/var/lib/kubelet/pods/aa692ff7-2356-4ff4-b9fe-08884e4081cf/volumes" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.901444 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfd3ab8a-e243-40a4-b8fc-b9c094339c78" path="/var/lib/kubelet/pods/cfd3ab8a-e243-40a4-b8fc-b9c094339c78/volumes" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.901944 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e35fc291-f4f0-41fe-93d3-dc2683811a39" path="/var/lib/kubelet/pods/e35fc291-f4f0-41fe-93d3-dc2683811a39/volumes" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.902874 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.929417 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv9gx\" (UniqueName: \"kubernetes.io/projected/a700ac64-3ee9-4e7b-8da8-1cb7537fee78-kube-api-access-wv9gx\") pod \"kube-state-metrics-0\" (UID: \"a700ac64-3ee9-4e7b-8da8-1cb7537fee78\") " pod="openstack/kube-state-metrics-0" Nov 26 00:49:31 crc kubenswrapper[4766]: I1126 00:49:31.949388 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a700ac64-3ee9-4e7b-8da8-1cb7537fee78-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a700ac64-3ee9-4e7b-8da8-1cb7537fee78\") " pod="openstack/kube-state-metrics-0" Nov 26 00:49:32 crc kubenswrapper[4766]: I1126 00:49:32.015881 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 26 00:49:32 crc kubenswrapper[4766]: I1126 00:49:32.049779 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 26 00:49:32 crc kubenswrapper[4766]: I1126 00:49:32.582217 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fr5xr" event={"ID":"580ecfc1-6334-44d3-a833-a68f98364860","Type":"ContainerStarted","Data":"5d6c296ba873fad78789f27343c9e823f2e4ea3222fcc68fd6f75fdc3d6157e6"} Nov 26 00:49:32 crc kubenswrapper[4766]: I1126 00:49:32.602320 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4cf2f787-c351-49ce-a347-dc0975ea8bc4","Type":"ContainerStarted","Data":"c96b36aecd7c39ff650d96ebf18b9bd7a95cf8312d64112ccd87246b79e05c82"} Nov 26 00:49:32 crc kubenswrapper[4766]: I1126 00:49:32.625030 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84fb747687-wsb8c" Nov 26 00:49:32 crc kubenswrapper[4766]: I1126 00:49:32.626795 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"f2ec2417-bd6f-4166-b42c-407424769274","Type":"ContainerStarted","Data":"ba5955c49225ad780a31a47ca526a02643a3a8182158c1a2e62f33f650f6ec20"} Nov 26 00:49:32 crc kubenswrapper[4766]: I1126 00:49:32.626923 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"f2ec2417-bd6f-4166-b42c-407424769274","Type":"ContainerStarted","Data":"04be57b4c1b7134b347bd46670ab085a96350d3708312c6b593c461b7e5fd96d"} Nov 26 00:49:32 crc kubenswrapper[4766]: I1126 00:49:32.627004 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 26 00:49:32 crc kubenswrapper[4766]: I1126 00:49:32.627107 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 00:49:32 crc kubenswrapper[4766]: I1126 00:49:32.627814 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="66040b18-6e3b-43dc-8d90-a3eeab8f31df" containerName="nova-api-log" containerID="cri-o://252626f0723e2a175f72d821ec6908dff31e1409f3ca28fd691e29d2869b2be6" gracePeriod=30 Nov 26 00:49:32 crc kubenswrapper[4766]: I1126 00:49:32.628354 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="66040b18-6e3b-43dc-8d90-a3eeab8f31df" containerName="nova-api-api" containerID="cri-o://063c2b5642c0debff8bdd6d1fb44f28d827eb58ca69c0244076b2830713167a0" gracePeriod=30 Nov 26 00:49:32 crc kubenswrapper[4766]: I1126 00:49:32.644488 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fr5xr" Nov 26 00:49:32 crc kubenswrapper[4766]: I1126 00:49:32.644668 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fr5xr" Nov 26 00:49:32 crc kubenswrapper[4766]: I1126 00:49:32.651643 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fr5xr" podStartSLOduration=4.632626769 podStartE2EDuration="10.651622379s" podCreationTimestamp="2025-11-26 00:49:22 +0000 UTC" firstStartedPulling="2025-11-26 00:49:24.927626417 +0000 UTC m=+1545.776396847" lastFinishedPulling="2025-11-26 00:49:30.946622027 +0000 UTC m=+1551.795392457" observedRunningTime="2025-11-26 00:49:32.640942013 +0000 UTC m=+1553.489712453" watchObservedRunningTime="2025-11-26 00:49:32.651622379 +0000 UTC m=+1553.500392809" Nov 26 00:49:32 crc kubenswrapper[4766]: I1126 00:49:32.801376 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.80135503 podStartE2EDuration="2.80135503s" podCreationTimestamp="2025-11-26 00:49:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:49:32.763823935 +0000 UTC m=+1553.612594365" watchObservedRunningTime="2025-11-26 00:49:32.80135503 +0000 UTC m=+1553.650125460" Nov 26 00:49:32 crc kubenswrapper[4766]: I1126 00:49:32.912132 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 00:49:32 crc kubenswrapper[4766]: I1126 00:49:32.945536 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 00:49:32 crc kubenswrapper[4766]: I1126 00:49:32.978896 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 26 00:49:32 crc kubenswrapper[4766]: I1126 00:49:32.994970 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.017883 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.019043 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.048313 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.216643 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08a8f08a-1eb6-424c-874d-15efb5083a8d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"08a8f08a-1eb6-424c-874d-15efb5083a8d\") " pod="openstack/nova-metadata-0" Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.219408 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08a8f08a-1eb6-424c-874d-15efb5083a8d-config-data\") pod \"nova-metadata-0\" (UID: \"08a8f08a-1eb6-424c-874d-15efb5083a8d\") " pod="openstack/nova-metadata-0" Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.219483 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/08a8f08a-1eb6-424c-874d-15efb5083a8d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"08a8f08a-1eb6-424c-874d-15efb5083a8d\") " pod="openstack/nova-metadata-0" Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.219594 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fx6cc\" (UniqueName: \"kubernetes.io/projected/08a8f08a-1eb6-424c-874d-15efb5083a8d-kube-api-access-fx6cc\") pod \"nova-metadata-0\" (UID: \"08a8f08a-1eb6-424c-874d-15efb5083a8d\") " pod="openstack/nova-metadata-0" Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.219643 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08a8f08a-1eb6-424c-874d-15efb5083a8d-logs\") pod \"nova-metadata-0\" (UID: \"08a8f08a-1eb6-424c-874d-15efb5083a8d\") " pod="openstack/nova-metadata-0" Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.234951 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-ff4gf"] Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.262576 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84fb747687-wsb8c"] Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.297248 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.310714 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84fb747687-wsb8c"] Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.327827 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fx6cc\" (UniqueName: \"kubernetes.io/projected/08a8f08a-1eb6-424c-874d-15efb5083a8d-kube-api-access-fx6cc\") pod \"nova-metadata-0\" (UID: \"08a8f08a-1eb6-424c-874d-15efb5083a8d\") " pod="openstack/nova-metadata-0" Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.327886 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08a8f08a-1eb6-424c-874d-15efb5083a8d-logs\") pod \"nova-metadata-0\" (UID: \"08a8f08a-1eb6-424c-874d-15efb5083a8d\") " pod="openstack/nova-metadata-0" Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.327956 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08a8f08a-1eb6-424c-874d-15efb5083a8d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"08a8f08a-1eb6-424c-874d-15efb5083a8d\") " pod="openstack/nova-metadata-0" Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.328001 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08a8f08a-1eb6-424c-874d-15efb5083a8d-config-data\") pod \"nova-metadata-0\" (UID: \"08a8f08a-1eb6-424c-874d-15efb5083a8d\") " pod="openstack/nova-metadata-0" Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.328036 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/08a8f08a-1eb6-424c-874d-15efb5083a8d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"08a8f08a-1eb6-424c-874d-15efb5083a8d\") " pod="openstack/nova-metadata-0" Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.329059 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08a8f08a-1eb6-424c-874d-15efb5083a8d-logs\") pod \"nova-metadata-0\" (UID: \"08a8f08a-1eb6-424c-874d-15efb5083a8d\") " pod="openstack/nova-metadata-0" Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.332478 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08a8f08a-1eb6-424c-874d-15efb5083a8d-config-data\") pod \"nova-metadata-0\" (UID: \"08a8f08a-1eb6-424c-874d-15efb5083a8d\") " pod="openstack/nova-metadata-0" Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.333137 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08a8f08a-1eb6-424c-874d-15efb5083a8d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"08a8f08a-1eb6-424c-874d-15efb5083a8d\") " pod="openstack/nova-metadata-0" Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.345241 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/08a8f08a-1eb6-424c-874d-15efb5083a8d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"08a8f08a-1eb6-424c-874d-15efb5083a8d\") " pod="openstack/nova-metadata-0" Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.345311 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.357263 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fx6cc\" (UniqueName: \"kubernetes.io/projected/08a8f08a-1eb6-424c-874d-15efb5083a8d-kube-api-access-fx6cc\") pod \"nova-metadata-0\" (UID: \"08a8f08a-1eb6-424c-874d-15efb5083a8d\") " pod="openstack/nova-metadata-0" Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.497403 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 00:49:33 crc kubenswrapper[4766]: E1126 00:49:33.571119 4766 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="00f3b60244d89847b3a53fd77058301dc3ec73e6fc1ebe02e2044b53df8adf65" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 26 00:49:33 crc kubenswrapper[4766]: E1126 00:49:33.585158 4766 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="00f3b60244d89847b3a53fd77058301dc3ec73e6fc1ebe02e2044b53df8adf65" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 26 00:49:33 crc kubenswrapper[4766]: E1126 00:49:33.597152 4766 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="00f3b60244d89847b3a53fd77058301dc3ec73e6fc1ebe02e2044b53df8adf65" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 26 00:49:33 crc kubenswrapper[4766]: E1126 00:49:33.597219 4766 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab" containerName="nova-scheduler-scheduler" Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.653359 4766 generic.go:334] "Generic (PLEG): container finished" podID="66040b18-6e3b-43dc-8d90-a3eeab8f31df" containerID="252626f0723e2a175f72d821ec6908dff31e1409f3ca28fd691e29d2869b2be6" exitCode=143 Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.653443 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"66040b18-6e3b-43dc-8d90-a3eeab8f31df","Type":"ContainerDied","Data":"252626f0723e2a175f72d821ec6908dff31e1409f3ca28fd691e29d2869b2be6"} Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.658580 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a700ac64-3ee9-4e7b-8da8-1cb7537fee78","Type":"ContainerStarted","Data":"0baae3337ff2378cad55ee8d10b2430eb002d0f4b29cd867814ff8af885af6db"} Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.660640 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" event={"ID":"90d82ce9-7909-40ac-9c8f-b15b6f68f24f","Type":"ContainerStarted","Data":"0e66e7a89154eecc5b3e106b8cbb55b12a241e065600bc55833aaef92382d7f2"} Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.661977 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"51bb050a-0106-4f7c-acf2-8c6c19f93d39","Type":"ContainerStarted","Data":"6ab076471d79aee42d7d6a49f766d9c1e4bad13028cd3c527756eae0224d84c0"} Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.664891 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4cf2f787-c351-49ce-a347-dc0975ea8bc4","Type":"ContainerStarted","Data":"936be501855d9d0a717d88938e2fce70865de323730ebf2f76c8c6a8e138114f"} Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.715459 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.715442536 podStartE2EDuration="3.715442536s" podCreationTimestamp="2025-11-26 00:49:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:49:33.699387665 +0000 UTC m=+1554.548158095" watchObservedRunningTime="2025-11-26 00:49:33.715442536 +0000 UTC m=+1554.564212966" Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.785445 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-fr5xr" podUID="580ecfc1-6334-44d3-a833-a68f98364860" containerName="registry-server" probeResult="failure" output=< Nov 26 00:49:33 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 00:49:33 crc kubenswrapper[4766]: > Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.886556 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69d94442-932b-417d-8857-f17acfc62174" path="/var/lib/kubelet/pods/69d94442-932b-417d-8857-f17acfc62174/volumes" Nov 26 00:49:33 crc kubenswrapper[4766]: I1126 00:49:33.887857 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87b59634-29ce-4283-a280-ebd27426348a" path="/var/lib/kubelet/pods/87b59634-29ce-4283-a280-ebd27426348a/volumes" Nov 26 00:49:34 crc kubenswrapper[4766]: W1126 00:49:34.159821 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08a8f08a_1eb6_424c_874d_15efb5083a8d.slice/crio-06f16064bc03bdcf4a82b6d3e08ea8a397f73de53afd6363f86fc7be4f194c62 WatchSource:0}: Error finding container 06f16064bc03bdcf4a82b6d3e08ea8a397f73de53afd6363f86fc7be4f194c62: Status 404 returned error can't find the container with id 06f16064bc03bdcf4a82b6d3e08ea8a397f73de53afd6363f86fc7be4f194c62 Nov 26 00:49:34 crc kubenswrapper[4766]: I1126 00:49:34.170587 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 00:49:34 crc kubenswrapper[4766]: I1126 00:49:34.679733 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a700ac64-3ee9-4e7b-8da8-1cb7537fee78","Type":"ContainerStarted","Data":"7dc705d83ebcb4e3756d32d6e1b0998156c1613b6b0965058cd3d4673a608aee"} Nov 26 00:49:34 crc kubenswrapper[4766]: I1126 00:49:34.680047 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 26 00:49:34 crc kubenswrapper[4766]: I1126 00:49:34.685326 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08a8f08a-1eb6-424c-874d-15efb5083a8d","Type":"ContainerStarted","Data":"a15f6b81a4cb1d4b9585ef657ff231e5573f9028d2c4e0eab7f21eaaafaffdad"} Nov 26 00:49:34 crc kubenswrapper[4766]: I1126 00:49:34.685360 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08a8f08a-1eb6-424c-874d-15efb5083a8d","Type":"ContainerStarted","Data":"06f16064bc03bdcf4a82b6d3e08ea8a397f73de53afd6363f86fc7be4f194c62"} Nov 26 00:49:34 crc kubenswrapper[4766]: I1126 00:49:34.687559 4766 generic.go:334] "Generic (PLEG): container finished" podID="90d82ce9-7909-40ac-9c8f-b15b6f68f24f" containerID="64ee6e6eb224e1d80393c27b907b09bb3ed85df968b78774ec928cde0defffcb" exitCode=0 Nov 26 00:49:34 crc kubenswrapper[4766]: I1126 00:49:34.687644 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" event={"ID":"90d82ce9-7909-40ac-9c8f-b15b6f68f24f","Type":"ContainerDied","Data":"64ee6e6eb224e1d80393c27b907b09bb3ed85df968b78774ec928cde0defffcb"} Nov 26 00:49:34 crc kubenswrapper[4766]: I1126 00:49:34.690156 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"51bb050a-0106-4f7c-acf2-8c6c19f93d39","Type":"ContainerStarted","Data":"ab820d05183b80791b197241b403e25e3744486138b27691a808fc4526e0c43b"} Nov 26 00:49:34 crc kubenswrapper[4766]: I1126 00:49:34.707154 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.2244220869999998 podStartE2EDuration="3.707132164s" podCreationTimestamp="2025-11-26 00:49:31 +0000 UTC" firstStartedPulling="2025-11-26 00:49:33.222213046 +0000 UTC m=+1554.070983476" lastFinishedPulling="2025-11-26 00:49:33.704923123 +0000 UTC m=+1554.553693553" observedRunningTime="2025-11-26 00:49:34.700496559 +0000 UTC m=+1555.549266989" watchObservedRunningTime="2025-11-26 00:49:34.707132164 +0000 UTC m=+1555.555902594" Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.064680 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=4.328225683 podStartE2EDuration="5.064664252s" podCreationTimestamp="2025-11-26 00:49:30 +0000 UTC" firstStartedPulling="2025-11-26 00:49:33.017188438 +0000 UTC m=+1553.865958868" lastFinishedPulling="2025-11-26 00:49:33.753627007 +0000 UTC m=+1554.602397437" observedRunningTime="2025-11-26 00:49:34.769853727 +0000 UTC m=+1555.618624157" watchObservedRunningTime="2025-11-26 00:49:35.064664252 +0000 UTC m=+1555.913434682" Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.077645 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rwk26"] Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.078001 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rwk26" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" containerName="registry-server" containerID="cri-o://bceafd245f3339232ecc96719a07509eae7eaa5d91b5564c4c361ab21fbea67c" gracePeriod=2 Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.702095 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" event={"ID":"90d82ce9-7909-40ac-9c8f-b15b6f68f24f","Type":"ContainerStarted","Data":"e0fc6733b38804513dacda343aa208bb447365560aaa4baf35e5bbe9390d787a"} Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.702417 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.704944 4766 generic.go:334] "Generic (PLEG): container finished" podID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" containerID="bceafd245f3339232ecc96719a07509eae7eaa5d91b5564c4c361ab21fbea67c" exitCode=0 Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.705022 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwk26" event={"ID":"9786e879-74cf-4b41-8d2f-482ba9e8d54f","Type":"ContainerDied","Data":"bceafd245f3339232ecc96719a07509eae7eaa5d91b5564c4c361ab21fbea67c"} Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.705050 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwk26" event={"ID":"9786e879-74cf-4b41-8d2f-482ba9e8d54f","Type":"ContainerDied","Data":"3f38432d89f6eb0024cd341789f516d200590be31b0c40465d29d2246d4c5698"} Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.705066 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f38432d89f6eb0024cd341789f516d200590be31b0c40465d29d2246d4c5698" Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.707516 4766 generic.go:334] "Generic (PLEG): container finished" podID="638a59e9-68c3-43cf-ae3b-4510e7df42ec" containerID="89528903de6bdb75b16ed3af567fb4d84f8c699933dfe36d9f49023cb06484b2" exitCode=0 Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.707567 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-mfr52" event={"ID":"638a59e9-68c3-43cf-ae3b-4510e7df42ec","Type":"ContainerDied","Data":"89528903de6bdb75b16ed3af567fb4d84f8c699933dfe36d9f49023cb06484b2"} Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.710976 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08a8f08a-1eb6-424c-874d-15efb5083a8d","Type":"ContainerStarted","Data":"b76d44319915d6e4df3307e2d68144e17141e42e5fbcc5ddfb665e3e09cb5a09"} Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.739737 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" podStartSLOduration=5.7396421 podStartE2EDuration="5.7396421s" podCreationTimestamp="2025-11-26 00:49:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:49:35.723597961 +0000 UTC m=+1556.572368391" watchObservedRunningTime="2025-11-26 00:49:35.7396421 +0000 UTC m=+1556.588412530" Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.746832 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rwk26" Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.784390 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.784362965 podStartE2EDuration="3.784362965s" podCreationTimestamp="2025-11-26 00:49:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:49:35.760573142 +0000 UTC m=+1556.609343572" watchObservedRunningTime="2025-11-26 00:49:35.784362965 +0000 UTC m=+1556.633133395" Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.805801 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.831986 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9786e879-74cf-4b41-8d2f-482ba9e8d54f-utilities\") pod \"9786e879-74cf-4b41-8d2f-482ba9e8d54f\" (UID: \"9786e879-74cf-4b41-8d2f-482ba9e8d54f\") " Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.832047 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwjbj\" (UniqueName: \"kubernetes.io/projected/9786e879-74cf-4b41-8d2f-482ba9e8d54f-kube-api-access-xwjbj\") pod \"9786e879-74cf-4b41-8d2f-482ba9e8d54f\" (UID: \"9786e879-74cf-4b41-8d2f-482ba9e8d54f\") " Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.832093 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9786e879-74cf-4b41-8d2f-482ba9e8d54f-catalog-content\") pod \"9786e879-74cf-4b41-8d2f-482ba9e8d54f\" (UID: \"9786e879-74cf-4b41-8d2f-482ba9e8d54f\") " Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.833978 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9786e879-74cf-4b41-8d2f-482ba9e8d54f-utilities" (OuterVolumeSpecName: "utilities") pod "9786e879-74cf-4b41-8d2f-482ba9e8d54f" (UID: "9786e879-74cf-4b41-8d2f-482ba9e8d54f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.840831 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9786e879-74cf-4b41-8d2f-482ba9e8d54f-kube-api-access-xwjbj" (OuterVolumeSpecName: "kube-api-access-xwjbj") pod "9786e879-74cf-4b41-8d2f-482ba9e8d54f" (UID: "9786e879-74cf-4b41-8d2f-482ba9e8d54f"). InnerVolumeSpecName "kube-api-access-xwjbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.925116 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9786e879-74cf-4b41-8d2f-482ba9e8d54f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9786e879-74cf-4b41-8d2f-482ba9e8d54f" (UID: "9786e879-74cf-4b41-8d2f-482ba9e8d54f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.936041 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwjbj\" (UniqueName: \"kubernetes.io/projected/9786e879-74cf-4b41-8d2f-482ba9e8d54f-kube-api-access-xwjbj\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.936084 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9786e879-74cf-4b41-8d2f-482ba9e8d54f-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:35 crc kubenswrapper[4766]: I1126 00:49:35.936099 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9786e879-74cf-4b41-8d2f-482ba9e8d54f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.151012 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.151439 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" containerName="sg-core" containerID="cri-o://d03a7c6e516f1357e128e452e362980fce2050b94ca9696ba2c21ac109e2698f" gracePeriod=30 Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.151469 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" containerName="ceilometer-notification-agent" containerID="cri-o://f1318ae2848a1d1d3ada1d25bf19bfc50609a50cc494c97f8ff30704a752b896" gracePeriod=30 Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.151501 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" containerName="proxy-httpd" containerID="cri-o://5016c3aa7d56fbe841e4e97c9618cfd6e543eefc0a230562e677ff7b237bebdf" gracePeriod=30 Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.151330 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" containerName="ceilometer-central-agent" containerID="cri-o://772ef3e6ad9388bd5ffed5f4e3e8599d67b0cb08ccd470bc826d64dfa68e21b3" gracePeriod=30 Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.749775 4766 generic.go:334] "Generic (PLEG): container finished" podID="8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab" containerID="00f3b60244d89847b3a53fd77058301dc3ec73e6fc1ebe02e2044b53df8adf65" exitCode=0 Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.749968 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab","Type":"ContainerDied","Data":"00f3b60244d89847b3a53fd77058301dc3ec73e6fc1ebe02e2044b53df8adf65"} Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.757740 4766 generic.go:334] "Generic (PLEG): container finished" podID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" containerID="5016c3aa7d56fbe841e4e97c9618cfd6e543eefc0a230562e677ff7b237bebdf" exitCode=0 Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.757778 4766 generic.go:334] "Generic (PLEG): container finished" podID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" containerID="d03a7c6e516f1357e128e452e362980fce2050b94ca9696ba2c21ac109e2698f" exitCode=2 Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.757791 4766 generic.go:334] "Generic (PLEG): container finished" podID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" containerID="772ef3e6ad9388bd5ffed5f4e3e8599d67b0cb08ccd470bc826d64dfa68e21b3" exitCode=0 Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.757847 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdd2abaa-1d30-4422-80c9-496e89a9aad9","Type":"ContainerDied","Data":"5016c3aa7d56fbe841e4e97c9618cfd6e543eefc0a230562e677ff7b237bebdf"} Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.757891 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdd2abaa-1d30-4422-80c9-496e89a9aad9","Type":"ContainerDied","Data":"d03a7c6e516f1357e128e452e362980fce2050b94ca9696ba2c21ac109e2698f"} Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.757903 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdd2abaa-1d30-4422-80c9-496e89a9aad9","Type":"ContainerDied","Data":"772ef3e6ad9388bd5ffed5f4e3e8599d67b0cb08ccd470bc826d64dfa68e21b3"} Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.761221 4766 generic.go:334] "Generic (PLEG): container finished" podID="66040b18-6e3b-43dc-8d90-a3eeab8f31df" containerID="063c2b5642c0debff8bdd6d1fb44f28d827eb58ca69c0244076b2830713167a0" exitCode=0 Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.761373 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"66040b18-6e3b-43dc-8d90-a3eeab8f31df","Type":"ContainerDied","Data":"063c2b5642c0debff8bdd6d1fb44f28d827eb58ca69c0244076b2830713167a0"} Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.761419 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"66040b18-6e3b-43dc-8d90-a3eeab8f31df","Type":"ContainerDied","Data":"57e358d9aab099e9ca0ab3973018cece1b324a254e3cd8f33a3d2b3de9899eef"} Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.761438 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57e358d9aab099e9ca0ab3973018cece1b324a254e3cd8f33a3d2b3de9899eef" Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.761513 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rwk26" Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.871700 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.880279 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rwk26"] Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.915032 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rwk26"] Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.939330 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ljrdw" Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.967017 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xdmn\" (UniqueName: \"kubernetes.io/projected/66040b18-6e3b-43dc-8d90-a3eeab8f31df-kube-api-access-4xdmn\") pod \"66040b18-6e3b-43dc-8d90-a3eeab8f31df\" (UID: \"66040b18-6e3b-43dc-8d90-a3eeab8f31df\") " Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.967110 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66040b18-6e3b-43dc-8d90-a3eeab8f31df-config-data\") pod \"66040b18-6e3b-43dc-8d90-a3eeab8f31df\" (UID: \"66040b18-6e3b-43dc-8d90-a3eeab8f31df\") " Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.967136 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66040b18-6e3b-43dc-8d90-a3eeab8f31df-logs\") pod \"66040b18-6e3b-43dc-8d90-a3eeab8f31df\" (UID: \"66040b18-6e3b-43dc-8d90-a3eeab8f31df\") " Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.967212 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66040b18-6e3b-43dc-8d90-a3eeab8f31df-combined-ca-bundle\") pod \"66040b18-6e3b-43dc-8d90-a3eeab8f31df\" (UID: \"66040b18-6e3b-43dc-8d90-a3eeab8f31df\") " Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.969173 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66040b18-6e3b-43dc-8d90-a3eeab8f31df-logs" (OuterVolumeSpecName: "logs") pod "66040b18-6e3b-43dc-8d90-a3eeab8f31df" (UID: "66040b18-6e3b-43dc-8d90-a3eeab8f31df"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.976205 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66040b18-6e3b-43dc-8d90-a3eeab8f31df-kube-api-access-4xdmn" (OuterVolumeSpecName: "kube-api-access-4xdmn") pod "66040b18-6e3b-43dc-8d90-a3eeab8f31df" (UID: "66040b18-6e3b-43dc-8d90-a3eeab8f31df"). InnerVolumeSpecName "kube-api-access-4xdmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:36 crc kubenswrapper[4766]: I1126 00:49:36.998918 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ljrdw" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.036704 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66040b18-6e3b-43dc-8d90-a3eeab8f31df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66040b18-6e3b-43dc-8d90-a3eeab8f31df" (UID: "66040b18-6e3b-43dc-8d90-a3eeab8f31df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.054747 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66040b18-6e3b-43dc-8d90-a3eeab8f31df-config-data" (OuterVolumeSpecName: "config-data") pod "66040b18-6e3b-43dc-8d90-a3eeab8f31df" (UID: "66040b18-6e3b-43dc-8d90-a3eeab8f31df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.069749 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xdmn\" (UniqueName: \"kubernetes.io/projected/66040b18-6e3b-43dc-8d90-a3eeab8f31df-kube-api-access-4xdmn\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.069781 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66040b18-6e3b-43dc-8d90-a3eeab8f31df-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.069792 4766 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66040b18-6e3b-43dc-8d90-a3eeab8f31df-logs\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.069801 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66040b18-6e3b-43dc-8d90-a3eeab8f31df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.283382 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.218:3000/\": dial tcp 10.217.0.218:3000: connect: connection refused" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.304456 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.313372 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-mfr52" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.387360 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab-config-data\") pod \"8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab\" (UID: \"8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab\") " Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.387663 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zczv\" (UniqueName: \"kubernetes.io/projected/8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab-kube-api-access-6zczv\") pod \"8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab\" (UID: \"8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab\") " Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.387703 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab-combined-ca-bundle\") pod \"8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab\" (UID: \"8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab\") " Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.390905 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab-kube-api-access-6zczv" (OuterVolumeSpecName: "kube-api-access-6zczv") pod "8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab" (UID: "8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab"). InnerVolumeSpecName "kube-api-access-6zczv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.447134 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab" (UID: "8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.449384 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab-config-data" (OuterVolumeSpecName: "config-data") pod "8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab" (UID: "8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.490633 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l486f\" (UniqueName: \"kubernetes.io/projected/638a59e9-68c3-43cf-ae3b-4510e7df42ec-kube-api-access-l486f\") pod \"638a59e9-68c3-43cf-ae3b-4510e7df42ec\" (UID: \"638a59e9-68c3-43cf-ae3b-4510e7df42ec\") " Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.490743 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/638a59e9-68c3-43cf-ae3b-4510e7df42ec-scripts\") pod \"638a59e9-68c3-43cf-ae3b-4510e7df42ec\" (UID: \"638a59e9-68c3-43cf-ae3b-4510e7df42ec\") " Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.491143 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/638a59e9-68c3-43cf-ae3b-4510e7df42ec-combined-ca-bundle\") pod \"638a59e9-68c3-43cf-ae3b-4510e7df42ec\" (UID: \"638a59e9-68c3-43cf-ae3b-4510e7df42ec\") " Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.491208 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/638a59e9-68c3-43cf-ae3b-4510e7df42ec-config-data\") pod \"638a59e9-68c3-43cf-ae3b-4510e7df42ec\" (UID: \"638a59e9-68c3-43cf-ae3b-4510e7df42ec\") " Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.501934 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.501967 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zczv\" (UniqueName: \"kubernetes.io/projected/8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab-kube-api-access-6zczv\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.501984 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.510862 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/638a59e9-68c3-43cf-ae3b-4510e7df42ec-kube-api-access-l486f" (OuterVolumeSpecName: "kube-api-access-l486f") pod "638a59e9-68c3-43cf-ae3b-4510e7df42ec" (UID: "638a59e9-68c3-43cf-ae3b-4510e7df42ec"). InnerVolumeSpecName "kube-api-access-l486f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.543037 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/638a59e9-68c3-43cf-ae3b-4510e7df42ec-scripts" (OuterVolumeSpecName: "scripts") pod "638a59e9-68c3-43cf-ae3b-4510e7df42ec" (UID: "638a59e9-68c3-43cf-ae3b-4510e7df42ec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.558338 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/638a59e9-68c3-43cf-ae3b-4510e7df42ec-config-data" (OuterVolumeSpecName: "config-data") pod "638a59e9-68c3-43cf-ae3b-4510e7df42ec" (UID: "638a59e9-68c3-43cf-ae3b-4510e7df42ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.581813 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/638a59e9-68c3-43cf-ae3b-4510e7df42ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "638a59e9-68c3-43cf-ae3b-4510e7df42ec" (UID: "638a59e9-68c3-43cf-ae3b-4510e7df42ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.603798 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/638a59e9-68c3-43cf-ae3b-4510e7df42ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.603848 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/638a59e9-68c3-43cf-ae3b-4510e7df42ec-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.603861 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l486f\" (UniqueName: \"kubernetes.io/projected/638a59e9-68c3-43cf-ae3b-4510e7df42ec-kube-api-access-l486f\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.603876 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/638a59e9-68c3-43cf-ae3b-4510e7df42ec-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.772757 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab","Type":"ContainerDied","Data":"689a90d00fef65123c6f346a7fdeaae541d711e63c450a3a218d0737bf367dc3"} Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.772807 4766 scope.go:117] "RemoveContainer" containerID="00f3b60244d89847b3a53fd77058301dc3ec73e6fc1ebe02e2044b53df8adf65" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.772944 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.776386 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-mfr52" event={"ID":"638a59e9-68c3-43cf-ae3b-4510e7df42ec","Type":"ContainerDied","Data":"7fb6d2a387f1b4b8391caf5d7fe480d89f001fb0cb3fb6782f708142812f530f"} Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.776422 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-mfr52" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.776438 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fb6d2a387f1b4b8391caf5d7fe480d89f001fb0cb3fb6782f708142812f530f" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.776425 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.865896 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" path="/var/lib/kubelet/pods/9786e879-74cf-4b41-8d2f-482ba9e8d54f/volumes" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.866935 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.866965 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.866981 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 00:49:37 crc kubenswrapper[4766]: E1126 00:49:37.867387 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="638a59e9-68c3-43cf-ae3b-4510e7df42ec" containerName="aodh-db-sync" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.867408 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="638a59e9-68c3-43cf-ae3b-4510e7df42ec" containerName="aodh-db-sync" Nov 26 00:49:37 crc kubenswrapper[4766]: E1126 00:49:37.867425 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab" containerName="nova-scheduler-scheduler" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.867432 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab" containerName="nova-scheduler-scheduler" Nov 26 00:49:37 crc kubenswrapper[4766]: E1126 00:49:37.867442 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66040b18-6e3b-43dc-8d90-a3eeab8f31df" containerName="nova-api-api" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.867450 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="66040b18-6e3b-43dc-8d90-a3eeab8f31df" containerName="nova-api-api" Nov 26 00:49:37 crc kubenswrapper[4766]: E1126 00:49:37.867467 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" containerName="extract-content" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.867473 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" containerName="extract-content" Nov 26 00:49:37 crc kubenswrapper[4766]: E1126 00:49:37.867503 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" containerName="extract-utilities" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.867508 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" containerName="extract-utilities" Nov 26 00:49:37 crc kubenswrapper[4766]: E1126 00:49:37.867521 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66040b18-6e3b-43dc-8d90-a3eeab8f31df" containerName="nova-api-log" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.867527 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="66040b18-6e3b-43dc-8d90-a3eeab8f31df" containerName="nova-api-log" Nov 26 00:49:37 crc kubenswrapper[4766]: E1126 00:49:37.867543 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" containerName="registry-server" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.867549 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" containerName="registry-server" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.870847 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="66040b18-6e3b-43dc-8d90-a3eeab8f31df" containerName="nova-api-api" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.870875 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="66040b18-6e3b-43dc-8d90-a3eeab8f31df" containerName="nova-api-log" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.870900 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="9786e879-74cf-4b41-8d2f-482ba9e8d54f" containerName="registry-server" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.870920 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="638a59e9-68c3-43cf-ae3b-4510e7df42ec" containerName="aodh-db-sync" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.870931 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab" containerName="nova-scheduler-scheduler" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.872063 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.874490 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.883806 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.904646 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.918564 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.933185 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.935666 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.939086 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.939311 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.939987 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 26 00:49:37 crc kubenswrapper[4766]: I1126 00:49:37.945736 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.014161 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b43a5476-de61-4907-ba9c-3195b7b2a85b-logs\") pod \"nova-api-0\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " pod="openstack/nova-api-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.014208 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skhst\" (UniqueName: \"kubernetes.io/projected/7bc2e0a9-02f7-4089-9820-62e9e07a98b8-kube-api-access-skhst\") pod \"nova-scheduler-0\" (UID: \"7bc2e0a9-02f7-4089-9820-62e9e07a98b8\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.014245 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " pod="openstack/nova-api-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.014298 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-config-data\") pod \"nova-api-0\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " pod="openstack/nova-api-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.014315 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-public-tls-certs\") pod \"nova-api-0\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " pod="openstack/nova-api-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.014582 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bc2e0a9-02f7-4089-9820-62e9e07a98b8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7bc2e0a9-02f7-4089-9820-62e9e07a98b8\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.014711 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn754\" (UniqueName: \"kubernetes.io/projected/b43a5476-de61-4907-ba9c-3195b7b2a85b-kube-api-access-sn754\") pod \"nova-api-0\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " pod="openstack/nova-api-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.014834 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " pod="openstack/nova-api-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.014924 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bc2e0a9-02f7-4089-9820-62e9e07a98b8-config-data\") pod \"nova-scheduler-0\" (UID: \"7bc2e0a9-02f7-4089-9820-62e9e07a98b8\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.117237 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b43a5476-de61-4907-ba9c-3195b7b2a85b-logs\") pod \"nova-api-0\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " pod="openstack/nova-api-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.117290 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skhst\" (UniqueName: \"kubernetes.io/projected/7bc2e0a9-02f7-4089-9820-62e9e07a98b8-kube-api-access-skhst\") pod \"nova-scheduler-0\" (UID: \"7bc2e0a9-02f7-4089-9820-62e9e07a98b8\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.117329 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " pod="openstack/nova-api-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.117357 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-config-data\") pod \"nova-api-0\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " pod="openstack/nova-api-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.117377 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-public-tls-certs\") pod \"nova-api-0\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " pod="openstack/nova-api-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.117795 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bc2e0a9-02f7-4089-9820-62e9e07a98b8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7bc2e0a9-02f7-4089-9820-62e9e07a98b8\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.117877 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn754\" (UniqueName: \"kubernetes.io/projected/b43a5476-de61-4907-ba9c-3195b7b2a85b-kube-api-access-sn754\") pod \"nova-api-0\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " pod="openstack/nova-api-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.117795 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b43a5476-de61-4907-ba9c-3195b7b2a85b-logs\") pod \"nova-api-0\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " pod="openstack/nova-api-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.118109 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " pod="openstack/nova-api-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.118198 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bc2e0a9-02f7-4089-9820-62e9e07a98b8-config-data\") pod \"nova-scheduler-0\" (UID: \"7bc2e0a9-02f7-4089-9820-62e9e07a98b8\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.123011 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " pod="openstack/nova-api-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.123197 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-public-tls-certs\") pod \"nova-api-0\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " pod="openstack/nova-api-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.123230 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " pod="openstack/nova-api-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.128371 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-config-data\") pod \"nova-api-0\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " pod="openstack/nova-api-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.128907 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bc2e0a9-02f7-4089-9820-62e9e07a98b8-config-data\") pod \"nova-scheduler-0\" (UID: \"7bc2e0a9-02f7-4089-9820-62e9e07a98b8\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.132226 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bc2e0a9-02f7-4089-9820-62e9e07a98b8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7bc2e0a9-02f7-4089-9820-62e9e07a98b8\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.134940 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skhst\" (UniqueName: \"kubernetes.io/projected/7bc2e0a9-02f7-4089-9820-62e9e07a98b8-kube-api-access-skhst\") pod \"nova-scheduler-0\" (UID: \"7bc2e0a9-02f7-4089-9820-62e9e07a98b8\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.141763 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn754\" (UniqueName: \"kubernetes.io/projected/b43a5476-de61-4907-ba9c-3195b7b2a85b-kube-api-access-sn754\") pod \"nova-api-0\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " pod="openstack/nova-api-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.200446 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.262437 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.499294 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.499609 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.735603 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.782964 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.792904 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.798513 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-9jvhv" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.798820 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.798963 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.835576 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:49:38 crc kubenswrapper[4766]: E1126 00:49:38.836130 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.836447 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mbqtt" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.836501 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mbqtt" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.842982 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7bc2e0a9-02f7-4089-9820-62e9e07a98b8","Type":"ContainerStarted","Data":"6e85851445dbdd739e2c2b5a7154eb1835aa28d79b1aa873ea7984eefb9947d7"} Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.856061 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.870101 4766 generic.go:334] "Generic (PLEG): container finished" podID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" containerID="f1318ae2848a1d1d3ada1d25bf19bfc50609a50cc494c97f8ff30704a752b896" exitCode=0 Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.870149 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdd2abaa-1d30-4422-80c9-496e89a9aad9","Type":"ContainerDied","Data":"f1318ae2848a1d1d3ada1d25bf19bfc50609a50cc494c97f8ff30704a752b896"} Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.910317 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.954772 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0f78adf-e52e-42dc-b856-a3c8e9620e01-scripts\") pod \"aodh-0\" (UID: \"e0f78adf-e52e-42dc-b856-a3c8e9620e01\") " pod="openstack/aodh-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.955397 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbhx6\" (UniqueName: \"kubernetes.io/projected/e0f78adf-e52e-42dc-b856-a3c8e9620e01-kube-api-access-dbhx6\") pod \"aodh-0\" (UID: \"e0f78adf-e52e-42dc-b856-a3c8e9620e01\") " pod="openstack/aodh-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.955439 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f78adf-e52e-42dc-b856-a3c8e9620e01-config-data\") pod \"aodh-0\" (UID: \"e0f78adf-e52e-42dc-b856-a3c8e9620e01\") " pod="openstack/aodh-0" Nov 26 00:49:38 crc kubenswrapper[4766]: I1126 00:49:38.955473 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f78adf-e52e-42dc-b856-a3c8e9620e01-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e0f78adf-e52e-42dc-b856-a3c8e9620e01\") " pod="openstack/aodh-0" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.060611 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0f78adf-e52e-42dc-b856-a3c8e9620e01-scripts\") pod \"aodh-0\" (UID: \"e0f78adf-e52e-42dc-b856-a3c8e9620e01\") " pod="openstack/aodh-0" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.060909 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbhx6\" (UniqueName: \"kubernetes.io/projected/e0f78adf-e52e-42dc-b856-a3c8e9620e01-kube-api-access-dbhx6\") pod \"aodh-0\" (UID: \"e0f78adf-e52e-42dc-b856-a3c8e9620e01\") " pod="openstack/aodh-0" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.061047 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f78adf-e52e-42dc-b856-a3c8e9620e01-config-data\") pod \"aodh-0\" (UID: \"e0f78adf-e52e-42dc-b856-a3c8e9620e01\") " pod="openstack/aodh-0" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.061151 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f78adf-e52e-42dc-b856-a3c8e9620e01-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e0f78adf-e52e-42dc-b856-a3c8e9620e01\") " pod="openstack/aodh-0" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.069892 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0f78adf-e52e-42dc-b856-a3c8e9620e01-scripts\") pod \"aodh-0\" (UID: \"e0f78adf-e52e-42dc-b856-a3c8e9620e01\") " pod="openstack/aodh-0" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.071867 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f78adf-e52e-42dc-b856-a3c8e9620e01-config-data\") pod \"aodh-0\" (UID: \"e0f78adf-e52e-42dc-b856-a3c8e9620e01\") " pod="openstack/aodh-0" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.078422 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f78adf-e52e-42dc-b856-a3c8e9620e01-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e0f78adf-e52e-42dc-b856-a3c8e9620e01\") " pod="openstack/aodh-0" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.084379 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbhx6\" (UniqueName: \"kubernetes.io/projected/e0f78adf-e52e-42dc-b856-a3c8e9620e01-kube-api-access-dbhx6\") pod \"aodh-0\" (UID: \"e0f78adf-e52e-42dc-b856-a3c8e9620e01\") " pod="openstack/aodh-0" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.146512 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.190838 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.265874 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-sg-core-conf-yaml\") pod \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.266212 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdd2abaa-1d30-4422-80c9-496e89a9aad9-run-httpd\") pod \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.266319 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-combined-ca-bundle\") pod \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.266475 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgxgw\" (UniqueName: \"kubernetes.io/projected/bdd2abaa-1d30-4422-80c9-496e89a9aad9-kube-api-access-jgxgw\") pod \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.266593 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdd2abaa-1d30-4422-80c9-496e89a9aad9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bdd2abaa-1d30-4422-80c9-496e89a9aad9" (UID: "bdd2abaa-1d30-4422-80c9-496e89a9aad9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.267084 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdd2abaa-1d30-4422-80c9-496e89a9aad9-log-httpd\") pod \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.267126 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdd2abaa-1d30-4422-80c9-496e89a9aad9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bdd2abaa-1d30-4422-80c9-496e89a9aad9" (UID: "bdd2abaa-1d30-4422-80c9-496e89a9aad9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.267146 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-config-data\") pod \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.267172 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-scripts\") pod \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\" (UID: \"bdd2abaa-1d30-4422-80c9-496e89a9aad9\") " Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.268112 4766 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdd2abaa-1d30-4422-80c9-496e89a9aad9-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.268141 4766 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdd2abaa-1d30-4422-80c9-496e89a9aad9-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.272077 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdd2abaa-1d30-4422-80c9-496e89a9aad9-kube-api-access-jgxgw" (OuterVolumeSpecName: "kube-api-access-jgxgw") pod "bdd2abaa-1d30-4422-80c9-496e89a9aad9" (UID: "bdd2abaa-1d30-4422-80c9-496e89a9aad9"). InnerVolumeSpecName "kube-api-access-jgxgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.277097 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-scripts" (OuterVolumeSpecName: "scripts") pod "bdd2abaa-1d30-4422-80c9-496e89a9aad9" (UID: "bdd2abaa-1d30-4422-80c9-496e89a9aad9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.300792 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bdd2abaa-1d30-4422-80c9-496e89a9aad9" (UID: "bdd2abaa-1d30-4422-80c9-496e89a9aad9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.373721 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgxgw\" (UniqueName: \"kubernetes.io/projected/bdd2abaa-1d30-4422-80c9-496e89a9aad9-kube-api-access-jgxgw\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.373968 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.373978 4766 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.509211 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bdd2abaa-1d30-4422-80c9-496e89a9aad9" (UID: "bdd2abaa-1d30-4422-80c9-496e89a9aad9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.578425 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.633012 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-config-data" (OuterVolumeSpecName: "config-data") pod "bdd2abaa-1d30-4422-80c9-496e89a9aad9" (UID: "bdd2abaa-1d30-4422-80c9-496e89a9aad9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.681992 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdd2abaa-1d30-4422-80c9-496e89a9aad9-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.866837 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66040b18-6e3b-43dc-8d90-a3eeab8f31df" path="/var/lib/kubelet/pods/66040b18-6e3b-43dc-8d90-a3eeab8f31df/volumes" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.893833 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab" path="/var/lib/kubelet/pods/8ca1cd1c-801f-41d2-bf7e-fd17f87f40ab/volumes" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.895090 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.897071 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdd2abaa-1d30-4422-80c9-496e89a9aad9","Type":"ContainerDied","Data":"0a967a8f03be6a9be05cfb69b3c9a733cadd958dbc86857bda20484c0d3d6cb5"} Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.897174 4766 scope.go:117] "RemoveContainer" containerID="5016c3aa7d56fbe841e4e97c9618cfd6e543eefc0a230562e677ff7b237bebdf" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.897354 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.915360 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ljrdw"] Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.915770 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ljrdw" podUID="ed38267b-18fd-4735-a51c-a09d735b756a" containerName="registry-server" containerID="cri-o://297475e998d0cc5924523d72b30842ab3c6c55e3f765a8e00dcfdc69a8da1c67" gracePeriod=2 Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.916228 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b43a5476-de61-4907-ba9c-3195b7b2a85b","Type":"ContainerStarted","Data":"6cdcd6892b943b97af3c212902847fad1897dec736571401062d6c6a5c4af9d5"} Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.916274 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b43a5476-de61-4907-ba9c-3195b7b2a85b","Type":"ContainerStarted","Data":"bb7696f3522a08358b6b07533673e087f8c1c019fb87943dd1dbed6860eb68c8"} Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.920801 4766 scope.go:117] "RemoveContainer" containerID="d03a7c6e516f1357e128e452e362980fce2050b94ca9696ba2c21ac109e2698f" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.925930 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e0f78adf-e52e-42dc-b856-a3c8e9620e01","Type":"ContainerStarted","Data":"977695b9754c81cdc9f7bc3b4ce02122329925314c11e698a116ab09f940947c"} Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.930563 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mbqtt" podUID="ed786021-3dca-48e5-b286-f1a06dcbe58e" containerName="registry-server" probeResult="failure" output=< Nov 26 00:49:39 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 00:49:39 crc kubenswrapper[4766]: > Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.936711 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7bc2e0a9-02f7-4089-9820-62e9e07a98b8","Type":"ContainerStarted","Data":"8f83a9bc9d7da1da297b7839f2bed88a4d83028c7c989816fd2a3394ec77ceb2"} Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.964167 4766 scope.go:117] "RemoveContainer" containerID="f1318ae2848a1d1d3ada1d25bf19bfc50609a50cc494c97f8ff30704a752b896" Nov 26 00:49:39 crc kubenswrapper[4766]: I1126 00:49:39.998701 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.026311 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.026457 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.026443921 podStartE2EDuration="3.026443921s" podCreationTimestamp="2025-11-26 00:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:49:39.981241925 +0000 UTC m=+1560.830012355" watchObservedRunningTime="2025-11-26 00:49:40.026443921 +0000 UTC m=+1560.875214351" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.117157 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:49:40 crc kubenswrapper[4766]: E1126 00:49:40.117717 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" containerName="ceilometer-central-agent" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.117736 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" containerName="ceilometer-central-agent" Nov 26 00:49:40 crc kubenswrapper[4766]: E1126 00:49:40.117758 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" containerName="ceilometer-notification-agent" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.117766 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" containerName="ceilometer-notification-agent" Nov 26 00:49:40 crc kubenswrapper[4766]: E1126 00:49:40.117778 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" containerName="sg-core" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.117787 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" containerName="sg-core" Nov 26 00:49:40 crc kubenswrapper[4766]: E1126 00:49:40.117823 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" containerName="proxy-httpd" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.117830 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" containerName="proxy-httpd" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.118023 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" containerName="ceilometer-notification-agent" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.118051 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" containerName="proxy-httpd" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.118064 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" containerName="sg-core" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.118075 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" containerName="ceilometer-central-agent" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.120076 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.123363 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.123544 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.123728 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.132793 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.151003 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:49:40 crc kubenswrapper[4766]: E1126 00:49:40.190211 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceilometer-tls-certs combined-ca-bundle config-data kube-api-access-r79z6 log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="3cf74a1d-4f84-4b48-afc1-16b97f4178e0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.193708 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-scripts\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.193802 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.193866 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-run-httpd\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.193889 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-log-httpd\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.194365 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r79z6\" (UniqueName: \"kubernetes.io/projected/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-kube-api-access-r79z6\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.194390 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.194429 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-config-data\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.194462 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.206570 4766 scope.go:117] "RemoveContainer" containerID="772ef3e6ad9388bd5ffed5f4e3e8599d67b0cb08ccd470bc826d64dfa68e21b3" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.295794 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-config-data\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.295854 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.295886 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-scripts\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.295950 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.296021 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-run-httpd\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.296035 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-log-httpd\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.296081 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r79z6\" (UniqueName: \"kubernetes.io/projected/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-kube-api-access-r79z6\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.296101 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.302192 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.302778 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-run-httpd\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.302996 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-log-httpd\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.303619 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.307238 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-scripts\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.319076 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-config-data\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.328594 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r79z6\" (UniqueName: \"kubernetes.io/projected/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-kube-api-access-r79z6\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.330441 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.561065 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.589344 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ljrdw" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.730887 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-695zw\" (UniqueName: \"kubernetes.io/projected/ed38267b-18fd-4735-a51c-a09d735b756a-kube-api-access-695zw\") pod \"ed38267b-18fd-4735-a51c-a09d735b756a\" (UID: \"ed38267b-18fd-4735-a51c-a09d735b756a\") " Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.731111 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed38267b-18fd-4735-a51c-a09d735b756a-utilities\") pod \"ed38267b-18fd-4735-a51c-a09d735b756a\" (UID: \"ed38267b-18fd-4735-a51c-a09d735b756a\") " Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.731173 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed38267b-18fd-4735-a51c-a09d735b756a-catalog-content\") pod \"ed38267b-18fd-4735-a51c-a09d735b756a\" (UID: \"ed38267b-18fd-4735-a51c-a09d735b756a\") " Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.733277 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed38267b-18fd-4735-a51c-a09d735b756a-utilities" (OuterVolumeSpecName: "utilities") pod "ed38267b-18fd-4735-a51c-a09d735b756a" (UID: "ed38267b-18fd-4735-a51c-a09d735b756a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.736390 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed38267b-18fd-4735-a51c-a09d735b756a-kube-api-access-695zw" (OuterVolumeSpecName: "kube-api-access-695zw") pod "ed38267b-18fd-4735-a51c-a09d735b756a" (UID: "ed38267b-18fd-4735-a51c-a09d735b756a"). InnerVolumeSpecName "kube-api-access-695zw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.753618 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed38267b-18fd-4735-a51c-a09d735b756a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ed38267b-18fd-4735-a51c-a09d735b756a" (UID: "ed38267b-18fd-4735-a51c-a09d735b756a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.806158 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.828543 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.834114 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-695zw\" (UniqueName: \"kubernetes.io/projected/ed38267b-18fd-4735-a51c-a09d735b756a-kube-api-access-695zw\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.834162 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed38267b-18fd-4735-a51c-a09d735b756a-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.834175 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed38267b-18fd-4735-a51c-a09d735b756a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.949699 4766 generic.go:334] "Generic (PLEG): container finished" podID="ed38267b-18fd-4735-a51c-a09d735b756a" containerID="297475e998d0cc5924523d72b30842ab3c6c55e3f765a8e00dcfdc69a8da1c67" exitCode=0 Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.949755 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ljrdw" event={"ID":"ed38267b-18fd-4735-a51c-a09d735b756a","Type":"ContainerDied","Data":"297475e998d0cc5924523d72b30842ab3c6c55e3f765a8e00dcfdc69a8da1c67"} Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.949788 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ljrdw" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.949819 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ljrdw" event={"ID":"ed38267b-18fd-4735-a51c-a09d735b756a","Type":"ContainerDied","Data":"198720d9532ee700314c562036153a48578986b18764731e1bd82031317ab36d"} Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.949850 4766 scope.go:117] "RemoveContainer" containerID="297475e998d0cc5924523d72b30842ab3c6c55e3f765a8e00dcfdc69a8da1c67" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.954791 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b43a5476-de61-4907-ba9c-3195b7b2a85b","Type":"ContainerStarted","Data":"ad1db70a1729eb31b751ce658a1098e209d2020294f81089fc95e8813fc123ec"} Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.984494 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.985465 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e0f78adf-e52e-42dc-b856-a3c8e9620e01","Type":"ContainerStarted","Data":"6c5850bc3228dd1178546f12ee6f36f448853eef9f2e36ed77de97ee52952bd6"} Nov 26 00:49:40 crc kubenswrapper[4766]: I1126 00:49:40.995320 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.995305292 podStartE2EDuration="3.995305292s" podCreationTimestamp="2025-11-26 00:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:49:40.993671391 +0000 UTC m=+1561.842441811" watchObservedRunningTime="2025-11-26 00:49:40.995305292 +0000 UTC m=+1561.844075722" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.006936 4766 scope.go:117] "RemoveContainer" containerID="0aca943ab636156e086d709729b9f4b92a4f5a345b5a80bfa04e114a69d02009" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.017358 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.046833 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ljrdw"] Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.063852 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.100530 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ljrdw"] Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.142703 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-run-httpd\") pod \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.142832 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-log-httpd\") pod \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.142918 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-combined-ca-bundle\") pod \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.142948 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-config-data\") pod \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.143001 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-sg-core-conf-yaml\") pod \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.143173 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r79z6\" (UniqueName: \"kubernetes.io/projected/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-kube-api-access-r79z6\") pod \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.143236 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-ceilometer-tls-certs\") pod \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.143343 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-scripts\") pod \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\" (UID: \"3cf74a1d-4f84-4b48-afc1-16b97f4178e0\") " Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.145332 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3cf74a1d-4f84-4b48-afc1-16b97f4178e0" (UID: "3cf74a1d-4f84-4b48-afc1-16b97f4178e0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.145883 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3cf74a1d-4f84-4b48-afc1-16b97f4178e0" (UID: "3cf74a1d-4f84-4b48-afc1-16b97f4178e0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.160836 4766 scope.go:117] "RemoveContainer" containerID="d9cd1a627d0d5e8681c265f4866275225a4a95c3fc9fe5d42d7090d131edb692" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.180519 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3cf74a1d-4f84-4b48-afc1-16b97f4178e0" (UID: "3cf74a1d-4f84-4b48-afc1-16b97f4178e0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.182019 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-config-data" (OuterVolumeSpecName: "config-data") pod "3cf74a1d-4f84-4b48-afc1-16b97f4178e0" (UID: "3cf74a1d-4f84-4b48-afc1-16b97f4178e0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.182564 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3cf74a1d-4f84-4b48-afc1-16b97f4178e0" (UID: "3cf74a1d-4f84-4b48-afc1-16b97f4178e0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.182903 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "3cf74a1d-4f84-4b48-afc1-16b97f4178e0" (UID: "3cf74a1d-4f84-4b48-afc1-16b97f4178e0"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.183102 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-kube-api-access-r79z6" (OuterVolumeSpecName: "kube-api-access-r79z6") pod "3cf74a1d-4f84-4b48-afc1-16b97f4178e0" (UID: "3cf74a1d-4f84-4b48-afc1-16b97f4178e0"). InnerVolumeSpecName "kube-api-access-r79z6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.187738 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-scripts" (OuterVolumeSpecName: "scripts") pod "3cf74a1d-4f84-4b48-afc1-16b97f4178e0" (UID: "3cf74a1d-4f84-4b48-afc1-16b97f4178e0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.246342 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.246374 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.246383 4766 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.246392 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r79z6\" (UniqueName: \"kubernetes.io/projected/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-kube-api-access-r79z6\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.246405 4766 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.246412 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.246422 4766 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.246429 4766 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cf74a1d-4f84-4b48-afc1-16b97f4178e0-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.331096 4766 scope.go:117] "RemoveContainer" containerID="297475e998d0cc5924523d72b30842ab3c6c55e3f765a8e00dcfdc69a8da1c67" Nov 26 00:49:41 crc kubenswrapper[4766]: E1126 00:49:41.340041 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"297475e998d0cc5924523d72b30842ab3c6c55e3f765a8e00dcfdc69a8da1c67\": container with ID starting with 297475e998d0cc5924523d72b30842ab3c6c55e3f765a8e00dcfdc69a8da1c67 not found: ID does not exist" containerID="297475e998d0cc5924523d72b30842ab3c6c55e3f765a8e00dcfdc69a8da1c67" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.340077 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"297475e998d0cc5924523d72b30842ab3c6c55e3f765a8e00dcfdc69a8da1c67"} err="failed to get container status \"297475e998d0cc5924523d72b30842ab3c6c55e3f765a8e00dcfdc69a8da1c67\": rpc error: code = NotFound desc = could not find container \"297475e998d0cc5924523d72b30842ab3c6c55e3f765a8e00dcfdc69a8da1c67\": container with ID starting with 297475e998d0cc5924523d72b30842ab3c6c55e3f765a8e00dcfdc69a8da1c67 not found: ID does not exist" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.340097 4766 scope.go:117] "RemoveContainer" containerID="0aca943ab636156e086d709729b9f4b92a4f5a345b5a80bfa04e114a69d02009" Nov 26 00:49:41 crc kubenswrapper[4766]: E1126 00:49:41.346710 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0aca943ab636156e086d709729b9f4b92a4f5a345b5a80bfa04e114a69d02009\": container with ID starting with 0aca943ab636156e086d709729b9f4b92a4f5a345b5a80bfa04e114a69d02009 not found: ID does not exist" containerID="0aca943ab636156e086d709729b9f4b92a4f5a345b5a80bfa04e114a69d02009" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.346743 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0aca943ab636156e086d709729b9f4b92a4f5a345b5a80bfa04e114a69d02009"} err="failed to get container status \"0aca943ab636156e086d709729b9f4b92a4f5a345b5a80bfa04e114a69d02009\": rpc error: code = NotFound desc = could not find container \"0aca943ab636156e086d709729b9f4b92a4f5a345b5a80bfa04e114a69d02009\": container with ID starting with 0aca943ab636156e086d709729b9f4b92a4f5a345b5a80bfa04e114a69d02009 not found: ID does not exist" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.346762 4766 scope.go:117] "RemoveContainer" containerID="d9cd1a627d0d5e8681c265f4866275225a4a95c3fc9fe5d42d7090d131edb692" Nov 26 00:49:41 crc kubenswrapper[4766]: E1126 00:49:41.352017 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9cd1a627d0d5e8681c265f4866275225a4a95c3fc9fe5d42d7090d131edb692\": container with ID starting with d9cd1a627d0d5e8681c265f4866275225a4a95c3fc9fe5d42d7090d131edb692 not found: ID does not exist" containerID="d9cd1a627d0d5e8681c265f4866275225a4a95c3fc9fe5d42d7090d131edb692" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.352062 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9cd1a627d0d5e8681c265f4866275225a4a95c3fc9fe5d42d7090d131edb692"} err="failed to get container status \"d9cd1a627d0d5e8681c265f4866275225a4a95c3fc9fe5d42d7090d131edb692\": rpc error: code = NotFound desc = could not find container \"d9cd1a627d0d5e8681c265f4866275225a4a95c3fc9fe5d42d7090d131edb692\": container with ID starting with d9cd1a627d0d5e8681c265f4866275225a4a95c3fc9fe5d42d7090d131edb692 not found: ID does not exist" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.602676 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-dn4zp"] Nov 26 00:49:41 crc kubenswrapper[4766]: E1126 00:49:41.603194 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed38267b-18fd-4735-a51c-a09d735b756a" containerName="extract-utilities" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.603211 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed38267b-18fd-4735-a51c-a09d735b756a" containerName="extract-utilities" Nov 26 00:49:41 crc kubenswrapper[4766]: E1126 00:49:41.603237 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed38267b-18fd-4735-a51c-a09d735b756a" containerName="registry-server" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.603244 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed38267b-18fd-4735-a51c-a09d735b756a" containerName="registry-server" Nov 26 00:49:41 crc kubenswrapper[4766]: E1126 00:49:41.603258 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed38267b-18fd-4735-a51c-a09d735b756a" containerName="extract-content" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.603264 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed38267b-18fd-4735-a51c-a09d735b756a" containerName="extract-content" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.603477 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed38267b-18fd-4735-a51c-a09d735b756a" containerName="registry-server" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.604340 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dn4zp" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.606354 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.606620 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.644990 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dn4zp"] Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.784203 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4252a4a9-9a21-4006-8d77-1128fdedd601-scripts\") pod \"nova-cell1-cell-mapping-dn4zp\" (UID: \"4252a4a9-9a21-4006-8d77-1128fdedd601\") " pod="openstack/nova-cell1-cell-mapping-dn4zp" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.784385 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4252a4a9-9a21-4006-8d77-1128fdedd601-config-data\") pod \"nova-cell1-cell-mapping-dn4zp\" (UID: \"4252a4a9-9a21-4006-8d77-1128fdedd601\") " pod="openstack/nova-cell1-cell-mapping-dn4zp" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.784423 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8klff\" (UniqueName: \"kubernetes.io/projected/4252a4a9-9a21-4006-8d77-1128fdedd601-kube-api-access-8klff\") pod \"nova-cell1-cell-mapping-dn4zp\" (UID: \"4252a4a9-9a21-4006-8d77-1128fdedd601\") " pod="openstack/nova-cell1-cell-mapping-dn4zp" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.784534 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4252a4a9-9a21-4006-8d77-1128fdedd601-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dn4zp\" (UID: \"4252a4a9-9a21-4006-8d77-1128fdedd601\") " pod="openstack/nova-cell1-cell-mapping-dn4zp" Nov 26 00:49:41 crc kubenswrapper[4766]: E1126 00:49:41.800991 4766 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdd2abaa_1d30_4422_80c9_496e89a9aad9.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdd2abaa_1d30_4422_80c9_496e89a9aad9.slice/crio-0a967a8f03be6a9be05cfb69b3c9a733cadd958dbc86857bda20484c0d3d6cb5\": RecentStats: unable to find data in memory cache]" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.816797 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.848425 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdd2abaa-1d30-4422-80c9-496e89a9aad9" path="/var/lib/kubelet/pods/bdd2abaa-1d30-4422-80c9-496e89a9aad9/volumes" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.849362 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed38267b-18fd-4735-a51c-a09d735b756a" path="/var/lib/kubelet/pods/ed38267b-18fd-4735-a51c-a09d735b756a/volumes" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.887398 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4252a4a9-9a21-4006-8d77-1128fdedd601-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dn4zp\" (UID: \"4252a4a9-9a21-4006-8d77-1128fdedd601\") " pod="openstack/nova-cell1-cell-mapping-dn4zp" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.887461 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4252a4a9-9a21-4006-8d77-1128fdedd601-scripts\") pod \"nova-cell1-cell-mapping-dn4zp\" (UID: \"4252a4a9-9a21-4006-8d77-1128fdedd601\") " pod="openstack/nova-cell1-cell-mapping-dn4zp" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.887610 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4252a4a9-9a21-4006-8d77-1128fdedd601-config-data\") pod \"nova-cell1-cell-mapping-dn4zp\" (UID: \"4252a4a9-9a21-4006-8d77-1128fdedd601\") " pod="openstack/nova-cell1-cell-mapping-dn4zp" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.887640 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8klff\" (UniqueName: \"kubernetes.io/projected/4252a4a9-9a21-4006-8d77-1128fdedd601-kube-api-access-8klff\") pod \"nova-cell1-cell-mapping-dn4zp\" (UID: \"4252a4a9-9a21-4006-8d77-1128fdedd601\") " pod="openstack/nova-cell1-cell-mapping-dn4zp" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.896457 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4252a4a9-9a21-4006-8d77-1128fdedd601-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dn4zp\" (UID: \"4252a4a9-9a21-4006-8d77-1128fdedd601\") " pod="openstack/nova-cell1-cell-mapping-dn4zp" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.897749 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4252a4a9-9a21-4006-8d77-1128fdedd601-scripts\") pod \"nova-cell1-cell-mapping-dn4zp\" (UID: \"4252a4a9-9a21-4006-8d77-1128fdedd601\") " pod="openstack/nova-cell1-cell-mapping-dn4zp" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.903210 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4252a4a9-9a21-4006-8d77-1128fdedd601-config-data\") pod \"nova-cell1-cell-mapping-dn4zp\" (UID: \"4252a4a9-9a21-4006-8d77-1128fdedd601\") " pod="openstack/nova-cell1-cell-mapping-dn4zp" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.926963 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8klff\" (UniqueName: \"kubernetes.io/projected/4252a4a9-9a21-4006-8d77-1128fdedd601-kube-api-access-8klff\") pod \"nova-cell1-cell-mapping-dn4zp\" (UID: \"4252a4a9-9a21-4006-8d77-1128fdedd601\") " pod="openstack/nova-cell1-cell-mapping-dn4zp" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.930584 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dn4zp" Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.936945 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-xtpgd"] Nov 26 00:49:41 crc kubenswrapper[4766]: I1126 00:49:41.937163 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" podUID="22305d36-865a-4fe0-8337-92009c1c3c29" containerName="dnsmasq-dns" containerID="cri-o://a24b82cf93de432af5c5aaac20f2f469f4bb029ec14c1b455b23cf107a1d967d" gracePeriod=10 Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.023466 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.153218 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.299390 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.318480 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" podUID="22305d36-865a-4fe0-8337-92009c1c3c29" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.196:5353: connect: connection refused" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.332524 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.356080 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.376868 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.387247 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.387484 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.387773 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.409256 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.516080 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.516141 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62490687-08a8-4e40-a3f7-d8c64aeb5d24-log-httpd\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.516188 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.516235 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-scripts\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.516270 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-config-data\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.516326 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c64np\" (UniqueName: \"kubernetes.io/projected/62490687-08a8-4e40-a3f7-d8c64aeb5d24-kube-api-access-c64np\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.516367 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62490687-08a8-4e40-a3f7-d8c64aeb5d24-run-httpd\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.516456 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.618910 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-config-data\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.619829 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c64np\" (UniqueName: \"kubernetes.io/projected/62490687-08a8-4e40-a3f7-d8c64aeb5d24-kube-api-access-c64np\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.619951 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62490687-08a8-4e40-a3f7-d8c64aeb5d24-run-httpd\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.620103 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.620286 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.620342 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62490687-08a8-4e40-a3f7-d8c64aeb5d24-run-httpd\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.620359 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62490687-08a8-4e40-a3f7-d8c64aeb5d24-log-httpd\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.620516 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.620612 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-scripts\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.621168 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62490687-08a8-4e40-a3f7-d8c64aeb5d24-log-httpd\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.627893 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.629603 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-scripts\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.632498 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.646599 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-config-data\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.649151 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c64np\" (UniqueName: \"kubernetes.io/projected/62490687-08a8-4e40-a3f7-d8c64aeb5d24-kube-api-access-c64np\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.666555 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.709173 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:49:42 crc kubenswrapper[4766]: I1126 00:49:42.890366 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dn4zp"] Nov 26 00:49:43 crc kubenswrapper[4766]: I1126 00:49:43.029904 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 26 00:49:43 crc kubenswrapper[4766]: I1126 00:49:43.047135 4766 generic.go:334] "Generic (PLEG): container finished" podID="22305d36-865a-4fe0-8337-92009c1c3c29" containerID="a24b82cf93de432af5c5aaac20f2f469f4bb029ec14c1b455b23cf107a1d967d" exitCode=0 Nov 26 00:49:43 crc kubenswrapper[4766]: I1126 00:49:43.047433 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" event={"ID":"22305d36-865a-4fe0-8337-92009c1c3c29","Type":"ContainerDied","Data":"a24b82cf93de432af5c5aaac20f2f469f4bb029ec14c1b455b23cf107a1d967d"} Nov 26 00:49:43 crc kubenswrapper[4766]: I1126 00:49:43.201341 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 26 00:49:43 crc kubenswrapper[4766]: I1126 00:49:43.500052 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 26 00:49:43 crc kubenswrapper[4766]: I1126 00:49:43.500337 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 26 00:49:43 crc kubenswrapper[4766]: I1126 00:49:43.577590 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:49:43 crc kubenswrapper[4766]: I1126 00:49:43.661195 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-dns-swift-storage-0\") pod \"22305d36-865a-4fe0-8337-92009c1c3c29\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " Nov 26 00:49:43 crc kubenswrapper[4766]: I1126 00:49:43.661467 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-config\") pod \"22305d36-865a-4fe0-8337-92009c1c3c29\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " Nov 26 00:49:43 crc kubenswrapper[4766]: I1126 00:49:43.661903 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-ovsdbserver-nb\") pod \"22305d36-865a-4fe0-8337-92009c1c3c29\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " Nov 26 00:49:43 crc kubenswrapper[4766]: I1126 00:49:43.662025 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnd8l\" (UniqueName: \"kubernetes.io/projected/22305d36-865a-4fe0-8337-92009c1c3c29-kube-api-access-wnd8l\") pod \"22305d36-865a-4fe0-8337-92009c1c3c29\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " Nov 26 00:49:43 crc kubenswrapper[4766]: I1126 00:49:43.662049 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-dns-svc\") pod \"22305d36-865a-4fe0-8337-92009c1c3c29\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " Nov 26 00:49:43 crc kubenswrapper[4766]: I1126 00:49:43.662079 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-ovsdbserver-sb\") pod \"22305d36-865a-4fe0-8337-92009c1c3c29\" (UID: \"22305d36-865a-4fe0-8337-92009c1c3c29\") " Nov 26 00:49:43 crc kubenswrapper[4766]: I1126 00:49:43.671499 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22305d36-865a-4fe0-8337-92009c1c3c29-kube-api-access-wnd8l" (OuterVolumeSpecName: "kube-api-access-wnd8l") pod "22305d36-865a-4fe0-8337-92009c1c3c29" (UID: "22305d36-865a-4fe0-8337-92009c1c3c29"). InnerVolumeSpecName "kube-api-access-wnd8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:43 crc kubenswrapper[4766]: I1126 00:49:43.769935 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnd8l\" (UniqueName: \"kubernetes.io/projected/22305d36-865a-4fe0-8337-92009c1c3c29-kube-api-access-wnd8l\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:43 crc kubenswrapper[4766]: I1126 00:49:43.790869 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-fr5xr" podUID="580ecfc1-6334-44d3-a833-a68f98364860" containerName="registry-server" probeResult="failure" output=< Nov 26 00:49:43 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 00:49:43 crc kubenswrapper[4766]: > Nov 26 00:49:43 crc kubenswrapper[4766]: I1126 00:49:43.880014 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cf74a1d-4f84-4b48-afc1-16b97f4178e0" path="/var/lib/kubelet/pods/3cf74a1d-4f84-4b48-afc1-16b97f4178e0/volumes" Nov 26 00:49:44 crc kubenswrapper[4766]: I1126 00:49:44.088531 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "22305d36-865a-4fe0-8337-92009c1c3c29" (UID: "22305d36-865a-4fe0-8337-92009c1c3c29"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:49:44 crc kubenswrapper[4766]: I1126 00:49:44.088967 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dn4zp" event={"ID":"4252a4a9-9a21-4006-8d77-1128fdedd601","Type":"ContainerStarted","Data":"b90b0a27b26e4357ff8dc14b9a17baaaab2d5052fa961c2c602d3b18c7cfad9b"} Nov 26 00:49:44 crc kubenswrapper[4766]: I1126 00:49:44.101701 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:44 crc kubenswrapper[4766]: I1126 00:49:44.109219 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" event={"ID":"22305d36-865a-4fe0-8337-92009c1c3c29","Type":"ContainerDied","Data":"ae8f81a11891320b7c03174056ad671c03e5aeefa79dbfe2cb5db6346bf64469"} Nov 26 00:49:44 crc kubenswrapper[4766]: I1126 00:49:44.109268 4766 scope.go:117] "RemoveContainer" containerID="a24b82cf93de432af5c5aaac20f2f469f4bb029ec14c1b455b23cf107a1d967d" Nov 26 00:49:44 crc kubenswrapper[4766]: I1126 00:49:44.109424 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-xtpgd" Nov 26 00:49:44 crc kubenswrapper[4766]: I1126 00:49:44.164797 4766 scope.go:117] "RemoveContainer" containerID="a615525881a275d3591fca34860f3c0ccc52cf8df614d018dcb30b4f1d99e563" Nov 26 00:49:44 crc kubenswrapper[4766]: I1126 00:49:44.219010 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "22305d36-865a-4fe0-8337-92009c1c3c29" (UID: "22305d36-865a-4fe0-8337-92009c1c3c29"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:49:44 crc kubenswrapper[4766]: I1126 00:49:44.220147 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:49:44 crc kubenswrapper[4766]: W1126 00:49:44.236200 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62490687_08a8_4e40_a3f7_d8c64aeb5d24.slice/crio-b3e9807dba0caa3d3b46315aab907b7c60dd37cfd8fff70d37b60cc2810b2922 WatchSource:0}: Error finding container b3e9807dba0caa3d3b46315aab907b7c60dd37cfd8fff70d37b60cc2810b2922: Status 404 returned error can't find the container with id b3e9807dba0caa3d3b46315aab907b7c60dd37cfd8fff70d37b60cc2810b2922 Nov 26 00:49:44 crc kubenswrapper[4766]: I1126 00:49:44.238379 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-config" (OuterVolumeSpecName: "config") pod "22305d36-865a-4fe0-8337-92009c1c3c29" (UID: "22305d36-865a-4fe0-8337-92009c1c3c29"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:49:44 crc kubenswrapper[4766]: I1126 00:49:44.239883 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "22305d36-865a-4fe0-8337-92009c1c3c29" (UID: "22305d36-865a-4fe0-8337-92009c1c3c29"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:49:44 crc kubenswrapper[4766]: I1126 00:49:44.266208 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "22305d36-865a-4fe0-8337-92009c1c3c29" (UID: "22305d36-865a-4fe0-8337-92009c1c3c29"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:49:44 crc kubenswrapper[4766]: I1126 00:49:44.306214 4766 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:44 crc kubenswrapper[4766]: I1126 00:49:44.306249 4766 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:44 crc kubenswrapper[4766]: I1126 00:49:44.306262 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:44 crc kubenswrapper[4766]: I1126 00:49:44.306272 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/22305d36-865a-4fe0-8337-92009c1c3c29-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:44 crc kubenswrapper[4766]: I1126 00:49:44.447921 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-xtpgd"] Nov 26 00:49:44 crc kubenswrapper[4766]: I1126 00:49:44.459661 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-xtpgd"] Nov 26 00:49:44 crc kubenswrapper[4766]: I1126 00:49:44.551837 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="08a8f08a-1eb6-424c-874d-15efb5083a8d" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.240:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 26 00:49:44 crc kubenswrapper[4766]: I1126 00:49:44.551848 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="08a8f08a-1eb6-424c-874d-15efb5083a8d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.240:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 26 00:49:45 crc kubenswrapper[4766]: I1126 00:49:45.120754 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e0f78adf-e52e-42dc-b856-a3c8e9620e01","Type":"ContainerStarted","Data":"7221b55c72c50568fb76398244c357c04af3ff796ed59da4993c4dea62ae3d6e"} Nov 26 00:49:45 crc kubenswrapper[4766]: I1126 00:49:45.122901 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62490687-08a8-4e40-a3f7-d8c64aeb5d24","Type":"ContainerStarted","Data":"81a18a1a1a33fce2ef59e20de6029e3f5ecb95513bd41c361e5d3a25170b3a8c"} Nov 26 00:49:45 crc kubenswrapper[4766]: I1126 00:49:45.122974 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62490687-08a8-4e40-a3f7-d8c64aeb5d24","Type":"ContainerStarted","Data":"b3e9807dba0caa3d3b46315aab907b7c60dd37cfd8fff70d37b60cc2810b2922"} Nov 26 00:49:45 crc kubenswrapper[4766]: I1126 00:49:45.125448 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dn4zp" event={"ID":"4252a4a9-9a21-4006-8d77-1128fdedd601","Type":"ContainerStarted","Data":"92aef5fe319c6e4705cccd3f94d4735c9ef2561999c3cbc1bcb70eb61d7d4ffd"} Nov 26 00:49:45 crc kubenswrapper[4766]: I1126 00:49:45.158339 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-dn4zp" podStartSLOduration=4.158316037 podStartE2EDuration="4.158316037s" podCreationTimestamp="2025-11-26 00:49:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:49:45.14237863 +0000 UTC m=+1565.991149060" watchObservedRunningTime="2025-11-26 00:49:45.158316037 +0000 UTC m=+1566.007086467" Nov 26 00:49:45 crc kubenswrapper[4766]: I1126 00:49:45.838199 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22305d36-865a-4fe0-8337-92009c1c3c29" path="/var/lib/kubelet/pods/22305d36-865a-4fe0-8337-92009c1c3c29/volumes" Nov 26 00:49:47 crc kubenswrapper[4766]: I1126 00:49:47.171849 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62490687-08a8-4e40-a3f7-d8c64aeb5d24","Type":"ContainerStarted","Data":"6673e046e2197a4b6fac11035a178781f4e83f5c3084eaaeefdd9e783da3f0ff"} Nov 26 00:49:47 crc kubenswrapper[4766]: I1126 00:49:47.175920 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e0f78adf-e52e-42dc-b856-a3c8e9620e01","Type":"ContainerStarted","Data":"80a8950bd48d7905ed4f8f6eaa2d4b9a1abf0751bfc2a3cc0a263f28c226c0c7"} Nov 26 00:49:48 crc kubenswrapper[4766]: I1126 00:49:48.188330 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62490687-08a8-4e40-a3f7-d8c64aeb5d24","Type":"ContainerStarted","Data":"c41bd398fb0aeb4ea7712daecc1565be2bc9a263b141abc17405ebb24c8de9ce"} Nov 26 00:49:48 crc kubenswrapper[4766]: I1126 00:49:48.201692 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 26 00:49:48 crc kubenswrapper[4766]: I1126 00:49:48.237973 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 26 00:49:48 crc kubenswrapper[4766]: I1126 00:49:48.262768 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 26 00:49:48 crc kubenswrapper[4766]: I1126 00:49:48.262813 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 26 00:49:49 crc kubenswrapper[4766]: I1126 00:49:49.201538 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e0f78adf-e52e-42dc-b856-a3c8e9620e01","Type":"ContainerStarted","Data":"ccb1658dcf0b4f2f56b342f572244b3b93c706ecf897af96f191ac8d30a7d977"} Nov 26 00:49:49 crc kubenswrapper[4766]: I1126 00:49:49.201898 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="e0f78adf-e52e-42dc-b856-a3c8e9620e01" containerName="aodh-api" containerID="cri-o://6c5850bc3228dd1178546f12ee6f36f448853eef9f2e36ed77de97ee52952bd6" gracePeriod=30 Nov 26 00:49:49 crc kubenswrapper[4766]: I1126 00:49:49.202239 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="e0f78adf-e52e-42dc-b856-a3c8e9620e01" containerName="aodh-notifier" containerID="cri-o://80a8950bd48d7905ed4f8f6eaa2d4b9a1abf0751bfc2a3cc0a263f28c226c0c7" gracePeriod=30 Nov 26 00:49:49 crc kubenswrapper[4766]: I1126 00:49:49.202351 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="e0f78adf-e52e-42dc-b856-a3c8e9620e01" containerName="aodh-listener" containerID="cri-o://ccb1658dcf0b4f2f56b342f572244b3b93c706ecf897af96f191ac8d30a7d977" gracePeriod=30 Nov 26 00:49:49 crc kubenswrapper[4766]: I1126 00:49:49.202419 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="e0f78adf-e52e-42dc-b856-a3c8e9620e01" containerName="aodh-evaluator" containerID="cri-o://7221b55c72c50568fb76398244c357c04af3ff796ed59da4993c4dea62ae3d6e" gracePeriod=30 Nov 26 00:49:49 crc kubenswrapper[4766]: I1126 00:49:49.214408 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62490687-08a8-4e40-a3f7-d8c64aeb5d24","Type":"ContainerStarted","Data":"b2138ff204dd583ad9f427fd6407d82c57b26b12b9bb1433ac804293133a3552"} Nov 26 00:49:49 crc kubenswrapper[4766]: I1126 00:49:49.214446 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 26 00:49:49 crc kubenswrapper[4766]: I1126 00:49:49.265452 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.845558549 podStartE2EDuration="11.26543008s" podCreationTimestamp="2025-11-26 00:49:38 +0000 UTC" firstStartedPulling="2025-11-26 00:49:39.809344092 +0000 UTC m=+1560.658114512" lastFinishedPulling="2025-11-26 00:49:48.229215613 +0000 UTC m=+1569.077986043" observedRunningTime="2025-11-26 00:49:49.2300943 +0000 UTC m=+1570.078864720" watchObservedRunningTime="2025-11-26 00:49:49.26543008 +0000 UTC m=+1570.114200510" Nov 26 00:49:49 crc kubenswrapper[4766]: I1126 00:49:49.283905 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b43a5476-de61-4907-ba9c-3195b7b2a85b" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.242:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 26 00:49:49 crc kubenswrapper[4766]: I1126 00:49:49.284169 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b43a5476-de61-4907-ba9c-3195b7b2a85b" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.242:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 26 00:49:49 crc kubenswrapper[4766]: I1126 00:49:49.304989 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.051603399 podStartE2EDuration="7.304970925s" podCreationTimestamp="2025-11-26 00:49:42 +0000 UTC" firstStartedPulling="2025-11-26 00:49:44.251599466 +0000 UTC m=+1565.100369896" lastFinishedPulling="2025-11-26 00:49:48.504966992 +0000 UTC m=+1569.353737422" observedRunningTime="2025-11-26 00:49:49.286181347 +0000 UTC m=+1570.134951777" watchObservedRunningTime="2025-11-26 00:49:49.304970925 +0000 UTC m=+1570.153741355" Nov 26 00:49:49 crc kubenswrapper[4766]: I1126 00:49:49.326361 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 26 00:49:50 crc kubenswrapper[4766]: I1126 00:49:50.008671 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mbqtt" podUID="ed786021-3dca-48e5-b286-f1a06dcbe58e" containerName="registry-server" probeResult="failure" output=< Nov 26 00:49:50 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 00:49:50 crc kubenswrapper[4766]: > Nov 26 00:49:50 crc kubenswrapper[4766]: I1126 00:49:50.225549 4766 generic.go:334] "Generic (PLEG): container finished" podID="e0f78adf-e52e-42dc-b856-a3c8e9620e01" containerID="80a8950bd48d7905ed4f8f6eaa2d4b9a1abf0751bfc2a3cc0a263f28c226c0c7" exitCode=0 Nov 26 00:49:50 crc kubenswrapper[4766]: I1126 00:49:50.225577 4766 generic.go:334] "Generic (PLEG): container finished" podID="e0f78adf-e52e-42dc-b856-a3c8e9620e01" containerID="7221b55c72c50568fb76398244c357c04af3ff796ed59da4993c4dea62ae3d6e" exitCode=0 Nov 26 00:49:50 crc kubenswrapper[4766]: I1126 00:49:50.225585 4766 generic.go:334] "Generic (PLEG): container finished" podID="e0f78adf-e52e-42dc-b856-a3c8e9620e01" containerID="6c5850bc3228dd1178546f12ee6f36f448853eef9f2e36ed77de97ee52952bd6" exitCode=0 Nov 26 00:49:50 crc kubenswrapper[4766]: I1126 00:49:50.225630 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e0f78adf-e52e-42dc-b856-a3c8e9620e01","Type":"ContainerDied","Data":"80a8950bd48d7905ed4f8f6eaa2d4b9a1abf0751bfc2a3cc0a263f28c226c0c7"} Nov 26 00:49:50 crc kubenswrapper[4766]: I1126 00:49:50.225686 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e0f78adf-e52e-42dc-b856-a3c8e9620e01","Type":"ContainerDied","Data":"7221b55c72c50568fb76398244c357c04af3ff796ed59da4993c4dea62ae3d6e"} Nov 26 00:49:50 crc kubenswrapper[4766]: I1126 00:49:50.225696 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e0f78adf-e52e-42dc-b856-a3c8e9620e01","Type":"ContainerDied","Data":"6c5850bc3228dd1178546f12ee6f36f448853eef9f2e36ed77de97ee52952bd6"} Nov 26 00:49:52 crc kubenswrapper[4766]: E1126 00:49:52.219877 4766 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdd2abaa_1d30_4422_80c9_496e89a9aad9.slice/crio-0a967a8f03be6a9be05cfb69b3c9a733cadd958dbc86857bda20484c0d3d6cb5\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdd2abaa_1d30_4422_80c9_496e89a9aad9.slice\": RecentStats: unable to find data in memory cache]" Nov 26 00:49:52 crc kubenswrapper[4766]: I1126 00:49:52.255468 4766 generic.go:334] "Generic (PLEG): container finished" podID="4252a4a9-9a21-4006-8d77-1128fdedd601" containerID="92aef5fe319c6e4705cccd3f94d4735c9ef2561999c3cbc1bcb70eb61d7d4ffd" exitCode=0 Nov 26 00:49:52 crc kubenswrapper[4766]: I1126 00:49:52.255729 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dn4zp" event={"ID":"4252a4a9-9a21-4006-8d77-1128fdedd601","Type":"ContainerDied","Data":"92aef5fe319c6e4705cccd3f94d4735c9ef2561999c3cbc1bcb70eb61d7d4ffd"} Nov 26 00:49:52 crc kubenswrapper[4766]: I1126 00:49:52.709210 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fr5xr" Nov 26 00:49:52 crc kubenswrapper[4766]: I1126 00:49:52.767626 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fr5xr" Nov 26 00:49:52 crc kubenswrapper[4766]: I1126 00:49:52.948180 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fr5xr"] Nov 26 00:49:53 crc kubenswrapper[4766]: I1126 00:49:53.148246 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sswvf"] Nov 26 00:49:53 crc kubenswrapper[4766]: I1126 00:49:53.148487 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sswvf" podUID="6b171321-6be1-4d99-ba59-ec6e80499900" containerName="registry-server" containerID="cri-o://152ff07a3de0e95758d71246f12cd1c9dd222b2ebf5cf8ef62f79997239a0aa6" gracePeriod=2 Nov 26 00:49:53 crc kubenswrapper[4766]: I1126 00:49:53.279062 4766 generic.go:334] "Generic (PLEG): container finished" podID="6b171321-6be1-4d99-ba59-ec6e80499900" containerID="152ff07a3de0e95758d71246f12cd1c9dd222b2ebf5cf8ef62f79997239a0aa6" exitCode=0 Nov 26 00:49:53 crc kubenswrapper[4766]: I1126 00:49:53.279142 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sswvf" event={"ID":"6b171321-6be1-4d99-ba59-ec6e80499900","Type":"ContainerDied","Data":"152ff07a3de0e95758d71246f12cd1c9dd222b2ebf5cf8ef62f79997239a0aa6"} Nov 26 00:49:53 crc kubenswrapper[4766]: I1126 00:49:53.347719 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t7c6w"] Nov 26 00:49:53 crc kubenswrapper[4766]: I1126 00:49:53.348005 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t7c6w" podUID="64e2c3e5-3cf6-4310-bc0b-325d569af577" containerName="registry-server" containerID="cri-o://a57ca2c45ee39ed364cb91e5ea745c245dcfc3c7b4c1bd7d7893717ba99a5ea0" gracePeriod=2 Nov 26 00:49:53 crc kubenswrapper[4766]: E1126 00:49:53.429199 4766 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a57ca2c45ee39ed364cb91e5ea745c245dcfc3c7b4c1bd7d7893717ba99a5ea0 is running failed: container process not found" containerID="a57ca2c45ee39ed364cb91e5ea745c245dcfc3c7b4c1bd7d7893717ba99a5ea0" cmd=["grpc_health_probe","-addr=:50051"] Nov 26 00:49:53 crc kubenswrapper[4766]: E1126 00:49:53.429775 4766 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a57ca2c45ee39ed364cb91e5ea745c245dcfc3c7b4c1bd7d7893717ba99a5ea0 is running failed: container process not found" containerID="a57ca2c45ee39ed364cb91e5ea745c245dcfc3c7b4c1bd7d7893717ba99a5ea0" cmd=["grpc_health_probe","-addr=:50051"] Nov 26 00:49:53 crc kubenswrapper[4766]: E1126 00:49:53.430146 4766 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a57ca2c45ee39ed364cb91e5ea745c245dcfc3c7b4c1bd7d7893717ba99a5ea0 is running failed: container process not found" containerID="a57ca2c45ee39ed364cb91e5ea745c245dcfc3c7b4c1bd7d7893717ba99a5ea0" cmd=["grpc_health_probe","-addr=:50051"] Nov 26 00:49:53 crc kubenswrapper[4766]: E1126 00:49:53.430176 4766 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a57ca2c45ee39ed364cb91e5ea745c245dcfc3c7b4c1bd7d7893717ba99a5ea0 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-t7c6w" podUID="64e2c3e5-3cf6-4310-bc0b-325d569af577" containerName="registry-server" Nov 26 00:49:53 crc kubenswrapper[4766]: I1126 00:49:53.516138 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 26 00:49:53 crc kubenswrapper[4766]: I1126 00:49:53.519338 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 26 00:49:53 crc kubenswrapper[4766]: I1126 00:49:53.521994 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 26 00:49:53 crc kubenswrapper[4766]: I1126 00:49:53.831909 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:49:53 crc kubenswrapper[4766]: E1126 00:49:53.832328 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:49:53 crc kubenswrapper[4766]: I1126 00:49:53.833945 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sswvf" Nov 26 00:49:53 crc kubenswrapper[4766]: I1126 00:49:53.955057 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b171321-6be1-4d99-ba59-ec6e80499900-catalog-content\") pod \"6b171321-6be1-4d99-ba59-ec6e80499900\" (UID: \"6b171321-6be1-4d99-ba59-ec6e80499900\") " Nov 26 00:49:53 crc kubenswrapper[4766]: I1126 00:49:53.958824 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b171321-6be1-4d99-ba59-ec6e80499900-utilities\") pod \"6b171321-6be1-4d99-ba59-ec6e80499900\" (UID: \"6b171321-6be1-4d99-ba59-ec6e80499900\") " Nov 26 00:49:53 crc kubenswrapper[4766]: I1126 00:49:53.958886 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v78w6\" (UniqueName: \"kubernetes.io/projected/6b171321-6be1-4d99-ba59-ec6e80499900-kube-api-access-v78w6\") pod \"6b171321-6be1-4d99-ba59-ec6e80499900\" (UID: \"6b171321-6be1-4d99-ba59-ec6e80499900\") " Nov 26 00:49:53 crc kubenswrapper[4766]: I1126 00:49:53.959244 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b171321-6be1-4d99-ba59-ec6e80499900-utilities" (OuterVolumeSpecName: "utilities") pod "6b171321-6be1-4d99-ba59-ec6e80499900" (UID: "6b171321-6be1-4d99-ba59-ec6e80499900"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:49:53 crc kubenswrapper[4766]: I1126 00:49:53.959954 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b171321-6be1-4d99-ba59-ec6e80499900-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:53 crc kubenswrapper[4766]: I1126 00:49:53.967070 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b171321-6be1-4d99-ba59-ec6e80499900-kube-api-access-v78w6" (OuterVolumeSpecName: "kube-api-access-v78w6") pod "6b171321-6be1-4d99-ba59-ec6e80499900" (UID: "6b171321-6be1-4d99-ba59-ec6e80499900"). InnerVolumeSpecName "kube-api-access-v78w6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:53 crc kubenswrapper[4766]: I1126 00:49:53.982855 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t7c6w" Nov 26 00:49:53 crc kubenswrapper[4766]: I1126 00:49:53.988767 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dn4zp" Nov 26 00:49:53 crc kubenswrapper[4766]: I1126 00:49:53.997224 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b171321-6be1-4d99-ba59-ec6e80499900-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b171321-6be1-4d99-ba59-ec6e80499900" (UID: "6b171321-6be1-4d99-ba59-ec6e80499900"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.061470 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4252a4a9-9a21-4006-8d77-1128fdedd601-combined-ca-bundle\") pod \"4252a4a9-9a21-4006-8d77-1128fdedd601\" (UID: \"4252a4a9-9a21-4006-8d77-1128fdedd601\") " Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.061633 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4252a4a9-9a21-4006-8d77-1128fdedd601-config-data\") pod \"4252a4a9-9a21-4006-8d77-1128fdedd601\" (UID: \"4252a4a9-9a21-4006-8d77-1128fdedd601\") " Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.061705 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8klff\" (UniqueName: \"kubernetes.io/projected/4252a4a9-9a21-4006-8d77-1128fdedd601-kube-api-access-8klff\") pod \"4252a4a9-9a21-4006-8d77-1128fdedd601\" (UID: \"4252a4a9-9a21-4006-8d77-1128fdedd601\") " Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.061734 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cf7c2\" (UniqueName: \"kubernetes.io/projected/64e2c3e5-3cf6-4310-bc0b-325d569af577-kube-api-access-cf7c2\") pod \"64e2c3e5-3cf6-4310-bc0b-325d569af577\" (UID: \"64e2c3e5-3cf6-4310-bc0b-325d569af577\") " Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.061766 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4252a4a9-9a21-4006-8d77-1128fdedd601-scripts\") pod \"4252a4a9-9a21-4006-8d77-1128fdedd601\" (UID: \"4252a4a9-9a21-4006-8d77-1128fdedd601\") " Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.061864 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64e2c3e5-3cf6-4310-bc0b-325d569af577-utilities\") pod \"64e2c3e5-3cf6-4310-bc0b-325d569af577\" (UID: \"64e2c3e5-3cf6-4310-bc0b-325d569af577\") " Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.061944 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64e2c3e5-3cf6-4310-bc0b-325d569af577-catalog-content\") pod \"64e2c3e5-3cf6-4310-bc0b-325d569af577\" (UID: \"64e2c3e5-3cf6-4310-bc0b-325d569af577\") " Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.062756 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v78w6\" (UniqueName: \"kubernetes.io/projected/6b171321-6be1-4d99-ba59-ec6e80499900-kube-api-access-v78w6\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.062786 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b171321-6be1-4d99-ba59-ec6e80499900-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.063445 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64e2c3e5-3cf6-4310-bc0b-325d569af577-utilities" (OuterVolumeSpecName: "utilities") pod "64e2c3e5-3cf6-4310-bc0b-325d569af577" (UID: "64e2c3e5-3cf6-4310-bc0b-325d569af577"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.067303 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4252a4a9-9a21-4006-8d77-1128fdedd601-kube-api-access-8klff" (OuterVolumeSpecName: "kube-api-access-8klff") pod "4252a4a9-9a21-4006-8d77-1128fdedd601" (UID: "4252a4a9-9a21-4006-8d77-1128fdedd601"). InnerVolumeSpecName "kube-api-access-8klff". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.067422 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4252a4a9-9a21-4006-8d77-1128fdedd601-scripts" (OuterVolumeSpecName: "scripts") pod "4252a4a9-9a21-4006-8d77-1128fdedd601" (UID: "4252a4a9-9a21-4006-8d77-1128fdedd601"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.069394 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64e2c3e5-3cf6-4310-bc0b-325d569af577-kube-api-access-cf7c2" (OuterVolumeSpecName: "kube-api-access-cf7c2") pod "64e2c3e5-3cf6-4310-bc0b-325d569af577" (UID: "64e2c3e5-3cf6-4310-bc0b-325d569af577"). InnerVolumeSpecName "kube-api-access-cf7c2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.088215 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4252a4a9-9a21-4006-8d77-1128fdedd601-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4252a4a9-9a21-4006-8d77-1128fdedd601" (UID: "4252a4a9-9a21-4006-8d77-1128fdedd601"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.098978 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4252a4a9-9a21-4006-8d77-1128fdedd601-config-data" (OuterVolumeSpecName: "config-data") pod "4252a4a9-9a21-4006-8d77-1128fdedd601" (UID: "4252a4a9-9a21-4006-8d77-1128fdedd601"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.117927 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64e2c3e5-3cf6-4310-bc0b-325d569af577-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "64e2c3e5-3cf6-4310-bc0b-325d569af577" (UID: "64e2c3e5-3cf6-4310-bc0b-325d569af577"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.165152 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4252a4a9-9a21-4006-8d77-1128fdedd601-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.165475 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8klff\" (UniqueName: \"kubernetes.io/projected/4252a4a9-9a21-4006-8d77-1128fdedd601-kube-api-access-8klff\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.165489 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cf7c2\" (UniqueName: \"kubernetes.io/projected/64e2c3e5-3cf6-4310-bc0b-325d569af577-kube-api-access-cf7c2\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.165500 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4252a4a9-9a21-4006-8d77-1128fdedd601-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.165510 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64e2c3e5-3cf6-4310-bc0b-325d569af577-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.165520 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64e2c3e5-3cf6-4310-bc0b-325d569af577-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.165530 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4252a4a9-9a21-4006-8d77-1128fdedd601-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.326374 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dn4zp" event={"ID":"4252a4a9-9a21-4006-8d77-1128fdedd601","Type":"ContainerDied","Data":"b90b0a27b26e4357ff8dc14b9a17baaaab2d5052fa961c2c602d3b18c7cfad9b"} Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.326405 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b90b0a27b26e4357ff8dc14b9a17baaaab2d5052fa961c2c602d3b18c7cfad9b" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.326459 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dn4zp" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.357147 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sswvf" event={"ID":"6b171321-6be1-4d99-ba59-ec6e80499900","Type":"ContainerDied","Data":"384014c196579f784761066a0d07652f0244b28f24841eb8a210f9e7fbb92a9f"} Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.357340 4766 scope.go:117] "RemoveContainer" containerID="152ff07a3de0e95758d71246f12cd1c9dd222b2ebf5cf8ef62f79997239a0aa6" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.357518 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sswvf" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.377708 4766 generic.go:334] "Generic (PLEG): container finished" podID="64e2c3e5-3cf6-4310-bc0b-325d569af577" containerID="a57ca2c45ee39ed364cb91e5ea745c245dcfc3c7b4c1bd7d7893717ba99a5ea0" exitCode=0 Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.377829 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7c6w" event={"ID":"64e2c3e5-3cf6-4310-bc0b-325d569af577","Type":"ContainerDied","Data":"a57ca2c45ee39ed364cb91e5ea745c245dcfc3c7b4c1bd7d7893717ba99a5ea0"} Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.377877 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7c6w" event={"ID":"64e2c3e5-3cf6-4310-bc0b-325d569af577","Type":"ContainerDied","Data":"b3efcc07eebad05c660c127eb6eb054d59cd79880cdb62cefb9709a63e13fdce"} Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.377877 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fr5xr" podUID="580ecfc1-6334-44d3-a833-a68f98364860" containerName="registry-server" containerID="cri-o://5d6c296ba873fad78789f27343c9e823f2e4ea3222fcc68fd6f75fdc3d6157e6" gracePeriod=2 Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.378045 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t7c6w" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.389435 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.405848 4766 scope.go:117] "RemoveContainer" containerID="b39cf8696a92e8d2038ebd430f3d20761def0b2745b98cc2925cd613bcf8b7a1" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.465144 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sswvf"] Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.525179 4766 scope.go:117] "RemoveContainer" containerID="5d7ab06e44b293e6afb2de2cdd4c0b2d4e2349124daa7a7eb90d39d26614a03c" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.566533 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sswvf"] Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.596789 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.597071 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b43a5476-de61-4907-ba9c-3195b7b2a85b" containerName="nova-api-log" containerID="cri-o://6cdcd6892b943b97af3c212902847fad1897dec736571401062d6c6a5c4af9d5" gracePeriod=30 Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.597759 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b43a5476-de61-4907-ba9c-3195b7b2a85b" containerName="nova-api-api" containerID="cri-o://ad1db70a1729eb31b751ce658a1098e209d2020294f81089fc95e8813fc123ec" gracePeriod=30 Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.616374 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.616634 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="7bc2e0a9-02f7-4089-9820-62e9e07a98b8" containerName="nova-scheduler-scheduler" containerID="cri-o://8f83a9bc9d7da1da297b7839f2bed88a4d83028c7c989816fd2a3394ec77ceb2" gracePeriod=30 Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.682711 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t7c6w"] Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.704775 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t7c6w"] Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.708273 4766 scope.go:117] "RemoveContainer" containerID="a57ca2c45ee39ed364cb91e5ea745c245dcfc3c7b4c1bd7d7893717ba99a5ea0" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.724608 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.798573 4766 scope.go:117] "RemoveContainer" containerID="af9cae77ddaefbb2f49db9dd147e090e5b53033753a7c7d5d83d725142424bcb" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.837309 4766 scope.go:117] "RemoveContainer" containerID="07e60ff7bb673aad0dcef7603fd849a25b95bda239d813b1124eeb14b684401d" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.896880 4766 scope.go:117] "RemoveContainer" containerID="a57ca2c45ee39ed364cb91e5ea745c245dcfc3c7b4c1bd7d7893717ba99a5ea0" Nov 26 00:49:54 crc kubenswrapper[4766]: E1126 00:49:54.913382 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a57ca2c45ee39ed364cb91e5ea745c245dcfc3c7b4c1bd7d7893717ba99a5ea0\": container with ID starting with a57ca2c45ee39ed364cb91e5ea745c245dcfc3c7b4c1bd7d7893717ba99a5ea0 not found: ID does not exist" containerID="a57ca2c45ee39ed364cb91e5ea745c245dcfc3c7b4c1bd7d7893717ba99a5ea0" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.913433 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a57ca2c45ee39ed364cb91e5ea745c245dcfc3c7b4c1bd7d7893717ba99a5ea0"} err="failed to get container status \"a57ca2c45ee39ed364cb91e5ea745c245dcfc3c7b4c1bd7d7893717ba99a5ea0\": rpc error: code = NotFound desc = could not find container \"a57ca2c45ee39ed364cb91e5ea745c245dcfc3c7b4c1bd7d7893717ba99a5ea0\": container with ID starting with a57ca2c45ee39ed364cb91e5ea745c245dcfc3c7b4c1bd7d7893717ba99a5ea0 not found: ID does not exist" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.913464 4766 scope.go:117] "RemoveContainer" containerID="af9cae77ddaefbb2f49db9dd147e090e5b53033753a7c7d5d83d725142424bcb" Nov 26 00:49:54 crc kubenswrapper[4766]: E1126 00:49:54.913912 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af9cae77ddaefbb2f49db9dd147e090e5b53033753a7c7d5d83d725142424bcb\": container with ID starting with af9cae77ddaefbb2f49db9dd147e090e5b53033753a7c7d5d83d725142424bcb not found: ID does not exist" containerID="af9cae77ddaefbb2f49db9dd147e090e5b53033753a7c7d5d83d725142424bcb" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.913934 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af9cae77ddaefbb2f49db9dd147e090e5b53033753a7c7d5d83d725142424bcb"} err="failed to get container status \"af9cae77ddaefbb2f49db9dd147e090e5b53033753a7c7d5d83d725142424bcb\": rpc error: code = NotFound desc = could not find container \"af9cae77ddaefbb2f49db9dd147e090e5b53033753a7c7d5d83d725142424bcb\": container with ID starting with af9cae77ddaefbb2f49db9dd147e090e5b53033753a7c7d5d83d725142424bcb not found: ID does not exist" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.913950 4766 scope.go:117] "RemoveContainer" containerID="07e60ff7bb673aad0dcef7603fd849a25b95bda239d813b1124eeb14b684401d" Nov 26 00:49:54 crc kubenswrapper[4766]: E1126 00:49:54.914324 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07e60ff7bb673aad0dcef7603fd849a25b95bda239d813b1124eeb14b684401d\": container with ID starting with 07e60ff7bb673aad0dcef7603fd849a25b95bda239d813b1124eeb14b684401d not found: ID does not exist" containerID="07e60ff7bb673aad0dcef7603fd849a25b95bda239d813b1124eeb14b684401d" Nov 26 00:49:54 crc kubenswrapper[4766]: I1126 00:49:54.914349 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07e60ff7bb673aad0dcef7603fd849a25b95bda239d813b1124eeb14b684401d"} err="failed to get container status \"07e60ff7bb673aad0dcef7603fd849a25b95bda239d813b1124eeb14b684401d\": rpc error: code = NotFound desc = could not find container \"07e60ff7bb673aad0dcef7603fd849a25b95bda239d813b1124eeb14b684401d\": container with ID starting with 07e60ff7bb673aad0dcef7603fd849a25b95bda239d813b1124eeb14b684401d not found: ID does not exist" Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.004286 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fr5xr" Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.104714 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkmdw\" (UniqueName: \"kubernetes.io/projected/580ecfc1-6334-44d3-a833-a68f98364860-kube-api-access-nkmdw\") pod \"580ecfc1-6334-44d3-a833-a68f98364860\" (UID: \"580ecfc1-6334-44d3-a833-a68f98364860\") " Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.104777 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/580ecfc1-6334-44d3-a833-a68f98364860-utilities\") pod \"580ecfc1-6334-44d3-a833-a68f98364860\" (UID: \"580ecfc1-6334-44d3-a833-a68f98364860\") " Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.105038 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/580ecfc1-6334-44d3-a833-a68f98364860-catalog-content\") pod \"580ecfc1-6334-44d3-a833-a68f98364860\" (UID: \"580ecfc1-6334-44d3-a833-a68f98364860\") " Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.105623 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/580ecfc1-6334-44d3-a833-a68f98364860-utilities" (OuterVolumeSpecName: "utilities") pod "580ecfc1-6334-44d3-a833-a68f98364860" (UID: "580ecfc1-6334-44d3-a833-a68f98364860"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.110681 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/580ecfc1-6334-44d3-a833-a68f98364860-kube-api-access-nkmdw" (OuterVolumeSpecName: "kube-api-access-nkmdw") pod "580ecfc1-6334-44d3-a833-a68f98364860" (UID: "580ecfc1-6334-44d3-a833-a68f98364860"). InnerVolumeSpecName "kube-api-access-nkmdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.145071 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/580ecfc1-6334-44d3-a833-a68f98364860-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "580ecfc1-6334-44d3-a833-a68f98364860" (UID: "580ecfc1-6334-44d3-a833-a68f98364860"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.207278 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/580ecfc1-6334-44d3-a833-a68f98364860-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.207321 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkmdw\" (UniqueName: \"kubernetes.io/projected/580ecfc1-6334-44d3-a833-a68f98364860-kube-api-access-nkmdw\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.207336 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/580ecfc1-6334-44d3-a833-a68f98364860-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.391466 4766 generic.go:334] "Generic (PLEG): container finished" podID="580ecfc1-6334-44d3-a833-a68f98364860" containerID="5d6c296ba873fad78789f27343c9e823f2e4ea3222fcc68fd6f75fdc3d6157e6" exitCode=0 Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.391533 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fr5xr" event={"ID":"580ecfc1-6334-44d3-a833-a68f98364860","Type":"ContainerDied","Data":"5d6c296ba873fad78789f27343c9e823f2e4ea3222fcc68fd6f75fdc3d6157e6"} Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.391573 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fr5xr" Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.391870 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fr5xr" event={"ID":"580ecfc1-6334-44d3-a833-a68f98364860","Type":"ContainerDied","Data":"c856bccb94c734e57f276e5e6e355d052046c97689085e1930b4faf0f57e8104"} Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.391906 4766 scope.go:117] "RemoveContainer" containerID="5d6c296ba873fad78789f27343c9e823f2e4ea3222fcc68fd6f75fdc3d6157e6" Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.403154 4766 generic.go:334] "Generic (PLEG): container finished" podID="b43a5476-de61-4907-ba9c-3195b7b2a85b" containerID="6cdcd6892b943b97af3c212902847fad1897dec736571401062d6c6a5c4af9d5" exitCode=143 Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.403227 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b43a5476-de61-4907-ba9c-3195b7b2a85b","Type":"ContainerDied","Data":"6cdcd6892b943b97af3c212902847fad1897dec736571401062d6c6a5c4af9d5"} Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.417398 4766 scope.go:117] "RemoveContainer" containerID="3020c9120c42026824f47a3298746387e75244633107d3b7564271252dfab3ae" Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.435165 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fr5xr"] Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.448124 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fr5xr"] Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.465970 4766 scope.go:117] "RemoveContainer" containerID="33bc57a0146c86e26237df37029ee1ba4ecb162095aa4abec2eaa16321318b07" Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.561360 4766 scope.go:117] "RemoveContainer" containerID="5d6c296ba873fad78789f27343c9e823f2e4ea3222fcc68fd6f75fdc3d6157e6" Nov 26 00:49:55 crc kubenswrapper[4766]: E1126 00:49:55.561897 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d6c296ba873fad78789f27343c9e823f2e4ea3222fcc68fd6f75fdc3d6157e6\": container with ID starting with 5d6c296ba873fad78789f27343c9e823f2e4ea3222fcc68fd6f75fdc3d6157e6 not found: ID does not exist" containerID="5d6c296ba873fad78789f27343c9e823f2e4ea3222fcc68fd6f75fdc3d6157e6" Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.561931 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d6c296ba873fad78789f27343c9e823f2e4ea3222fcc68fd6f75fdc3d6157e6"} err="failed to get container status \"5d6c296ba873fad78789f27343c9e823f2e4ea3222fcc68fd6f75fdc3d6157e6\": rpc error: code = NotFound desc = could not find container \"5d6c296ba873fad78789f27343c9e823f2e4ea3222fcc68fd6f75fdc3d6157e6\": container with ID starting with 5d6c296ba873fad78789f27343c9e823f2e4ea3222fcc68fd6f75fdc3d6157e6 not found: ID does not exist" Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.561954 4766 scope.go:117] "RemoveContainer" containerID="3020c9120c42026824f47a3298746387e75244633107d3b7564271252dfab3ae" Nov 26 00:49:55 crc kubenswrapper[4766]: E1126 00:49:55.562316 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3020c9120c42026824f47a3298746387e75244633107d3b7564271252dfab3ae\": container with ID starting with 3020c9120c42026824f47a3298746387e75244633107d3b7564271252dfab3ae not found: ID does not exist" containerID="3020c9120c42026824f47a3298746387e75244633107d3b7564271252dfab3ae" Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.562344 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3020c9120c42026824f47a3298746387e75244633107d3b7564271252dfab3ae"} err="failed to get container status \"3020c9120c42026824f47a3298746387e75244633107d3b7564271252dfab3ae\": rpc error: code = NotFound desc = could not find container \"3020c9120c42026824f47a3298746387e75244633107d3b7564271252dfab3ae\": container with ID starting with 3020c9120c42026824f47a3298746387e75244633107d3b7564271252dfab3ae not found: ID does not exist" Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.562361 4766 scope.go:117] "RemoveContainer" containerID="33bc57a0146c86e26237df37029ee1ba4ecb162095aa4abec2eaa16321318b07" Nov 26 00:49:55 crc kubenswrapper[4766]: E1126 00:49:55.562878 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33bc57a0146c86e26237df37029ee1ba4ecb162095aa4abec2eaa16321318b07\": container with ID starting with 33bc57a0146c86e26237df37029ee1ba4ecb162095aa4abec2eaa16321318b07 not found: ID does not exist" containerID="33bc57a0146c86e26237df37029ee1ba4ecb162095aa4abec2eaa16321318b07" Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.562901 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33bc57a0146c86e26237df37029ee1ba4ecb162095aa4abec2eaa16321318b07"} err="failed to get container status \"33bc57a0146c86e26237df37029ee1ba4ecb162095aa4abec2eaa16321318b07\": rpc error: code = NotFound desc = could not find container \"33bc57a0146c86e26237df37029ee1ba4ecb162095aa4abec2eaa16321318b07\": container with ID starting with 33bc57a0146c86e26237df37029ee1ba4ecb162095aa4abec2eaa16321318b07 not found: ID does not exist" Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.839554 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="580ecfc1-6334-44d3-a833-a68f98364860" path="/var/lib/kubelet/pods/580ecfc1-6334-44d3-a833-a68f98364860/volumes" Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.840223 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64e2c3e5-3cf6-4310-bc0b-325d569af577" path="/var/lib/kubelet/pods/64e2c3e5-3cf6-4310-bc0b-325d569af577/volumes" Nov 26 00:49:55 crc kubenswrapper[4766]: I1126 00:49:55.842946 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b171321-6be1-4d99-ba59-ec6e80499900" path="/var/lib/kubelet/pods/6b171321-6be1-4d99-ba59-ec6e80499900/volumes" Nov 26 00:49:56 crc kubenswrapper[4766]: I1126 00:49:56.415226 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="08a8f08a-1eb6-424c-874d-15efb5083a8d" containerName="nova-metadata-log" containerID="cri-o://a15f6b81a4cb1d4b9585ef657ff231e5573f9028d2c4e0eab7f21eaaafaffdad" gracePeriod=30 Nov 26 00:49:56 crc kubenswrapper[4766]: I1126 00:49:56.415320 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="08a8f08a-1eb6-424c-874d-15efb5083a8d" containerName="nova-metadata-metadata" containerID="cri-o://b76d44319915d6e4df3307e2d68144e17141e42e5fbcc5ddfb665e3e09cb5a09" gracePeriod=30 Nov 26 00:49:57 crc kubenswrapper[4766]: I1126 00:49:57.426694 4766 generic.go:334] "Generic (PLEG): container finished" podID="08a8f08a-1eb6-424c-874d-15efb5083a8d" containerID="a15f6b81a4cb1d4b9585ef657ff231e5573f9028d2c4e0eab7f21eaaafaffdad" exitCode=143 Nov 26 00:49:57 crc kubenswrapper[4766]: I1126 00:49:57.426799 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08a8f08a-1eb6-424c-874d-15efb5083a8d","Type":"ContainerDied","Data":"a15f6b81a4cb1d4b9585ef657ff231e5573f9028d2c4e0eab7f21eaaafaffdad"} Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.201356 4766 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f83a9bc9d7da1da297b7839f2bed88a4d83028c7c989816fd2a3394ec77ceb2 is running failed: container process not found" containerID="8f83a9bc9d7da1da297b7839f2bed88a4d83028c7c989816fd2a3394ec77ceb2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.202209 4766 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f83a9bc9d7da1da297b7839f2bed88a4d83028c7c989816fd2a3394ec77ceb2 is running failed: container process not found" containerID="8f83a9bc9d7da1da297b7839f2bed88a4d83028c7c989816fd2a3394ec77ceb2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.202525 4766 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f83a9bc9d7da1da297b7839f2bed88a4d83028c7c989816fd2a3394ec77ceb2 is running failed: container process not found" containerID="8f83a9bc9d7da1da297b7839f2bed88a4d83028c7c989816fd2a3394ec77ceb2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.202562 4766 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f83a9bc9d7da1da297b7839f2bed88a4d83028c7c989816fd2a3394ec77ceb2 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="7bc2e0a9-02f7-4089-9820-62e9e07a98b8" containerName="nova-scheduler-scheduler" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.292342 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.383000 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-internal-tls-certs\") pod \"b43a5476-de61-4907-ba9c-3195b7b2a85b\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.383222 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-config-data\") pod \"b43a5476-de61-4907-ba9c-3195b7b2a85b\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.383280 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-combined-ca-bundle\") pod \"b43a5476-de61-4907-ba9c-3195b7b2a85b\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.383311 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sn754\" (UniqueName: \"kubernetes.io/projected/b43a5476-de61-4907-ba9c-3195b7b2a85b-kube-api-access-sn754\") pod \"b43a5476-de61-4907-ba9c-3195b7b2a85b\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.383343 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-public-tls-certs\") pod \"b43a5476-de61-4907-ba9c-3195b7b2a85b\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.383452 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b43a5476-de61-4907-ba9c-3195b7b2a85b-logs\") pod \"b43a5476-de61-4907-ba9c-3195b7b2a85b\" (UID: \"b43a5476-de61-4907-ba9c-3195b7b2a85b\") " Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.384753 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b43a5476-de61-4907-ba9c-3195b7b2a85b-logs" (OuterVolumeSpecName: "logs") pod "b43a5476-de61-4907-ba9c-3195b7b2a85b" (UID: "b43a5476-de61-4907-ba9c-3195b7b2a85b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.388994 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b43a5476-de61-4907-ba9c-3195b7b2a85b-kube-api-access-sn754" (OuterVolumeSpecName: "kube-api-access-sn754") pod "b43a5476-de61-4907-ba9c-3195b7b2a85b" (UID: "b43a5476-de61-4907-ba9c-3195b7b2a85b"). InnerVolumeSpecName "kube-api-access-sn754". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.411111 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.414903 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b43a5476-de61-4907-ba9c-3195b7b2a85b" (UID: "b43a5476-de61-4907-ba9c-3195b7b2a85b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.422576 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-config-data" (OuterVolumeSpecName: "config-data") pod "b43a5476-de61-4907-ba9c-3195b7b2a85b" (UID: "b43a5476-de61-4907-ba9c-3195b7b2a85b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.458866 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b43a5476-de61-4907-ba9c-3195b7b2a85b" (UID: "b43a5476-de61-4907-ba9c-3195b7b2a85b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.458955 4766 generic.go:334] "Generic (PLEG): container finished" podID="7bc2e0a9-02f7-4089-9820-62e9e07a98b8" containerID="8f83a9bc9d7da1da297b7839f2bed88a4d83028c7c989816fd2a3394ec77ceb2" exitCode=0 Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.458985 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7bc2e0a9-02f7-4089-9820-62e9e07a98b8","Type":"ContainerDied","Data":"8f83a9bc9d7da1da297b7839f2bed88a4d83028c7c989816fd2a3394ec77ceb2"} Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.459354 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7bc2e0a9-02f7-4089-9820-62e9e07a98b8","Type":"ContainerDied","Data":"6e85851445dbdd739e2c2b5a7154eb1835aa28d79b1aa873ea7984eefb9947d7"} Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.459389 4766 scope.go:117] "RemoveContainer" containerID="8f83a9bc9d7da1da297b7839f2bed88a4d83028c7c989816fd2a3394ec77ceb2" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.459020 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.461409 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b43a5476-de61-4907-ba9c-3195b7b2a85b" (UID: "b43a5476-de61-4907-ba9c-3195b7b2a85b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.463712 4766 generic.go:334] "Generic (PLEG): container finished" podID="b43a5476-de61-4907-ba9c-3195b7b2a85b" containerID="ad1db70a1729eb31b751ce658a1098e209d2020294f81089fc95e8813fc123ec" exitCode=0 Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.463755 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b43a5476-de61-4907-ba9c-3195b7b2a85b","Type":"ContainerDied","Data":"ad1db70a1729eb31b751ce658a1098e209d2020294f81089fc95e8813fc123ec"} Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.463778 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b43a5476-de61-4907-ba9c-3195b7b2a85b","Type":"ContainerDied","Data":"bb7696f3522a08358b6b07533673e087f8c1c019fb87943dd1dbed6860eb68c8"} Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.463832 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.484693 4766 scope.go:117] "RemoveContainer" containerID="8f83a9bc9d7da1da297b7839f2bed88a4d83028c7c989816fd2a3394ec77ceb2" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.484938 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bc2e0a9-02f7-4089-9820-62e9e07a98b8-config-data\") pod \"7bc2e0a9-02f7-4089-9820-62e9e07a98b8\" (UID: \"7bc2e0a9-02f7-4089-9820-62e9e07a98b8\") " Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.484984 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bc2e0a9-02f7-4089-9820-62e9e07a98b8-combined-ca-bundle\") pod \"7bc2e0a9-02f7-4089-9820-62e9e07a98b8\" (UID: \"7bc2e0a9-02f7-4089-9820-62e9e07a98b8\") " Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.485158 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skhst\" (UniqueName: \"kubernetes.io/projected/7bc2e0a9-02f7-4089-9820-62e9e07a98b8-kube-api-access-skhst\") pod \"7bc2e0a9-02f7-4089-9820-62e9e07a98b8\" (UID: \"7bc2e0a9-02f7-4089-9820-62e9e07a98b8\") " Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.485186 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f83a9bc9d7da1da297b7839f2bed88a4d83028c7c989816fd2a3394ec77ceb2\": container with ID starting with 8f83a9bc9d7da1da297b7839f2bed88a4d83028c7c989816fd2a3394ec77ceb2 not found: ID does not exist" containerID="8f83a9bc9d7da1da297b7839f2bed88a4d83028c7c989816fd2a3394ec77ceb2" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.485221 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f83a9bc9d7da1da297b7839f2bed88a4d83028c7c989816fd2a3394ec77ceb2"} err="failed to get container status \"8f83a9bc9d7da1da297b7839f2bed88a4d83028c7c989816fd2a3394ec77ceb2\": rpc error: code = NotFound desc = could not find container \"8f83a9bc9d7da1da297b7839f2bed88a4d83028c7c989816fd2a3394ec77ceb2\": container with ID starting with 8f83a9bc9d7da1da297b7839f2bed88a4d83028c7c989816fd2a3394ec77ceb2 not found: ID does not exist" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.485245 4766 scope.go:117] "RemoveContainer" containerID="ad1db70a1729eb31b751ce658a1098e209d2020294f81089fc95e8813fc123ec" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.485601 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.485633 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.485643 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sn754\" (UniqueName: \"kubernetes.io/projected/b43a5476-de61-4907-ba9c-3195b7b2a85b-kube-api-access-sn754\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.485667 4766 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.485675 4766 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b43a5476-de61-4907-ba9c-3195b7b2a85b-logs\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.485684 4766 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b43a5476-de61-4907-ba9c-3195b7b2a85b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.491893 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bc2e0a9-02f7-4089-9820-62e9e07a98b8-kube-api-access-skhst" (OuterVolumeSpecName: "kube-api-access-skhst") pod "7bc2e0a9-02f7-4089-9820-62e9e07a98b8" (UID: "7bc2e0a9-02f7-4089-9820-62e9e07a98b8"). InnerVolumeSpecName "kube-api-access-skhst". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.516486 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.537111 4766 scope.go:117] "RemoveContainer" containerID="6cdcd6892b943b97af3c212902847fad1897dec736571401062d6c6a5c4af9d5" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.544500 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bc2e0a9-02f7-4089-9820-62e9e07a98b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7bc2e0a9-02f7-4089-9820-62e9e07a98b8" (UID: "7bc2e0a9-02f7-4089-9820-62e9e07a98b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.546560 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bc2e0a9-02f7-4089-9820-62e9e07a98b8-config-data" (OuterVolumeSpecName: "config-data") pod "7bc2e0a9-02f7-4089-9820-62e9e07a98b8" (UID: "7bc2e0a9-02f7-4089-9820-62e9e07a98b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.556244 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.570620 4766 scope.go:117] "RemoveContainer" containerID="ad1db70a1729eb31b751ce658a1098e209d2020294f81089fc95e8813fc123ec" Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.571348 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad1db70a1729eb31b751ce658a1098e209d2020294f81089fc95e8813fc123ec\": container with ID starting with ad1db70a1729eb31b751ce658a1098e209d2020294f81089fc95e8813fc123ec not found: ID does not exist" containerID="ad1db70a1729eb31b751ce658a1098e209d2020294f81089fc95e8813fc123ec" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.571387 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad1db70a1729eb31b751ce658a1098e209d2020294f81089fc95e8813fc123ec"} err="failed to get container status \"ad1db70a1729eb31b751ce658a1098e209d2020294f81089fc95e8813fc123ec\": rpc error: code = NotFound desc = could not find container \"ad1db70a1729eb31b751ce658a1098e209d2020294f81089fc95e8813fc123ec\": container with ID starting with ad1db70a1729eb31b751ce658a1098e209d2020294f81089fc95e8813fc123ec not found: ID does not exist" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.571411 4766 scope.go:117] "RemoveContainer" containerID="6cdcd6892b943b97af3c212902847fad1897dec736571401062d6c6a5c4af9d5" Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.574243 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cdcd6892b943b97af3c212902847fad1897dec736571401062d6c6a5c4af9d5\": container with ID starting with 6cdcd6892b943b97af3c212902847fad1897dec736571401062d6c6a5c4af9d5 not found: ID does not exist" containerID="6cdcd6892b943b97af3c212902847fad1897dec736571401062d6c6a5c4af9d5" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.574280 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cdcd6892b943b97af3c212902847fad1897dec736571401062d6c6a5c4af9d5"} err="failed to get container status \"6cdcd6892b943b97af3c212902847fad1897dec736571401062d6c6a5c4af9d5\": rpc error: code = NotFound desc = could not find container \"6cdcd6892b943b97af3c212902847fad1897dec736571401062d6c6a5c4af9d5\": container with ID starting with 6cdcd6892b943b97af3c212902847fad1897dec736571401062d6c6a5c4af9d5 not found: ID does not exist" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.583019 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.583578 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22305d36-865a-4fe0-8337-92009c1c3c29" containerName="dnsmasq-dns" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.583597 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="22305d36-865a-4fe0-8337-92009c1c3c29" containerName="dnsmasq-dns" Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.583615 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b43a5476-de61-4907-ba9c-3195b7b2a85b" containerName="nova-api-api" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.583623 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="b43a5476-de61-4907-ba9c-3195b7b2a85b" containerName="nova-api-api" Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.583638 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64e2c3e5-3cf6-4310-bc0b-325d569af577" containerName="extract-utilities" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.583644 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="64e2c3e5-3cf6-4310-bc0b-325d569af577" containerName="extract-utilities" Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.583677 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bc2e0a9-02f7-4089-9820-62e9e07a98b8" containerName="nova-scheduler-scheduler" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.583683 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bc2e0a9-02f7-4089-9820-62e9e07a98b8" containerName="nova-scheduler-scheduler" Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.583691 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4252a4a9-9a21-4006-8d77-1128fdedd601" containerName="nova-manage" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.583697 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="4252a4a9-9a21-4006-8d77-1128fdedd601" containerName="nova-manage" Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.583715 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="580ecfc1-6334-44d3-a833-a68f98364860" containerName="extract-utilities" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.583721 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="580ecfc1-6334-44d3-a833-a68f98364860" containerName="extract-utilities" Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.583734 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22305d36-865a-4fe0-8337-92009c1c3c29" containerName="init" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.583740 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="22305d36-865a-4fe0-8337-92009c1c3c29" containerName="init" Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.583750 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b43a5476-de61-4907-ba9c-3195b7b2a85b" containerName="nova-api-log" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.583756 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="b43a5476-de61-4907-ba9c-3195b7b2a85b" containerName="nova-api-log" Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.583773 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b171321-6be1-4d99-ba59-ec6e80499900" containerName="extract-content" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.583779 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b171321-6be1-4d99-ba59-ec6e80499900" containerName="extract-content" Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.583789 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="580ecfc1-6334-44d3-a833-a68f98364860" containerName="extract-content" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.583795 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="580ecfc1-6334-44d3-a833-a68f98364860" containerName="extract-content" Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.583806 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64e2c3e5-3cf6-4310-bc0b-325d569af577" containerName="registry-server" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.583812 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="64e2c3e5-3cf6-4310-bc0b-325d569af577" containerName="registry-server" Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.583824 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64e2c3e5-3cf6-4310-bc0b-325d569af577" containerName="extract-content" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.583830 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="64e2c3e5-3cf6-4310-bc0b-325d569af577" containerName="extract-content" Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.583839 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b171321-6be1-4d99-ba59-ec6e80499900" containerName="extract-utilities" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.583846 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b171321-6be1-4d99-ba59-ec6e80499900" containerName="extract-utilities" Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.583864 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b171321-6be1-4d99-ba59-ec6e80499900" containerName="registry-server" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.583870 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b171321-6be1-4d99-ba59-ec6e80499900" containerName="registry-server" Nov 26 00:49:58 crc kubenswrapper[4766]: E1126 00:49:58.583882 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="580ecfc1-6334-44d3-a833-a68f98364860" containerName="registry-server" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.583889 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="580ecfc1-6334-44d3-a833-a68f98364860" containerName="registry-server" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.584080 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="4252a4a9-9a21-4006-8d77-1128fdedd601" containerName="nova-manage" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.584096 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="22305d36-865a-4fe0-8337-92009c1c3c29" containerName="dnsmasq-dns" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.584112 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="64e2c3e5-3cf6-4310-bc0b-325d569af577" containerName="registry-server" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.584130 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bc2e0a9-02f7-4089-9820-62e9e07a98b8" containerName="nova-scheduler-scheduler" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.584141 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="b43a5476-de61-4907-ba9c-3195b7b2a85b" containerName="nova-api-log" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.584233 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="580ecfc1-6334-44d3-a833-a68f98364860" containerName="registry-server" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.584246 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b171321-6be1-4d99-ba59-ec6e80499900" containerName="registry-server" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.584259 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="b43a5476-de61-4907-ba9c-3195b7b2a85b" containerName="nova-api-api" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.586531 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.587906 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skhst\" (UniqueName: \"kubernetes.io/projected/7bc2e0a9-02f7-4089-9820-62e9e07a98b8-kube-api-access-skhst\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.587937 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bc2e0a9-02f7-4089-9820-62e9e07a98b8-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.587951 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bc2e0a9-02f7-4089-9820-62e9e07a98b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.589777 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.589859 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.590112 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.593978 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.689662 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a93e0b93-3cde-4098-a853-ea1b3c86c323-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a93e0b93-3cde-4098-a853-ea1b3c86c323\") " pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.689785 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a93e0b93-3cde-4098-a853-ea1b3c86c323-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a93e0b93-3cde-4098-a853-ea1b3c86c323\") " pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.689825 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a93e0b93-3cde-4098-a853-ea1b3c86c323-logs\") pod \"nova-api-0\" (UID: \"a93e0b93-3cde-4098-a853-ea1b3c86c323\") " pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.689852 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a93e0b93-3cde-4098-a853-ea1b3c86c323-public-tls-certs\") pod \"nova-api-0\" (UID: \"a93e0b93-3cde-4098-a853-ea1b3c86c323\") " pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.689910 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2w9x\" (UniqueName: \"kubernetes.io/projected/a93e0b93-3cde-4098-a853-ea1b3c86c323-kube-api-access-k2w9x\") pod \"nova-api-0\" (UID: \"a93e0b93-3cde-4098-a853-ea1b3c86c323\") " pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.689932 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a93e0b93-3cde-4098-a853-ea1b3c86c323-config-data\") pod \"nova-api-0\" (UID: \"a93e0b93-3cde-4098-a853-ea1b3c86c323\") " pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.791904 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a93e0b93-3cde-4098-a853-ea1b3c86c323-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a93e0b93-3cde-4098-a853-ea1b3c86c323\") " pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.792029 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a93e0b93-3cde-4098-a853-ea1b3c86c323-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a93e0b93-3cde-4098-a853-ea1b3c86c323\") " pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.792572 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a93e0b93-3cde-4098-a853-ea1b3c86c323-logs\") pod \"nova-api-0\" (UID: \"a93e0b93-3cde-4098-a853-ea1b3c86c323\") " pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.792623 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a93e0b93-3cde-4098-a853-ea1b3c86c323-public-tls-certs\") pod \"nova-api-0\" (UID: \"a93e0b93-3cde-4098-a853-ea1b3c86c323\") " pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.792669 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2w9x\" (UniqueName: \"kubernetes.io/projected/a93e0b93-3cde-4098-a853-ea1b3c86c323-kube-api-access-k2w9x\") pod \"nova-api-0\" (UID: \"a93e0b93-3cde-4098-a853-ea1b3c86c323\") " pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.792714 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a93e0b93-3cde-4098-a853-ea1b3c86c323-config-data\") pod \"nova-api-0\" (UID: \"a93e0b93-3cde-4098-a853-ea1b3c86c323\") " pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.794865 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.795819 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a93e0b93-3cde-4098-a853-ea1b3c86c323-logs\") pod \"nova-api-0\" (UID: \"a93e0b93-3cde-4098-a853-ea1b3c86c323\") " pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.796874 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a93e0b93-3cde-4098-a853-ea1b3c86c323-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a93e0b93-3cde-4098-a853-ea1b3c86c323\") " pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.797024 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a93e0b93-3cde-4098-a853-ea1b3c86c323-config-data\") pod \"nova-api-0\" (UID: \"a93e0b93-3cde-4098-a853-ea1b3c86c323\") " pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.798093 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a93e0b93-3cde-4098-a853-ea1b3c86c323-public-tls-certs\") pod \"nova-api-0\" (UID: \"a93e0b93-3cde-4098-a853-ea1b3c86c323\") " pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.801527 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a93e0b93-3cde-4098-a853-ea1b3c86c323-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a93e0b93-3cde-4098-a853-ea1b3c86c323\") " pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.805336 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.820468 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.822513 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.823679 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2w9x\" (UniqueName: \"kubernetes.io/projected/a93e0b93-3cde-4098-a853-ea1b3c86c323-kube-api-access-k2w9x\") pod \"nova-api-0\" (UID: \"a93e0b93-3cde-4098-a853-ea1b3c86c323\") " pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.825300 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.860501 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.894734 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qzls\" (UniqueName: \"kubernetes.io/projected/64dcd71b-3e04-470c-ab99-37105bd225dc-kube-api-access-9qzls\") pod \"nova-scheduler-0\" (UID: \"64dcd71b-3e04-470c-ab99-37105bd225dc\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.894786 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64dcd71b-3e04-470c-ab99-37105bd225dc-config-data\") pod \"nova-scheduler-0\" (UID: \"64dcd71b-3e04-470c-ab99-37105bd225dc\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.894916 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64dcd71b-3e04-470c-ab99-37105bd225dc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"64dcd71b-3e04-470c-ab99-37105bd225dc\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.908275 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.995794 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64dcd71b-3e04-470c-ab99-37105bd225dc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"64dcd71b-3e04-470c-ab99-37105bd225dc\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.995878 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qzls\" (UniqueName: \"kubernetes.io/projected/64dcd71b-3e04-470c-ab99-37105bd225dc-kube-api-access-9qzls\") pod \"nova-scheduler-0\" (UID: \"64dcd71b-3e04-470c-ab99-37105bd225dc\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.995905 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64dcd71b-3e04-470c-ab99-37105bd225dc-config-data\") pod \"nova-scheduler-0\" (UID: \"64dcd71b-3e04-470c-ab99-37105bd225dc\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.999267 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64dcd71b-3e04-470c-ab99-37105bd225dc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"64dcd71b-3e04-470c-ab99-37105bd225dc\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:58 crc kubenswrapper[4766]: I1126 00:49:58.999418 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64dcd71b-3e04-470c-ab99-37105bd225dc-config-data\") pod \"nova-scheduler-0\" (UID: \"64dcd71b-3e04-470c-ab99-37105bd225dc\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:59 crc kubenswrapper[4766]: I1126 00:49:59.013778 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qzls\" (UniqueName: \"kubernetes.io/projected/64dcd71b-3e04-470c-ab99-37105bd225dc-kube-api-access-9qzls\") pod \"nova-scheduler-0\" (UID: \"64dcd71b-3e04-470c-ab99-37105bd225dc\") " pod="openstack/nova-scheduler-0" Nov 26 00:49:59 crc kubenswrapper[4766]: I1126 00:49:59.247463 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 26 00:49:59 crc kubenswrapper[4766]: W1126 00:49:59.506645 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda93e0b93_3cde_4098_a853_ea1b3c86c323.slice/crio-530fc91057ac2889a3cf86f1c4aae85229c40c88c9e4b238a61642519e8514b9 WatchSource:0}: Error finding container 530fc91057ac2889a3cf86f1c4aae85229c40c88c9e4b238a61642519e8514b9: Status 404 returned error can't find the container with id 530fc91057ac2889a3cf86f1c4aae85229c40c88c9e4b238a61642519e8514b9 Nov 26 00:49:59 crc kubenswrapper[4766]: I1126 00:49:59.508395 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 26 00:49:59 crc kubenswrapper[4766]: I1126 00:49:59.591599 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="08a8f08a-1eb6-424c-874d-15efb5083a8d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.240:8775/\": read tcp 10.217.0.2:34418->10.217.0.240:8775: read: connection reset by peer" Nov 26 00:49:59 crc kubenswrapper[4766]: I1126 00:49:59.591606 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="08a8f08a-1eb6-424c-874d-15efb5083a8d" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.240:8775/\": read tcp 10.217.0.2:34428->10.217.0.240:8775: read: connection reset by peer" Nov 26 00:49:59 crc kubenswrapper[4766]: I1126 00:49:59.720014 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 00:49:59 crc kubenswrapper[4766]: W1126 00:49:59.722531 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64dcd71b_3e04_470c_ab99_37105bd225dc.slice/crio-ad72c1f5e20b42b67f9b637e42c170d9839b3b22a6b3b7bb7b27f3b65e232b0e WatchSource:0}: Error finding container ad72c1f5e20b42b67f9b637e42c170d9839b3b22a6b3b7bb7b27f3b65e232b0e: Status 404 returned error can't find the container with id ad72c1f5e20b42b67f9b637e42c170d9839b3b22a6b3b7bb7b27f3b65e232b0e Nov 26 00:49:59 crc kubenswrapper[4766]: I1126 00:49:59.858568 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bc2e0a9-02f7-4089-9820-62e9e07a98b8" path="/var/lib/kubelet/pods/7bc2e0a9-02f7-4089-9820-62e9e07a98b8/volumes" Nov 26 00:49:59 crc kubenswrapper[4766]: I1126 00:49:59.859269 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b43a5476-de61-4907-ba9c-3195b7b2a85b" path="/var/lib/kubelet/pods/b43a5476-de61-4907-ba9c-3195b7b2a85b/volumes" Nov 26 00:49:59 crc kubenswrapper[4766]: I1126 00:49:59.909272 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mbqtt" podUID="ed786021-3dca-48e5-b286-f1a06dcbe58e" containerName="registry-server" probeResult="failure" output=< Nov 26 00:49:59 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 00:49:59 crc kubenswrapper[4766]: > Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.041585 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.116607 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08a8f08a-1eb6-424c-874d-15efb5083a8d-logs\") pod \"08a8f08a-1eb6-424c-874d-15efb5083a8d\" (UID: \"08a8f08a-1eb6-424c-874d-15efb5083a8d\") " Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.116779 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08a8f08a-1eb6-424c-874d-15efb5083a8d-config-data\") pod \"08a8f08a-1eb6-424c-874d-15efb5083a8d\" (UID: \"08a8f08a-1eb6-424c-874d-15efb5083a8d\") " Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.116808 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/08a8f08a-1eb6-424c-874d-15efb5083a8d-nova-metadata-tls-certs\") pod \"08a8f08a-1eb6-424c-874d-15efb5083a8d\" (UID: \"08a8f08a-1eb6-424c-874d-15efb5083a8d\") " Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.116841 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fx6cc\" (UniqueName: \"kubernetes.io/projected/08a8f08a-1eb6-424c-874d-15efb5083a8d-kube-api-access-fx6cc\") pod \"08a8f08a-1eb6-424c-874d-15efb5083a8d\" (UID: \"08a8f08a-1eb6-424c-874d-15efb5083a8d\") " Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.116921 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08a8f08a-1eb6-424c-874d-15efb5083a8d-combined-ca-bundle\") pod \"08a8f08a-1eb6-424c-874d-15efb5083a8d\" (UID: \"08a8f08a-1eb6-424c-874d-15efb5083a8d\") " Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.117390 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08a8f08a-1eb6-424c-874d-15efb5083a8d-logs" (OuterVolumeSpecName: "logs") pod "08a8f08a-1eb6-424c-874d-15efb5083a8d" (UID: "08a8f08a-1eb6-424c-874d-15efb5083a8d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.117887 4766 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08a8f08a-1eb6-424c-874d-15efb5083a8d-logs\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.129859 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08a8f08a-1eb6-424c-874d-15efb5083a8d-kube-api-access-fx6cc" (OuterVolumeSpecName: "kube-api-access-fx6cc") pod "08a8f08a-1eb6-424c-874d-15efb5083a8d" (UID: "08a8f08a-1eb6-424c-874d-15efb5083a8d"). InnerVolumeSpecName "kube-api-access-fx6cc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.170905 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08a8f08a-1eb6-424c-874d-15efb5083a8d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "08a8f08a-1eb6-424c-874d-15efb5083a8d" (UID: "08a8f08a-1eb6-424c-874d-15efb5083a8d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.225769 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08a8f08a-1eb6-424c-874d-15efb5083a8d-config-data" (OuterVolumeSpecName: "config-data") pod "08a8f08a-1eb6-424c-874d-15efb5083a8d" (UID: "08a8f08a-1eb6-424c-874d-15efb5083a8d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.226421 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fx6cc\" (UniqueName: \"kubernetes.io/projected/08a8f08a-1eb6-424c-874d-15efb5083a8d-kube-api-access-fx6cc\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.226455 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08a8f08a-1eb6-424c-874d-15efb5083a8d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.230834 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08a8f08a-1eb6-424c-874d-15efb5083a8d-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "08a8f08a-1eb6-424c-874d-15efb5083a8d" (UID: "08a8f08a-1eb6-424c-874d-15efb5083a8d"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.329023 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08a8f08a-1eb6-424c-874d-15efb5083a8d-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.329072 4766 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/08a8f08a-1eb6-424c-874d-15efb5083a8d-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.503324 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"64dcd71b-3e04-470c-ab99-37105bd225dc","Type":"ContainerStarted","Data":"fda331f8469c941447d6088ec6a032aaaad09d6a805c5fb68ba9e7325286df5a"} Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.503375 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"64dcd71b-3e04-470c-ab99-37105bd225dc","Type":"ContainerStarted","Data":"ad72c1f5e20b42b67f9b637e42c170d9839b3b22a6b3b7bb7b27f3b65e232b0e"} Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.508421 4766 generic.go:334] "Generic (PLEG): container finished" podID="08a8f08a-1eb6-424c-874d-15efb5083a8d" containerID="b76d44319915d6e4df3307e2d68144e17141e42e5fbcc5ddfb665e3e09cb5a09" exitCode=0 Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.508488 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08a8f08a-1eb6-424c-874d-15efb5083a8d","Type":"ContainerDied","Data":"b76d44319915d6e4df3307e2d68144e17141e42e5fbcc5ddfb665e3e09cb5a09"} Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.508518 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08a8f08a-1eb6-424c-874d-15efb5083a8d","Type":"ContainerDied","Data":"06f16064bc03bdcf4a82b6d3e08ea8a397f73de53afd6363f86fc7be4f194c62"} Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.508538 4766 scope.go:117] "RemoveContainer" containerID="b76d44319915d6e4df3307e2d68144e17141e42e5fbcc5ddfb665e3e09cb5a09" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.508773 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.520733 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a93e0b93-3cde-4098-a853-ea1b3c86c323","Type":"ContainerStarted","Data":"9d13f3b309723be646012403af3cd62353b149fa35adca173991c2a49f83cd3f"} Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.520768 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a93e0b93-3cde-4098-a853-ea1b3c86c323","Type":"ContainerStarted","Data":"b37a4e1769fb4c03cfb102c4b4a1737f4cc338449eec5ad79d785640a738c496"} Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.520779 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a93e0b93-3cde-4098-a853-ea1b3c86c323","Type":"ContainerStarted","Data":"530fc91057ac2889a3cf86f1c4aae85229c40c88c9e4b238a61642519e8514b9"} Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.528088 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.528066152 podStartE2EDuration="2.528066152s" podCreationTimestamp="2025-11-26 00:49:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:50:00.517673483 +0000 UTC m=+1581.366443913" watchObservedRunningTime="2025-11-26 00:50:00.528066152 +0000 UTC m=+1581.376836602" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.564319 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.564300475 podStartE2EDuration="2.564300475s" podCreationTimestamp="2025-11-26 00:49:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:50:00.543415684 +0000 UTC m=+1581.392186124" watchObservedRunningTime="2025-11-26 00:50:00.564300475 +0000 UTC m=+1581.413070905" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.589851 4766 scope.go:117] "RemoveContainer" containerID="a15f6b81a4cb1d4b9585ef657ff231e5573f9028d2c4e0eab7f21eaaafaffdad" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.599870 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.618774 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.629454 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 26 00:50:00 crc kubenswrapper[4766]: E1126 00:50:00.630058 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08a8f08a-1eb6-424c-874d-15efb5083a8d" containerName="nova-metadata-metadata" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.630076 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="08a8f08a-1eb6-424c-874d-15efb5083a8d" containerName="nova-metadata-metadata" Nov 26 00:50:00 crc kubenswrapper[4766]: E1126 00:50:00.630110 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08a8f08a-1eb6-424c-874d-15efb5083a8d" containerName="nova-metadata-log" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.630119 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="08a8f08a-1eb6-424c-874d-15efb5083a8d" containerName="nova-metadata-log" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.630418 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="08a8f08a-1eb6-424c-874d-15efb5083a8d" containerName="nova-metadata-log" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.630434 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="08a8f08a-1eb6-424c-874d-15efb5083a8d" containerName="nova-metadata-metadata" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.632988 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.636310 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.639622 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.645395 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.676016 4766 scope.go:117] "RemoveContainer" containerID="b76d44319915d6e4df3307e2d68144e17141e42e5fbcc5ddfb665e3e09cb5a09" Nov 26 00:50:00 crc kubenswrapper[4766]: E1126 00:50:00.679976 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b76d44319915d6e4df3307e2d68144e17141e42e5fbcc5ddfb665e3e09cb5a09\": container with ID starting with b76d44319915d6e4df3307e2d68144e17141e42e5fbcc5ddfb665e3e09cb5a09 not found: ID does not exist" containerID="b76d44319915d6e4df3307e2d68144e17141e42e5fbcc5ddfb665e3e09cb5a09" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.680008 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b76d44319915d6e4df3307e2d68144e17141e42e5fbcc5ddfb665e3e09cb5a09"} err="failed to get container status \"b76d44319915d6e4df3307e2d68144e17141e42e5fbcc5ddfb665e3e09cb5a09\": rpc error: code = NotFound desc = could not find container \"b76d44319915d6e4df3307e2d68144e17141e42e5fbcc5ddfb665e3e09cb5a09\": container with ID starting with b76d44319915d6e4df3307e2d68144e17141e42e5fbcc5ddfb665e3e09cb5a09 not found: ID does not exist" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.680027 4766 scope.go:117] "RemoveContainer" containerID="a15f6b81a4cb1d4b9585ef657ff231e5573f9028d2c4e0eab7f21eaaafaffdad" Nov 26 00:50:00 crc kubenswrapper[4766]: E1126 00:50:00.683626 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a15f6b81a4cb1d4b9585ef657ff231e5573f9028d2c4e0eab7f21eaaafaffdad\": container with ID starting with a15f6b81a4cb1d4b9585ef657ff231e5573f9028d2c4e0eab7f21eaaafaffdad not found: ID does not exist" containerID="a15f6b81a4cb1d4b9585ef657ff231e5573f9028d2c4e0eab7f21eaaafaffdad" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.683686 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a15f6b81a4cb1d4b9585ef657ff231e5573f9028d2c4e0eab7f21eaaafaffdad"} err="failed to get container status \"a15f6b81a4cb1d4b9585ef657ff231e5573f9028d2c4e0eab7f21eaaafaffdad\": rpc error: code = NotFound desc = could not find container \"a15f6b81a4cb1d4b9585ef657ff231e5573f9028d2c4e0eab7f21eaaafaffdad\": container with ID starting with a15f6b81a4cb1d4b9585ef657ff231e5573f9028d2c4e0eab7f21eaaafaffdad not found: ID does not exist" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.741737 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpf9b\" (UniqueName: \"kubernetes.io/projected/1f172da7-f357-44b5-a4c0-5a2cc493f906-kube-api-access-jpf9b\") pod \"nova-metadata-0\" (UID: \"1f172da7-f357-44b5-a4c0-5a2cc493f906\") " pod="openstack/nova-metadata-0" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.741807 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f172da7-f357-44b5-a4c0-5a2cc493f906-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1f172da7-f357-44b5-a4c0-5a2cc493f906\") " pod="openstack/nova-metadata-0" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.741891 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f172da7-f357-44b5-a4c0-5a2cc493f906-logs\") pod \"nova-metadata-0\" (UID: \"1f172da7-f357-44b5-a4c0-5a2cc493f906\") " pod="openstack/nova-metadata-0" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.741924 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f172da7-f357-44b5-a4c0-5a2cc493f906-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1f172da7-f357-44b5-a4c0-5a2cc493f906\") " pod="openstack/nova-metadata-0" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.741992 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f172da7-f357-44b5-a4c0-5a2cc493f906-config-data\") pod \"nova-metadata-0\" (UID: \"1f172da7-f357-44b5-a4c0-5a2cc493f906\") " pod="openstack/nova-metadata-0" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.844153 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f172da7-f357-44b5-a4c0-5a2cc493f906-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1f172da7-f357-44b5-a4c0-5a2cc493f906\") " pod="openstack/nova-metadata-0" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.844261 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f172da7-f357-44b5-a4c0-5a2cc493f906-logs\") pod \"nova-metadata-0\" (UID: \"1f172da7-f357-44b5-a4c0-5a2cc493f906\") " pod="openstack/nova-metadata-0" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.844294 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f172da7-f357-44b5-a4c0-5a2cc493f906-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1f172da7-f357-44b5-a4c0-5a2cc493f906\") " pod="openstack/nova-metadata-0" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.844343 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f172da7-f357-44b5-a4c0-5a2cc493f906-config-data\") pod \"nova-metadata-0\" (UID: \"1f172da7-f357-44b5-a4c0-5a2cc493f906\") " pod="openstack/nova-metadata-0" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.844547 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpf9b\" (UniqueName: \"kubernetes.io/projected/1f172da7-f357-44b5-a4c0-5a2cc493f906-kube-api-access-jpf9b\") pod \"nova-metadata-0\" (UID: \"1f172da7-f357-44b5-a4c0-5a2cc493f906\") " pod="openstack/nova-metadata-0" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.844801 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f172da7-f357-44b5-a4c0-5a2cc493f906-logs\") pod \"nova-metadata-0\" (UID: \"1f172da7-f357-44b5-a4c0-5a2cc493f906\") " pod="openstack/nova-metadata-0" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.848259 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f172da7-f357-44b5-a4c0-5a2cc493f906-config-data\") pod \"nova-metadata-0\" (UID: \"1f172da7-f357-44b5-a4c0-5a2cc493f906\") " pod="openstack/nova-metadata-0" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.848554 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f172da7-f357-44b5-a4c0-5a2cc493f906-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1f172da7-f357-44b5-a4c0-5a2cc493f906\") " pod="openstack/nova-metadata-0" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.849005 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f172da7-f357-44b5-a4c0-5a2cc493f906-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1f172da7-f357-44b5-a4c0-5a2cc493f906\") " pod="openstack/nova-metadata-0" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.863188 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpf9b\" (UniqueName: \"kubernetes.io/projected/1f172da7-f357-44b5-a4c0-5a2cc493f906-kube-api-access-jpf9b\") pod \"nova-metadata-0\" (UID: \"1f172da7-f357-44b5-a4c0-5a2cc493f906\") " pod="openstack/nova-metadata-0" Nov 26 00:50:00 crc kubenswrapper[4766]: I1126 00:50:00.965301 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 00:50:01 crc kubenswrapper[4766]: I1126 00:50:01.478375 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 00:50:01 crc kubenswrapper[4766]: W1126 00:50:01.483428 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f172da7_f357_44b5_a4c0_5a2cc493f906.slice/crio-5bb5b658af996253fd7da7afc761834c8601e2d7913d8d64c512a70d81314c1e WatchSource:0}: Error finding container 5bb5b658af996253fd7da7afc761834c8601e2d7913d8d64c512a70d81314c1e: Status 404 returned error can't find the container with id 5bb5b658af996253fd7da7afc761834c8601e2d7913d8d64c512a70d81314c1e Nov 26 00:50:01 crc kubenswrapper[4766]: I1126 00:50:01.588970 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f172da7-f357-44b5-a4c0-5a2cc493f906","Type":"ContainerStarted","Data":"5bb5b658af996253fd7da7afc761834c8601e2d7913d8d64c512a70d81314c1e"} Nov 26 00:50:01 crc kubenswrapper[4766]: I1126 00:50:01.842031 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08a8f08a-1eb6-424c-874d-15efb5083a8d" path="/var/lib/kubelet/pods/08a8f08a-1eb6-424c-874d-15efb5083a8d/volumes" Nov 26 00:50:02 crc kubenswrapper[4766]: E1126 00:50:02.508795 4766 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdd2abaa_1d30_4422_80c9_496e89a9aad9.slice/crio-0a967a8f03be6a9be05cfb69b3c9a733cadd958dbc86857bda20484c0d3d6cb5\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdd2abaa_1d30_4422_80c9_496e89a9aad9.slice\": RecentStats: unable to find data in memory cache]" Nov 26 00:50:02 crc kubenswrapper[4766]: I1126 00:50:02.609379 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f172da7-f357-44b5-a4c0-5a2cc493f906","Type":"ContainerStarted","Data":"c24eba699f10cfc77993541f961aa0b52fd5113ecb8e4171a46d94eb5a6f5269"} Nov 26 00:50:02 crc kubenswrapper[4766]: I1126 00:50:02.609728 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f172da7-f357-44b5-a4c0-5a2cc493f906","Type":"ContainerStarted","Data":"c5f0be86f73aad513cac79d742822da40147318792f71481acc6610e5738a3ea"} Nov 26 00:50:02 crc kubenswrapper[4766]: I1126 00:50:02.625836 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.62581897 podStartE2EDuration="2.62581897s" podCreationTimestamp="2025-11-26 00:50:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:50:02.625368849 +0000 UTC m=+1583.474139279" watchObservedRunningTime="2025-11-26 00:50:02.62581897 +0000 UTC m=+1583.474589400" Nov 26 00:50:03 crc kubenswrapper[4766]: I1126 00:50:03.812036 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:50:03 crc kubenswrapper[4766]: I1126 00:50:03.812368 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" containerName="ceilometer-central-agent" containerID="cri-o://81a18a1a1a33fce2ef59e20de6029e3f5ecb95513bd41c361e5d3a25170b3a8c" gracePeriod=30 Nov 26 00:50:03 crc kubenswrapper[4766]: I1126 00:50:03.812425 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" containerName="sg-core" containerID="cri-o://c41bd398fb0aeb4ea7712daecc1565be2bc9a263b141abc17405ebb24c8de9ce" gracePeriod=30 Nov 26 00:50:03 crc kubenswrapper[4766]: I1126 00:50:03.812441 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" containerName="proxy-httpd" containerID="cri-o://b2138ff204dd583ad9f427fd6407d82c57b26b12b9bb1433ac804293133a3552" gracePeriod=30 Nov 26 00:50:03 crc kubenswrapper[4766]: I1126 00:50:03.812454 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" containerName="ceilometer-notification-agent" containerID="cri-o://6673e046e2197a4b6fac11035a178781f4e83f5c3084eaaeefdd9e783da3f0ff" gracePeriod=30 Nov 26 00:50:03 crc kubenswrapper[4766]: I1126 00:50:03.833111 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.246:3000/\": EOF" Nov 26 00:50:04 crc kubenswrapper[4766]: I1126 00:50:04.248360 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 26 00:50:04 crc kubenswrapper[4766]: I1126 00:50:04.633739 4766 generic.go:334] "Generic (PLEG): container finished" podID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" containerID="b2138ff204dd583ad9f427fd6407d82c57b26b12b9bb1433ac804293133a3552" exitCode=0 Nov 26 00:50:04 crc kubenswrapper[4766]: I1126 00:50:04.633776 4766 generic.go:334] "Generic (PLEG): container finished" podID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" containerID="c41bd398fb0aeb4ea7712daecc1565be2bc9a263b141abc17405ebb24c8de9ce" exitCode=2 Nov 26 00:50:04 crc kubenswrapper[4766]: I1126 00:50:04.633786 4766 generic.go:334] "Generic (PLEG): container finished" podID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" containerID="81a18a1a1a33fce2ef59e20de6029e3f5ecb95513bd41c361e5d3a25170b3a8c" exitCode=0 Nov 26 00:50:04 crc kubenswrapper[4766]: I1126 00:50:04.633790 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62490687-08a8-4e40-a3f7-d8c64aeb5d24","Type":"ContainerDied","Data":"b2138ff204dd583ad9f427fd6407d82c57b26b12b9bb1433ac804293133a3552"} Nov 26 00:50:04 crc kubenswrapper[4766]: I1126 00:50:04.633831 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62490687-08a8-4e40-a3f7-d8c64aeb5d24","Type":"ContainerDied","Data":"c41bd398fb0aeb4ea7712daecc1565be2bc9a263b141abc17405ebb24c8de9ce"} Nov 26 00:50:04 crc kubenswrapper[4766]: I1126 00:50:04.633844 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62490687-08a8-4e40-a3f7-d8c64aeb5d24","Type":"ContainerDied","Data":"81a18a1a1a33fce2ef59e20de6029e3f5ecb95513bd41c361e5d3a25170b3a8c"} Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.649420 4766 generic.go:334] "Generic (PLEG): container finished" podID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" containerID="6673e046e2197a4b6fac11035a178781f4e83f5c3084eaaeefdd9e783da3f0ff" exitCode=0 Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.649742 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62490687-08a8-4e40-a3f7-d8c64aeb5d24","Type":"ContainerDied","Data":"6673e046e2197a4b6fac11035a178781f4e83f5c3084eaaeefdd9e783da3f0ff"} Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.649773 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62490687-08a8-4e40-a3f7-d8c64aeb5d24","Type":"ContainerDied","Data":"b3e9807dba0caa3d3b46315aab907b7c60dd37cfd8fff70d37b60cc2810b2922"} Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.649786 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3e9807dba0caa3d3b46315aab907b7c60dd37cfd8fff70d37b60cc2810b2922" Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.660613 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.758108 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-scripts\") pod \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.758167 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c64np\" (UniqueName: \"kubernetes.io/projected/62490687-08a8-4e40-a3f7-d8c64aeb5d24-kube-api-access-c64np\") pod \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.758204 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62490687-08a8-4e40-a3f7-d8c64aeb5d24-log-httpd\") pod \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.758238 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-config-data\") pod \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.758372 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-combined-ca-bundle\") pod \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.758408 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-sg-core-conf-yaml\") pod \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.758481 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62490687-08a8-4e40-a3f7-d8c64aeb5d24-run-httpd\") pod \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.758640 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-ceilometer-tls-certs\") pod \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\" (UID: \"62490687-08a8-4e40-a3f7-d8c64aeb5d24\") " Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.758837 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62490687-08a8-4e40-a3f7-d8c64aeb5d24-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "62490687-08a8-4e40-a3f7-d8c64aeb5d24" (UID: "62490687-08a8-4e40-a3f7-d8c64aeb5d24"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.758917 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62490687-08a8-4e40-a3f7-d8c64aeb5d24-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "62490687-08a8-4e40-a3f7-d8c64aeb5d24" (UID: "62490687-08a8-4e40-a3f7-d8c64aeb5d24"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.759249 4766 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62490687-08a8-4e40-a3f7-d8c64aeb5d24-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.759293 4766 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62490687-08a8-4e40-a3f7-d8c64aeb5d24-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.764608 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62490687-08a8-4e40-a3f7-d8c64aeb5d24-kube-api-access-c64np" (OuterVolumeSpecName: "kube-api-access-c64np") pod "62490687-08a8-4e40-a3f7-d8c64aeb5d24" (UID: "62490687-08a8-4e40-a3f7-d8c64aeb5d24"). InnerVolumeSpecName "kube-api-access-c64np". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.770030 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-scripts" (OuterVolumeSpecName: "scripts") pod "62490687-08a8-4e40-a3f7-d8c64aeb5d24" (UID: "62490687-08a8-4e40-a3f7-d8c64aeb5d24"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.795268 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "62490687-08a8-4e40-a3f7-d8c64aeb5d24" (UID: "62490687-08a8-4e40-a3f7-d8c64aeb5d24"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.837277 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "62490687-08a8-4e40-a3f7-d8c64aeb5d24" (UID: "62490687-08a8-4e40-a3f7-d8c64aeb5d24"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.861871 4766 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.861908 4766 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.861919 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.861930 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c64np\" (UniqueName: \"kubernetes.io/projected/62490687-08a8-4e40-a3f7-d8c64aeb5d24-kube-api-access-c64np\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.876427 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62490687-08a8-4e40-a3f7-d8c64aeb5d24" (UID: "62490687-08a8-4e40-a3f7-d8c64aeb5d24"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.917186 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-config-data" (OuterVolumeSpecName: "config-data") pod "62490687-08a8-4e40-a3f7-d8c64aeb5d24" (UID: "62490687-08a8-4e40-a3f7-d8c64aeb5d24"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.963891 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.964147 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62490687-08a8-4e40-a3f7-d8c64aeb5d24-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.965906 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 26 00:50:05 crc kubenswrapper[4766]: I1126 00:50:05.965947 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.658479 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.732297 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.748730 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.775217 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:50:06 crc kubenswrapper[4766]: E1126 00:50:06.775912 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" containerName="proxy-httpd" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.775937 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" containerName="proxy-httpd" Nov 26 00:50:06 crc kubenswrapper[4766]: E1126 00:50:06.776007 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" containerName="ceilometer-central-agent" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.776018 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" containerName="ceilometer-central-agent" Nov 26 00:50:06 crc kubenswrapper[4766]: E1126 00:50:06.776029 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" containerName="sg-core" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.776036 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" containerName="sg-core" Nov 26 00:50:06 crc kubenswrapper[4766]: E1126 00:50:06.776068 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" containerName="ceilometer-notification-agent" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.776076 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" containerName="ceilometer-notification-agent" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.776340 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" containerName="ceilometer-notification-agent" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.776376 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" containerName="sg-core" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.776403 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" containerName="ceilometer-central-agent" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.776419 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" containerName="proxy-httpd" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.778581 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.780811 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.781373 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.783870 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.801919 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.827764 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:50:06 crc kubenswrapper[4766]: E1126 00:50:06.828142 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.883505 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.883902 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfvzv\" (UniqueName: \"kubernetes.io/projected/dbc768d1-ef8f-4c8f-baca-fbde7c820793-kube-api-access-vfvzv\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.884117 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.884165 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-scripts\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.884186 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-config-data\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.884247 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbc768d1-ef8f-4c8f-baca-fbde7c820793-run-httpd\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.884433 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbc768d1-ef8f-4c8f-baca-fbde7c820793-log-httpd\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.884521 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.985966 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.986027 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-scripts\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.986045 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-config-data\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.986077 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbc768d1-ef8f-4c8f-baca-fbde7c820793-run-httpd\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.986152 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbc768d1-ef8f-4c8f-baca-fbde7c820793-log-httpd\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.986192 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.986229 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.986278 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfvzv\" (UniqueName: \"kubernetes.io/projected/dbc768d1-ef8f-4c8f-baca-fbde7c820793-kube-api-access-vfvzv\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.986615 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbc768d1-ef8f-4c8f-baca-fbde7c820793-run-httpd\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.986691 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbc768d1-ef8f-4c8f-baca-fbde7c820793-log-httpd\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.990101 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.990170 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.990639 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-scripts\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.992359 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:06 crc kubenswrapper[4766]: I1126 00:50:06.993618 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-config-data\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:07 crc kubenswrapper[4766]: I1126 00:50:07.008902 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfvzv\" (UniqueName: \"kubernetes.io/projected/dbc768d1-ef8f-4c8f-baca-fbde7c820793-kube-api-access-vfvzv\") pod \"ceilometer-0\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " pod="openstack/ceilometer-0" Nov 26 00:50:07 crc kubenswrapper[4766]: I1126 00:50:07.105784 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:50:07 crc kubenswrapper[4766]: W1126 00:50:07.633021 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddbc768d1_ef8f_4c8f_baca_fbde7c820793.slice/crio-ca46fb3dea11a816ae02a9e7618138ad9c9c8c74ab94ed9135cfce07533476eb WatchSource:0}: Error finding container ca46fb3dea11a816ae02a9e7618138ad9c9c8c74ab94ed9135cfce07533476eb: Status 404 returned error can't find the container with id ca46fb3dea11a816ae02a9e7618138ad9c9c8c74ab94ed9135cfce07533476eb Nov 26 00:50:07 crc kubenswrapper[4766]: I1126 00:50:07.648076 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:50:07 crc kubenswrapper[4766]: I1126 00:50:07.673174 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbc768d1-ef8f-4c8f-baca-fbde7c820793","Type":"ContainerStarted","Data":"ca46fb3dea11a816ae02a9e7618138ad9c9c8c74ab94ed9135cfce07533476eb"} Nov 26 00:50:07 crc kubenswrapper[4766]: I1126 00:50:07.843618 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62490687-08a8-4e40-a3f7-d8c64aeb5d24" path="/var/lib/kubelet/pods/62490687-08a8-4e40-a3f7-d8c64aeb5d24/volumes" Nov 26 00:50:08 crc kubenswrapper[4766]: I1126 00:50:08.687583 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbc768d1-ef8f-4c8f-baca-fbde7c820793","Type":"ContainerStarted","Data":"efcafb1d3b6ab9a688d80751a6a8e50d35e35ca981bb7ae8b6994d88b053d737"} Nov 26 00:50:08 crc kubenswrapper[4766]: I1126 00:50:08.908909 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 26 00:50:08 crc kubenswrapper[4766]: I1126 00:50:08.909254 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 26 00:50:09 crc kubenswrapper[4766]: I1126 00:50:09.248137 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 26 00:50:09 crc kubenswrapper[4766]: I1126 00:50:09.284197 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 26 00:50:09 crc kubenswrapper[4766]: I1126 00:50:09.707908 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbc768d1-ef8f-4c8f-baca-fbde7c820793","Type":"ContainerStarted","Data":"4b5c5b7cc746ea7f62128481c01f60c0ad13e847bbbc90f3e1d919a11559ccac"} Nov 26 00:50:09 crc kubenswrapper[4766]: I1126 00:50:09.751421 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 26 00:50:09 crc kubenswrapper[4766]: I1126 00:50:09.886828 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mbqtt" podUID="ed786021-3dca-48e5-b286-f1a06dcbe58e" containerName="registry-server" probeResult="failure" output=< Nov 26 00:50:09 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 00:50:09 crc kubenswrapper[4766]: > Nov 26 00:50:09 crc kubenswrapper[4766]: I1126 00:50:09.920792 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a93e0b93-3cde-4098-a853-ea1b3c86c323" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.247:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 26 00:50:09 crc kubenswrapper[4766]: I1126 00:50:09.920802 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a93e0b93-3cde-4098-a853-ea1b3c86c323" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.247:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 26 00:50:10 crc kubenswrapper[4766]: I1126 00:50:10.726160 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbc768d1-ef8f-4c8f-baca-fbde7c820793","Type":"ContainerStarted","Data":"5715943ffc4af11e634fdc21dcee90a4abe8b3afe0a72df09d8a87d8e626eeec"} Nov 26 00:50:11 crc kubenswrapper[4766]: I1126 00:50:11.124156 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 26 00:50:11 crc kubenswrapper[4766]: I1126 00:50:11.124191 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 26 00:50:11 crc kubenswrapper[4766]: I1126 00:50:11.741366 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbc768d1-ef8f-4c8f-baca-fbde7c820793","Type":"ContainerStarted","Data":"d3d64f79241bbd8529f11a9ada179a12bb7068bb63af1244e64a039b89e414ac"} Nov 26 00:50:11 crc kubenswrapper[4766]: I1126 00:50:11.741723 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 26 00:50:11 crc kubenswrapper[4766]: I1126 00:50:11.770511 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.240614388 podStartE2EDuration="5.770493539s" podCreationTimestamp="2025-11-26 00:50:06 +0000 UTC" firstStartedPulling="2025-11-26 00:50:07.637570863 +0000 UTC m=+1588.486341313" lastFinishedPulling="2025-11-26 00:50:11.167450034 +0000 UTC m=+1592.016220464" observedRunningTime="2025-11-26 00:50:11.758535671 +0000 UTC m=+1592.607306091" watchObservedRunningTime="2025-11-26 00:50:11.770493539 +0000 UTC m=+1592.619263969" Nov 26 00:50:12 crc kubenswrapper[4766]: I1126 00:50:12.135885 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1f172da7-f357-44b5-a4c0-5a2cc493f906" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.249:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 26 00:50:12 crc kubenswrapper[4766]: I1126 00:50:12.136206 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1f172da7-f357-44b5-a4c0-5a2cc493f906" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.249:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 26 00:50:12 crc kubenswrapper[4766]: E1126 00:50:12.801076 4766 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdd2abaa_1d30_4422_80c9_496e89a9aad9.slice/crio-0a967a8f03be6a9be05cfb69b3c9a733cadd958dbc86857bda20484c0d3d6cb5\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdd2abaa_1d30_4422_80c9_496e89a9aad9.slice\": RecentStats: unable to find data in memory cache]" Nov 26 00:50:18 crc kubenswrapper[4766]: I1126 00:50:18.907415 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mbqtt" Nov 26 00:50:18 crc kubenswrapper[4766]: I1126 00:50:18.941629 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 26 00:50:18 crc kubenswrapper[4766]: I1126 00:50:18.942829 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 26 00:50:18 crc kubenswrapper[4766]: I1126 00:50:18.943034 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 26 00:50:18 crc kubenswrapper[4766]: I1126 00:50:18.952506 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 26 00:50:18 crc kubenswrapper[4766]: I1126 00:50:18.979303 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mbqtt" Nov 26 00:50:19 crc kubenswrapper[4766]: I1126 00:50:19.734959 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mbqtt"] Nov 26 00:50:19 crc kubenswrapper[4766]: I1126 00:50:19.804449 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 26 00:50:19 crc kubenswrapper[4766]: I1126 00:50:19.855248 4766 generic.go:334] "Generic (PLEG): container finished" podID="e0f78adf-e52e-42dc-b856-a3c8e9620e01" containerID="ccb1658dcf0b4f2f56b342f572244b3b93c706ecf897af96f191ac8d30a7d977" exitCode=137 Nov 26 00:50:19 crc kubenswrapper[4766]: I1126 00:50:19.856902 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 26 00:50:19 crc kubenswrapper[4766]: I1126 00:50:19.869077 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 26 00:50:19 crc kubenswrapper[4766]: I1126 00:50:19.869473 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e0f78adf-e52e-42dc-b856-a3c8e9620e01","Type":"ContainerDied","Data":"ccb1658dcf0b4f2f56b342f572244b3b93c706ecf897af96f191ac8d30a7d977"} Nov 26 00:50:19 crc kubenswrapper[4766]: I1126 00:50:19.869520 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e0f78adf-e52e-42dc-b856-a3c8e9620e01","Type":"ContainerDied","Data":"977695b9754c81cdc9f7bc3b4ce02122329925314c11e698a116ab09f940947c"} Nov 26 00:50:19 crc kubenswrapper[4766]: I1126 00:50:19.869548 4766 scope.go:117] "RemoveContainer" containerID="ccb1658dcf0b4f2f56b342f572244b3b93c706ecf897af96f191ac8d30a7d977" Nov 26 00:50:19 crc kubenswrapper[4766]: I1126 00:50:19.882128 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 26 00:50:19 crc kubenswrapper[4766]: I1126 00:50:19.922406 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0f78adf-e52e-42dc-b856-a3c8e9620e01-scripts\") pod \"e0f78adf-e52e-42dc-b856-a3c8e9620e01\" (UID: \"e0f78adf-e52e-42dc-b856-a3c8e9620e01\") " Nov 26 00:50:19 crc kubenswrapper[4766]: I1126 00:50:19.922508 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f78adf-e52e-42dc-b856-a3c8e9620e01-config-data\") pod \"e0f78adf-e52e-42dc-b856-a3c8e9620e01\" (UID: \"e0f78adf-e52e-42dc-b856-a3c8e9620e01\") " Nov 26 00:50:19 crc kubenswrapper[4766]: I1126 00:50:19.922710 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbhx6\" (UniqueName: \"kubernetes.io/projected/e0f78adf-e52e-42dc-b856-a3c8e9620e01-kube-api-access-dbhx6\") pod \"e0f78adf-e52e-42dc-b856-a3c8e9620e01\" (UID: \"e0f78adf-e52e-42dc-b856-a3c8e9620e01\") " Nov 26 00:50:19 crc kubenswrapper[4766]: I1126 00:50:19.922861 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f78adf-e52e-42dc-b856-a3c8e9620e01-combined-ca-bundle\") pod \"e0f78adf-e52e-42dc-b856-a3c8e9620e01\" (UID: \"e0f78adf-e52e-42dc-b856-a3c8e9620e01\") " Nov 26 00:50:19 crc kubenswrapper[4766]: I1126 00:50:19.924308 4766 scope.go:117] "RemoveContainer" containerID="80a8950bd48d7905ed4f8f6eaa2d4b9a1abf0751bfc2a3cc0a263f28c226c0c7" Nov 26 00:50:19 crc kubenswrapper[4766]: I1126 00:50:19.932368 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0f78adf-e52e-42dc-b856-a3c8e9620e01-kube-api-access-dbhx6" (OuterVolumeSpecName: "kube-api-access-dbhx6") pod "e0f78adf-e52e-42dc-b856-a3c8e9620e01" (UID: "e0f78adf-e52e-42dc-b856-a3c8e9620e01"). InnerVolumeSpecName "kube-api-access-dbhx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:50:19 crc kubenswrapper[4766]: I1126 00:50:19.937978 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0f78adf-e52e-42dc-b856-a3c8e9620e01-scripts" (OuterVolumeSpecName: "scripts") pod "e0f78adf-e52e-42dc-b856-a3c8e9620e01" (UID: "e0f78adf-e52e-42dc-b856-a3c8e9620e01"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.026157 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbhx6\" (UniqueName: \"kubernetes.io/projected/e0f78adf-e52e-42dc-b856-a3c8e9620e01-kube-api-access-dbhx6\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.026191 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0f78adf-e52e-42dc-b856-a3c8e9620e01-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.074831 4766 scope.go:117] "RemoveContainer" containerID="7221b55c72c50568fb76398244c357c04af3ff796ed59da4993c4dea62ae3d6e" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.099199 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0f78adf-e52e-42dc-b856-a3c8e9620e01-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0f78adf-e52e-42dc-b856-a3c8e9620e01" (UID: "e0f78adf-e52e-42dc-b856-a3c8e9620e01"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.101229 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0f78adf-e52e-42dc-b856-a3c8e9620e01-config-data" (OuterVolumeSpecName: "config-data") pod "e0f78adf-e52e-42dc-b856-a3c8e9620e01" (UID: "e0f78adf-e52e-42dc-b856-a3c8e9620e01"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.106959 4766 scope.go:117] "RemoveContainer" containerID="6c5850bc3228dd1178546f12ee6f36f448853eef9f2e36ed77de97ee52952bd6" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.126239 4766 scope.go:117] "RemoveContainer" containerID="ccb1658dcf0b4f2f56b342f572244b3b93c706ecf897af96f191ac8d30a7d977" Nov 26 00:50:20 crc kubenswrapper[4766]: E1126 00:50:20.127205 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccb1658dcf0b4f2f56b342f572244b3b93c706ecf897af96f191ac8d30a7d977\": container with ID starting with ccb1658dcf0b4f2f56b342f572244b3b93c706ecf897af96f191ac8d30a7d977 not found: ID does not exist" containerID="ccb1658dcf0b4f2f56b342f572244b3b93c706ecf897af96f191ac8d30a7d977" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.127262 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccb1658dcf0b4f2f56b342f572244b3b93c706ecf897af96f191ac8d30a7d977"} err="failed to get container status \"ccb1658dcf0b4f2f56b342f572244b3b93c706ecf897af96f191ac8d30a7d977\": rpc error: code = NotFound desc = could not find container \"ccb1658dcf0b4f2f56b342f572244b3b93c706ecf897af96f191ac8d30a7d977\": container with ID starting with ccb1658dcf0b4f2f56b342f572244b3b93c706ecf897af96f191ac8d30a7d977 not found: ID does not exist" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.127293 4766 scope.go:117] "RemoveContainer" containerID="80a8950bd48d7905ed4f8f6eaa2d4b9a1abf0751bfc2a3cc0a263f28c226c0c7" Nov 26 00:50:20 crc kubenswrapper[4766]: E1126 00:50:20.127683 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80a8950bd48d7905ed4f8f6eaa2d4b9a1abf0751bfc2a3cc0a263f28c226c0c7\": container with ID starting with 80a8950bd48d7905ed4f8f6eaa2d4b9a1abf0751bfc2a3cc0a263f28c226c0c7 not found: ID does not exist" containerID="80a8950bd48d7905ed4f8f6eaa2d4b9a1abf0751bfc2a3cc0a263f28c226c0c7" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.127719 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0f78adf-e52e-42dc-b856-a3c8e9620e01-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.127747 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0f78adf-e52e-42dc-b856-a3c8e9620e01-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.127716 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80a8950bd48d7905ed4f8f6eaa2d4b9a1abf0751bfc2a3cc0a263f28c226c0c7"} err="failed to get container status \"80a8950bd48d7905ed4f8f6eaa2d4b9a1abf0751bfc2a3cc0a263f28c226c0c7\": rpc error: code = NotFound desc = could not find container \"80a8950bd48d7905ed4f8f6eaa2d4b9a1abf0751bfc2a3cc0a263f28c226c0c7\": container with ID starting with 80a8950bd48d7905ed4f8f6eaa2d4b9a1abf0751bfc2a3cc0a263f28c226c0c7 not found: ID does not exist" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.127781 4766 scope.go:117] "RemoveContainer" containerID="7221b55c72c50568fb76398244c357c04af3ff796ed59da4993c4dea62ae3d6e" Nov 26 00:50:20 crc kubenswrapper[4766]: E1126 00:50:20.128269 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7221b55c72c50568fb76398244c357c04af3ff796ed59da4993c4dea62ae3d6e\": container with ID starting with 7221b55c72c50568fb76398244c357c04af3ff796ed59da4993c4dea62ae3d6e not found: ID does not exist" containerID="7221b55c72c50568fb76398244c357c04af3ff796ed59da4993c4dea62ae3d6e" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.128509 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7221b55c72c50568fb76398244c357c04af3ff796ed59da4993c4dea62ae3d6e"} err="failed to get container status \"7221b55c72c50568fb76398244c357c04af3ff796ed59da4993c4dea62ae3d6e\": rpc error: code = NotFound desc = could not find container \"7221b55c72c50568fb76398244c357c04af3ff796ed59da4993c4dea62ae3d6e\": container with ID starting with 7221b55c72c50568fb76398244c357c04af3ff796ed59da4993c4dea62ae3d6e not found: ID does not exist" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.128535 4766 scope.go:117] "RemoveContainer" containerID="6c5850bc3228dd1178546f12ee6f36f448853eef9f2e36ed77de97ee52952bd6" Nov 26 00:50:20 crc kubenswrapper[4766]: E1126 00:50:20.128878 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c5850bc3228dd1178546f12ee6f36f448853eef9f2e36ed77de97ee52952bd6\": container with ID starting with 6c5850bc3228dd1178546f12ee6f36f448853eef9f2e36ed77de97ee52952bd6 not found: ID does not exist" containerID="6c5850bc3228dd1178546f12ee6f36f448853eef9f2e36ed77de97ee52952bd6" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.128915 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c5850bc3228dd1178546f12ee6f36f448853eef9f2e36ed77de97ee52952bd6"} err="failed to get container status \"6c5850bc3228dd1178546f12ee6f36f448853eef9f2e36ed77de97ee52952bd6\": rpc error: code = NotFound desc = could not find container \"6c5850bc3228dd1178546f12ee6f36f448853eef9f2e36ed77de97ee52952bd6\": container with ID starting with 6c5850bc3228dd1178546f12ee6f36f448853eef9f2e36ed77de97ee52952bd6 not found: ID does not exist" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.207423 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.218914 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.231893 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 26 00:50:20 crc kubenswrapper[4766]: E1126 00:50:20.232447 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f78adf-e52e-42dc-b856-a3c8e9620e01" containerName="aodh-api" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.232472 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f78adf-e52e-42dc-b856-a3c8e9620e01" containerName="aodh-api" Nov 26 00:50:20 crc kubenswrapper[4766]: E1126 00:50:20.232500 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f78adf-e52e-42dc-b856-a3c8e9620e01" containerName="aodh-evaluator" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.232508 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f78adf-e52e-42dc-b856-a3c8e9620e01" containerName="aodh-evaluator" Nov 26 00:50:20 crc kubenswrapper[4766]: E1126 00:50:20.232532 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f78adf-e52e-42dc-b856-a3c8e9620e01" containerName="aodh-listener" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.232540 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f78adf-e52e-42dc-b856-a3c8e9620e01" containerName="aodh-listener" Nov 26 00:50:20 crc kubenswrapper[4766]: E1126 00:50:20.232548 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f78adf-e52e-42dc-b856-a3c8e9620e01" containerName="aodh-notifier" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.232554 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f78adf-e52e-42dc-b856-a3c8e9620e01" containerName="aodh-notifier" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.232793 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0f78adf-e52e-42dc-b856-a3c8e9620e01" containerName="aodh-evaluator" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.232827 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0f78adf-e52e-42dc-b856-a3c8e9620e01" containerName="aodh-api" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.232850 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0f78adf-e52e-42dc-b856-a3c8e9620e01" containerName="aodh-notifier" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.232862 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0f78adf-e52e-42dc-b856-a3c8e9620e01" containerName="aodh-listener" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.235094 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.238201 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.239165 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.239481 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.239625 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.239771 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-9jvhv" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.244531 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.332611 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-scripts\") pod \"aodh-0\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " pod="openstack/aodh-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.332915 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-internal-tls-certs\") pod \"aodh-0\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " pod="openstack/aodh-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.332996 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7846\" (UniqueName: \"kubernetes.io/projected/5383cd51-7225-43df-8637-9d8216505757-kube-api-access-v7846\") pod \"aodh-0\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " pod="openstack/aodh-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.333067 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-combined-ca-bundle\") pod \"aodh-0\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " pod="openstack/aodh-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.333209 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-public-tls-certs\") pod \"aodh-0\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " pod="openstack/aodh-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.333317 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-config-data\") pod \"aodh-0\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " pod="openstack/aodh-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.434980 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-combined-ca-bundle\") pod \"aodh-0\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " pod="openstack/aodh-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.435111 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-public-tls-certs\") pod \"aodh-0\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " pod="openstack/aodh-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.435196 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-config-data\") pod \"aodh-0\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " pod="openstack/aodh-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.435280 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-scripts\") pod \"aodh-0\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " pod="openstack/aodh-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.435326 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-internal-tls-certs\") pod \"aodh-0\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " pod="openstack/aodh-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.435359 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7846\" (UniqueName: \"kubernetes.io/projected/5383cd51-7225-43df-8637-9d8216505757-kube-api-access-v7846\") pod \"aodh-0\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " pod="openstack/aodh-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.440491 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-scripts\") pod \"aodh-0\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " pod="openstack/aodh-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.440513 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-combined-ca-bundle\") pod \"aodh-0\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " pod="openstack/aodh-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.440607 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-public-tls-certs\") pod \"aodh-0\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " pod="openstack/aodh-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.440505 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-internal-tls-certs\") pod \"aodh-0\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " pod="openstack/aodh-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.451279 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-config-data\") pod \"aodh-0\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " pod="openstack/aodh-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.456397 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7846\" (UniqueName: \"kubernetes.io/projected/5383cd51-7225-43df-8637-9d8216505757-kube-api-access-v7846\") pod \"aodh-0\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " pod="openstack/aodh-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.573977 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.827057 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:50:20 crc kubenswrapper[4766]: E1126 00:50:20.827444 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.886975 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mbqtt" podUID="ed786021-3dca-48e5-b286-f1a06dcbe58e" containerName="registry-server" containerID="cri-o://f1b33ec75dc749ccd1d290210016ab9fe7c2d1feac99e45044b55369a17d1428" gracePeriod=2 Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.978987 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.981537 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 26 00:50:20 crc kubenswrapper[4766]: I1126 00:50:20.999785 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.087315 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 26 00:50:21 crc kubenswrapper[4766]: W1126 00:50:21.095676 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5383cd51_7225_43df_8637_9d8216505757.slice/crio-7954bcc8eaacd5ef1bb30a78deee221b2d59b5452715a1256b4d9292f93fe08a WatchSource:0}: Error finding container 7954bcc8eaacd5ef1bb30a78deee221b2d59b5452715a1256b4d9292f93fe08a: Status 404 returned error can't find the container with id 7954bcc8eaacd5ef1bb30a78deee221b2d59b5452715a1256b4d9292f93fe08a Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.497128 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mbqtt" Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.567340 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed786021-3dca-48e5-b286-f1a06dcbe58e-catalog-content\") pod \"ed786021-3dca-48e5-b286-f1a06dcbe58e\" (UID: \"ed786021-3dca-48e5-b286-f1a06dcbe58e\") " Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.567571 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed786021-3dca-48e5-b286-f1a06dcbe58e-utilities\") pod \"ed786021-3dca-48e5-b286-f1a06dcbe58e\" (UID: \"ed786021-3dca-48e5-b286-f1a06dcbe58e\") " Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.567659 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vt5g\" (UniqueName: \"kubernetes.io/projected/ed786021-3dca-48e5-b286-f1a06dcbe58e-kube-api-access-7vt5g\") pod \"ed786021-3dca-48e5-b286-f1a06dcbe58e\" (UID: \"ed786021-3dca-48e5-b286-f1a06dcbe58e\") " Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.568150 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed786021-3dca-48e5-b286-f1a06dcbe58e-utilities" (OuterVolumeSpecName: "utilities") pod "ed786021-3dca-48e5-b286-f1a06dcbe58e" (UID: "ed786021-3dca-48e5-b286-f1a06dcbe58e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.572878 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed786021-3dca-48e5-b286-f1a06dcbe58e-kube-api-access-7vt5g" (OuterVolumeSpecName: "kube-api-access-7vt5g") pod "ed786021-3dca-48e5-b286-f1a06dcbe58e" (UID: "ed786021-3dca-48e5-b286-f1a06dcbe58e"). InnerVolumeSpecName "kube-api-access-7vt5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.670724 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed786021-3dca-48e5-b286-f1a06dcbe58e-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.670763 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vt5g\" (UniqueName: \"kubernetes.io/projected/ed786021-3dca-48e5-b286-f1a06dcbe58e-kube-api-access-7vt5g\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.689017 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed786021-3dca-48e5-b286-f1a06dcbe58e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ed786021-3dca-48e5-b286-f1a06dcbe58e" (UID: "ed786021-3dca-48e5-b286-f1a06dcbe58e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.772530 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed786021-3dca-48e5-b286-f1a06dcbe58e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.844141 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0f78adf-e52e-42dc-b856-a3c8e9620e01" path="/var/lib/kubelet/pods/e0f78adf-e52e-42dc-b856-a3c8e9620e01/volumes" Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.902387 4766 generic.go:334] "Generic (PLEG): container finished" podID="ed786021-3dca-48e5-b286-f1a06dcbe58e" containerID="f1b33ec75dc749ccd1d290210016ab9fe7c2d1feac99e45044b55369a17d1428" exitCode=0 Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.902454 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mbqtt" event={"ID":"ed786021-3dca-48e5-b286-f1a06dcbe58e","Type":"ContainerDied","Data":"f1b33ec75dc749ccd1d290210016ab9fe7c2d1feac99e45044b55369a17d1428"} Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.902834 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mbqtt" event={"ID":"ed786021-3dca-48e5-b286-f1a06dcbe58e","Type":"ContainerDied","Data":"c1bda5bbfb1bb6ee9188e20ec7d5e5a7f1a551c686d571397074f7b30801ddcd"} Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.902476 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mbqtt" Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.902891 4766 scope.go:117] "RemoveContainer" containerID="f1b33ec75dc749ccd1d290210016ab9fe7c2d1feac99e45044b55369a17d1428" Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.909404 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5383cd51-7225-43df-8637-9d8216505757","Type":"ContainerStarted","Data":"dbb697fcd0acc97e4a2057b818fc9792cc3251bef001e93c129eed46cab3d934"} Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.909430 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5383cd51-7225-43df-8637-9d8216505757","Type":"ContainerStarted","Data":"7954bcc8eaacd5ef1bb30a78deee221b2d59b5452715a1256b4d9292f93fe08a"} Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.914314 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.944410 4766 scope.go:117] "RemoveContainer" containerID="d2023a97f41b4d03c09189daceb3da6384094f0ab8c9cc4b5d6dfb213e0b6f8d" Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.981193 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mbqtt"] Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.992111 4766 scope.go:117] "RemoveContainer" containerID="efcf5a32b9c10dadb05b1e9595a2e97f628ee328999c92b6bebd546967f968f8" Nov 26 00:50:21 crc kubenswrapper[4766]: I1126 00:50:21.992855 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mbqtt"] Nov 26 00:50:22 crc kubenswrapper[4766]: I1126 00:50:22.016463 4766 scope.go:117] "RemoveContainer" containerID="f1b33ec75dc749ccd1d290210016ab9fe7c2d1feac99e45044b55369a17d1428" Nov 26 00:50:22 crc kubenswrapper[4766]: E1126 00:50:22.016861 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1b33ec75dc749ccd1d290210016ab9fe7c2d1feac99e45044b55369a17d1428\": container with ID starting with f1b33ec75dc749ccd1d290210016ab9fe7c2d1feac99e45044b55369a17d1428 not found: ID does not exist" containerID="f1b33ec75dc749ccd1d290210016ab9fe7c2d1feac99e45044b55369a17d1428" Nov 26 00:50:22 crc kubenswrapper[4766]: I1126 00:50:22.016896 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1b33ec75dc749ccd1d290210016ab9fe7c2d1feac99e45044b55369a17d1428"} err="failed to get container status \"f1b33ec75dc749ccd1d290210016ab9fe7c2d1feac99e45044b55369a17d1428\": rpc error: code = NotFound desc = could not find container \"f1b33ec75dc749ccd1d290210016ab9fe7c2d1feac99e45044b55369a17d1428\": container with ID starting with f1b33ec75dc749ccd1d290210016ab9fe7c2d1feac99e45044b55369a17d1428 not found: ID does not exist" Nov 26 00:50:22 crc kubenswrapper[4766]: I1126 00:50:22.016921 4766 scope.go:117] "RemoveContainer" containerID="d2023a97f41b4d03c09189daceb3da6384094f0ab8c9cc4b5d6dfb213e0b6f8d" Nov 26 00:50:22 crc kubenswrapper[4766]: E1126 00:50:22.017373 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2023a97f41b4d03c09189daceb3da6384094f0ab8c9cc4b5d6dfb213e0b6f8d\": container with ID starting with d2023a97f41b4d03c09189daceb3da6384094f0ab8c9cc4b5d6dfb213e0b6f8d not found: ID does not exist" containerID="d2023a97f41b4d03c09189daceb3da6384094f0ab8c9cc4b5d6dfb213e0b6f8d" Nov 26 00:50:22 crc kubenswrapper[4766]: I1126 00:50:22.017413 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2023a97f41b4d03c09189daceb3da6384094f0ab8c9cc4b5d6dfb213e0b6f8d"} err="failed to get container status \"d2023a97f41b4d03c09189daceb3da6384094f0ab8c9cc4b5d6dfb213e0b6f8d\": rpc error: code = NotFound desc = could not find container \"d2023a97f41b4d03c09189daceb3da6384094f0ab8c9cc4b5d6dfb213e0b6f8d\": container with ID starting with d2023a97f41b4d03c09189daceb3da6384094f0ab8c9cc4b5d6dfb213e0b6f8d not found: ID does not exist" Nov 26 00:50:22 crc kubenswrapper[4766]: I1126 00:50:22.017426 4766 scope.go:117] "RemoveContainer" containerID="efcf5a32b9c10dadb05b1e9595a2e97f628ee328999c92b6bebd546967f968f8" Nov 26 00:50:22 crc kubenswrapper[4766]: E1126 00:50:22.017696 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efcf5a32b9c10dadb05b1e9595a2e97f628ee328999c92b6bebd546967f968f8\": container with ID starting with efcf5a32b9c10dadb05b1e9595a2e97f628ee328999c92b6bebd546967f968f8 not found: ID does not exist" containerID="efcf5a32b9c10dadb05b1e9595a2e97f628ee328999c92b6bebd546967f968f8" Nov 26 00:50:22 crc kubenswrapper[4766]: I1126 00:50:22.017717 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efcf5a32b9c10dadb05b1e9595a2e97f628ee328999c92b6bebd546967f968f8"} err="failed to get container status \"efcf5a32b9c10dadb05b1e9595a2e97f628ee328999c92b6bebd546967f968f8\": rpc error: code = NotFound desc = could not find container \"efcf5a32b9c10dadb05b1e9595a2e97f628ee328999c92b6bebd546967f968f8\": container with ID starting with efcf5a32b9c10dadb05b1e9595a2e97f628ee328999c92b6bebd546967f968f8 not found: ID does not exist" Nov 26 00:50:22 crc kubenswrapper[4766]: I1126 00:50:22.923937 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5383cd51-7225-43df-8637-9d8216505757","Type":"ContainerStarted","Data":"e70c99dab2f3e5e3dbfa4be4d4bf349e0fd90a658a331bd6182eda41ad03b53a"} Nov 26 00:50:23 crc kubenswrapper[4766]: E1126 00:50:23.112072 4766 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdd2abaa_1d30_4422_80c9_496e89a9aad9.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdd2abaa_1d30_4422_80c9_496e89a9aad9.slice/crio-0a967a8f03be6a9be05cfb69b3c9a733cadd958dbc86857bda20484c0d3d6cb5\": RecentStats: unable to find data in memory cache]" Nov 26 00:50:23 crc kubenswrapper[4766]: I1126 00:50:23.840276 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed786021-3dca-48e5-b286-f1a06dcbe58e" path="/var/lib/kubelet/pods/ed786021-3dca-48e5-b286-f1a06dcbe58e/volumes" Nov 26 00:50:23 crc kubenswrapper[4766]: I1126 00:50:23.975110 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5383cd51-7225-43df-8637-9d8216505757","Type":"ContainerStarted","Data":"3e34c59bf5bf7ca25e183e71e40495c20d81fb1424f79727c366c75530b5398e"} Nov 26 00:50:24 crc kubenswrapper[4766]: I1126 00:50:24.991121 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5383cd51-7225-43df-8637-9d8216505757","Type":"ContainerStarted","Data":"f864a4e95e0b02c3766b4690228cf2871cb060db24cf9356761f03f29da7d731"} Nov 26 00:50:25 crc kubenswrapper[4766]: I1126 00:50:25.031038 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.275294777 podStartE2EDuration="5.030980069s" podCreationTimestamp="2025-11-26 00:50:20 +0000 UTC" firstStartedPulling="2025-11-26 00:50:21.098830035 +0000 UTC m=+1601.947600465" lastFinishedPulling="2025-11-26 00:50:23.854515327 +0000 UTC m=+1604.703285757" observedRunningTime="2025-11-26 00:50:25.016095528 +0000 UTC m=+1605.864865978" watchObservedRunningTime="2025-11-26 00:50:25.030980069 +0000 UTC m=+1605.879750539" Nov 26 00:50:32 crc kubenswrapper[4766]: I1126 00:50:32.826791 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:50:32 crc kubenswrapper[4766]: E1126 00:50:32.827928 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:50:33 crc kubenswrapper[4766]: E1126 00:50:33.478281 4766 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdd2abaa_1d30_4422_80c9_496e89a9aad9.slice/crio-0a967a8f03be6a9be05cfb69b3c9a733cadd958dbc86857bda20484c0d3d6cb5\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdd2abaa_1d30_4422_80c9_496e89a9aad9.slice\": RecentStats: unable to find data in memory cache]" Nov 26 00:50:37 crc kubenswrapper[4766]: I1126 00:50:37.117975 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 26 00:50:47 crc kubenswrapper[4766]: I1126 00:50:47.838849 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:50:47 crc kubenswrapper[4766]: E1126 00:50:47.840364 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:50:48 crc kubenswrapper[4766]: I1126 00:50:48.905120 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-9259k"] Nov 26 00:50:48 crc kubenswrapper[4766]: I1126 00:50:48.917670 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-9259k"] Nov 26 00:50:49 crc kubenswrapper[4766]: I1126 00:50:49.022194 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-8vfjx"] Nov 26 00:50:49 crc kubenswrapper[4766]: E1126 00:50:49.023131 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed786021-3dca-48e5-b286-f1a06dcbe58e" containerName="extract-utilities" Nov 26 00:50:49 crc kubenswrapper[4766]: I1126 00:50:49.023165 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed786021-3dca-48e5-b286-f1a06dcbe58e" containerName="extract-utilities" Nov 26 00:50:49 crc kubenswrapper[4766]: E1126 00:50:49.023236 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed786021-3dca-48e5-b286-f1a06dcbe58e" containerName="extract-content" Nov 26 00:50:49 crc kubenswrapper[4766]: I1126 00:50:49.023250 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed786021-3dca-48e5-b286-f1a06dcbe58e" containerName="extract-content" Nov 26 00:50:49 crc kubenswrapper[4766]: E1126 00:50:49.023277 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed786021-3dca-48e5-b286-f1a06dcbe58e" containerName="registry-server" Nov 26 00:50:49 crc kubenswrapper[4766]: I1126 00:50:49.023290 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed786021-3dca-48e5-b286-f1a06dcbe58e" containerName="registry-server" Nov 26 00:50:49 crc kubenswrapper[4766]: I1126 00:50:49.023680 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed786021-3dca-48e5-b286-f1a06dcbe58e" containerName="registry-server" Nov 26 00:50:49 crc kubenswrapper[4766]: I1126 00:50:49.024760 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-8vfjx" Nov 26 00:50:49 crc kubenswrapper[4766]: I1126 00:50:49.052691 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-8vfjx"] Nov 26 00:50:49 crc kubenswrapper[4766]: I1126 00:50:49.181010 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzzds\" (UniqueName: \"kubernetes.io/projected/75040a0b-cc48-4e77-9956-ca849299ff7b-kube-api-access-bzzds\") pod \"heat-db-sync-8vfjx\" (UID: \"75040a0b-cc48-4e77-9956-ca849299ff7b\") " pod="openstack/heat-db-sync-8vfjx" Nov 26 00:50:49 crc kubenswrapper[4766]: I1126 00:50:49.181110 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75040a0b-cc48-4e77-9956-ca849299ff7b-combined-ca-bundle\") pod \"heat-db-sync-8vfjx\" (UID: \"75040a0b-cc48-4e77-9956-ca849299ff7b\") " pod="openstack/heat-db-sync-8vfjx" Nov 26 00:50:49 crc kubenswrapper[4766]: I1126 00:50:49.181157 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75040a0b-cc48-4e77-9956-ca849299ff7b-config-data\") pod \"heat-db-sync-8vfjx\" (UID: \"75040a0b-cc48-4e77-9956-ca849299ff7b\") " pod="openstack/heat-db-sync-8vfjx" Nov 26 00:50:49 crc kubenswrapper[4766]: I1126 00:50:49.283063 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzzds\" (UniqueName: \"kubernetes.io/projected/75040a0b-cc48-4e77-9956-ca849299ff7b-kube-api-access-bzzds\") pod \"heat-db-sync-8vfjx\" (UID: \"75040a0b-cc48-4e77-9956-ca849299ff7b\") " pod="openstack/heat-db-sync-8vfjx" Nov 26 00:50:49 crc kubenswrapper[4766]: I1126 00:50:49.283154 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75040a0b-cc48-4e77-9956-ca849299ff7b-combined-ca-bundle\") pod \"heat-db-sync-8vfjx\" (UID: \"75040a0b-cc48-4e77-9956-ca849299ff7b\") " pod="openstack/heat-db-sync-8vfjx" Nov 26 00:50:49 crc kubenswrapper[4766]: I1126 00:50:49.283201 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75040a0b-cc48-4e77-9956-ca849299ff7b-config-data\") pod \"heat-db-sync-8vfjx\" (UID: \"75040a0b-cc48-4e77-9956-ca849299ff7b\") " pod="openstack/heat-db-sync-8vfjx" Nov 26 00:50:49 crc kubenswrapper[4766]: I1126 00:50:49.289471 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75040a0b-cc48-4e77-9956-ca849299ff7b-combined-ca-bundle\") pod \"heat-db-sync-8vfjx\" (UID: \"75040a0b-cc48-4e77-9956-ca849299ff7b\") " pod="openstack/heat-db-sync-8vfjx" Nov 26 00:50:49 crc kubenswrapper[4766]: I1126 00:50:49.295627 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75040a0b-cc48-4e77-9956-ca849299ff7b-config-data\") pod \"heat-db-sync-8vfjx\" (UID: \"75040a0b-cc48-4e77-9956-ca849299ff7b\") " pod="openstack/heat-db-sync-8vfjx" Nov 26 00:50:49 crc kubenswrapper[4766]: I1126 00:50:49.316734 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzzds\" (UniqueName: \"kubernetes.io/projected/75040a0b-cc48-4e77-9956-ca849299ff7b-kube-api-access-bzzds\") pod \"heat-db-sync-8vfjx\" (UID: \"75040a0b-cc48-4e77-9956-ca849299ff7b\") " pod="openstack/heat-db-sync-8vfjx" Nov 26 00:50:49 crc kubenswrapper[4766]: I1126 00:50:49.352670 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-8vfjx" Nov 26 00:50:49 crc kubenswrapper[4766]: I1126 00:50:49.837153 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fd45623-8db1-440a-960a-a848abbb0c90" path="/var/lib/kubelet/pods/2fd45623-8db1-440a-960a-a848abbb0c90/volumes" Nov 26 00:50:49 crc kubenswrapper[4766]: I1126 00:50:49.870296 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-8vfjx"] Nov 26 00:50:50 crc kubenswrapper[4766]: I1126 00:50:50.354394 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-8vfjx" event={"ID":"75040a0b-cc48-4e77-9956-ca849299ff7b","Type":"ContainerStarted","Data":"7d971d0341e5ea2c8acb2a1e9a8a1b1e72e08606994c9a09824a8575c0e661d7"} Nov 26 00:50:50 crc kubenswrapper[4766]: I1126 00:50:50.811576 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 26 00:50:51 crc kubenswrapper[4766]: I1126 00:50:51.270206 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:50:51 crc kubenswrapper[4766]: I1126 00:50:51.270891 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbc768d1-ef8f-4c8f-baca-fbde7c820793" containerName="ceilometer-central-agent" containerID="cri-o://efcafb1d3b6ab9a688d80751a6a8e50d35e35ca981bb7ae8b6994d88b053d737" gracePeriod=30 Nov 26 00:50:51 crc kubenswrapper[4766]: I1126 00:50:51.270993 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbc768d1-ef8f-4c8f-baca-fbde7c820793" containerName="sg-core" containerID="cri-o://5715943ffc4af11e634fdc21dcee90a4abe8b3afe0a72df09d8a87d8e626eeec" gracePeriod=30 Nov 26 00:50:51 crc kubenswrapper[4766]: I1126 00:50:51.270909 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbc768d1-ef8f-4c8f-baca-fbde7c820793" containerName="proxy-httpd" containerID="cri-o://d3d64f79241bbd8529f11a9ada179a12bb7068bb63af1244e64a039b89e414ac" gracePeriod=30 Nov 26 00:50:51 crc kubenswrapper[4766]: I1126 00:50:51.270960 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbc768d1-ef8f-4c8f-baca-fbde7c820793" containerName="ceilometer-notification-agent" containerID="cri-o://4b5c5b7cc746ea7f62128481c01f60c0ad13e847bbbc90f3e1d919a11559ccac" gracePeriod=30 Nov 26 00:50:51 crc kubenswrapper[4766]: I1126 00:50:51.763269 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 26 00:50:52 crc kubenswrapper[4766]: I1126 00:50:52.393427 4766 generic.go:334] "Generic (PLEG): container finished" podID="dbc768d1-ef8f-4c8f-baca-fbde7c820793" containerID="d3d64f79241bbd8529f11a9ada179a12bb7068bb63af1244e64a039b89e414ac" exitCode=0 Nov 26 00:50:52 crc kubenswrapper[4766]: I1126 00:50:52.393772 4766 generic.go:334] "Generic (PLEG): container finished" podID="dbc768d1-ef8f-4c8f-baca-fbde7c820793" containerID="5715943ffc4af11e634fdc21dcee90a4abe8b3afe0a72df09d8a87d8e626eeec" exitCode=2 Nov 26 00:50:52 crc kubenswrapper[4766]: I1126 00:50:52.393511 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbc768d1-ef8f-4c8f-baca-fbde7c820793","Type":"ContainerDied","Data":"d3d64f79241bbd8529f11a9ada179a12bb7068bb63af1244e64a039b89e414ac"} Nov 26 00:50:52 crc kubenswrapper[4766]: I1126 00:50:52.393832 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbc768d1-ef8f-4c8f-baca-fbde7c820793","Type":"ContainerDied","Data":"5715943ffc4af11e634fdc21dcee90a4abe8b3afe0a72df09d8a87d8e626eeec"} Nov 26 00:50:52 crc kubenswrapper[4766]: I1126 00:50:52.393848 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbc768d1-ef8f-4c8f-baca-fbde7c820793","Type":"ContainerDied","Data":"efcafb1d3b6ab9a688d80751a6a8e50d35e35ca981bb7ae8b6994d88b053d737"} Nov 26 00:50:52 crc kubenswrapper[4766]: I1126 00:50:52.393785 4766 generic.go:334] "Generic (PLEG): container finished" podID="dbc768d1-ef8f-4c8f-baca-fbde7c820793" containerID="efcafb1d3b6ab9a688d80751a6a8e50d35e35ca981bb7ae8b6994d88b053d737" exitCode=0 Nov 26 00:50:54 crc kubenswrapper[4766]: I1126 00:50:54.841172 4766 scope.go:117] "RemoveContainer" containerID="3230fd677c493d3befdf9f609024a491fa15783fbb4f924c1484a0015f258849" Nov 26 00:50:54 crc kubenswrapper[4766]: I1126 00:50:54.870860 4766 scope.go:117] "RemoveContainer" containerID="6d403e118929942f889cbb9cc1f0a5ff12be8b1d9c96aadbde1c60c3b7d47810" Nov 26 00:50:55 crc kubenswrapper[4766]: I1126 00:50:55.633891 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="8a944def-677c-4bdc-a7df-2586d0c7937f" containerName="rabbitmq" containerID="cri-o://8a31ae18469d427edc200ce8bf40458cbdf1027220aa0c2d71e7e146412710b9" gracePeriod=604796 Nov 26 00:50:56 crc kubenswrapper[4766]: I1126 00:50:56.051799 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="0527e2fe-678e-4712-ac14-3f2e1f3f0fe0" containerName="rabbitmq" containerID="cri-o://2406d6b7cfd725312646e13f515b83a477e306195d4a6f5b4240ac62783379e0" gracePeriod=604796 Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.260855 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.361630 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-ceilometer-tls-certs\") pod \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.361751 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-sg-core-conf-yaml\") pod \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.361824 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbc768d1-ef8f-4c8f-baca-fbde7c820793-log-httpd\") pod \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.361884 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbc768d1-ef8f-4c8f-baca-fbde7c820793-run-httpd\") pod \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.361939 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-config-data\") pod \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.362247 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbc768d1-ef8f-4c8f-baca-fbde7c820793-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "dbc768d1-ef8f-4c8f-baca-fbde7c820793" (UID: "dbc768d1-ef8f-4c8f-baca-fbde7c820793"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.362293 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbc768d1-ef8f-4c8f-baca-fbde7c820793-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "dbc768d1-ef8f-4c8f-baca-fbde7c820793" (UID: "dbc768d1-ef8f-4c8f-baca-fbde7c820793"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.362467 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfvzv\" (UniqueName: \"kubernetes.io/projected/dbc768d1-ef8f-4c8f-baca-fbde7c820793-kube-api-access-vfvzv\") pod \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.362529 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-scripts\") pod \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.362568 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-combined-ca-bundle\") pod \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\" (UID: \"dbc768d1-ef8f-4c8f-baca-fbde7c820793\") " Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.363109 4766 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbc768d1-ef8f-4c8f-baca-fbde7c820793-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.363124 4766 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbc768d1-ef8f-4c8f-baca-fbde7c820793-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.369339 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbc768d1-ef8f-4c8f-baca-fbde7c820793-kube-api-access-vfvzv" (OuterVolumeSpecName: "kube-api-access-vfvzv") pod "dbc768d1-ef8f-4c8f-baca-fbde7c820793" (UID: "dbc768d1-ef8f-4c8f-baca-fbde7c820793"). InnerVolumeSpecName "kube-api-access-vfvzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.379903 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-scripts" (OuterVolumeSpecName: "scripts") pod "dbc768d1-ef8f-4c8f-baca-fbde7c820793" (UID: "dbc768d1-ef8f-4c8f-baca-fbde7c820793"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.405066 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "dbc768d1-ef8f-4c8f-baca-fbde7c820793" (UID: "dbc768d1-ef8f-4c8f-baca-fbde7c820793"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.424323 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "dbc768d1-ef8f-4c8f-baca-fbde7c820793" (UID: "dbc768d1-ef8f-4c8f-baca-fbde7c820793"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.474335 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dbc768d1-ef8f-4c8f-baca-fbde7c820793" (UID: "dbc768d1-ef8f-4c8f-baca-fbde7c820793"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.479003 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfvzv\" (UniqueName: \"kubernetes.io/projected/dbc768d1-ef8f-4c8f-baca-fbde7c820793-kube-api-access-vfvzv\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.479040 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.479067 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.479080 4766 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.479091 4766 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.492711 4766 generic.go:334] "Generic (PLEG): container finished" podID="dbc768d1-ef8f-4c8f-baca-fbde7c820793" containerID="4b5c5b7cc746ea7f62128481c01f60c0ad13e847bbbc90f3e1d919a11559ccac" exitCode=0 Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.492759 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbc768d1-ef8f-4c8f-baca-fbde7c820793","Type":"ContainerDied","Data":"4b5c5b7cc746ea7f62128481c01f60c0ad13e847bbbc90f3e1d919a11559ccac"} Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.492789 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbc768d1-ef8f-4c8f-baca-fbde7c820793","Type":"ContainerDied","Data":"ca46fb3dea11a816ae02a9e7618138ad9c9c8c74ab94ed9135cfce07533476eb"} Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.492805 4766 scope.go:117] "RemoveContainer" containerID="d3d64f79241bbd8529f11a9ada179a12bb7068bb63af1244e64a039b89e414ac" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.492970 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.528253 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-config-data" (OuterVolumeSpecName: "config-data") pod "dbc768d1-ef8f-4c8f-baca-fbde7c820793" (UID: "dbc768d1-ef8f-4c8f-baca-fbde7c820793"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.581087 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbc768d1-ef8f-4c8f-baca-fbde7c820793-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.853084 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.861857 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.878463 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:50:57 crc kubenswrapper[4766]: E1126 00:50:57.879095 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbc768d1-ef8f-4c8f-baca-fbde7c820793" containerName="proxy-httpd" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.879187 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbc768d1-ef8f-4c8f-baca-fbde7c820793" containerName="proxy-httpd" Nov 26 00:50:57 crc kubenswrapper[4766]: E1126 00:50:57.879270 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbc768d1-ef8f-4c8f-baca-fbde7c820793" containerName="sg-core" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.879320 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbc768d1-ef8f-4c8f-baca-fbde7c820793" containerName="sg-core" Nov 26 00:50:57 crc kubenswrapper[4766]: E1126 00:50:57.879393 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbc768d1-ef8f-4c8f-baca-fbde7c820793" containerName="ceilometer-central-agent" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.879443 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbc768d1-ef8f-4c8f-baca-fbde7c820793" containerName="ceilometer-central-agent" Nov 26 00:50:57 crc kubenswrapper[4766]: E1126 00:50:57.879508 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbc768d1-ef8f-4c8f-baca-fbde7c820793" containerName="ceilometer-notification-agent" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.879564 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbc768d1-ef8f-4c8f-baca-fbde7c820793" containerName="ceilometer-notification-agent" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.879809 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbc768d1-ef8f-4c8f-baca-fbde7c820793" containerName="sg-core" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.879893 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbc768d1-ef8f-4c8f-baca-fbde7c820793" containerName="proxy-httpd" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.879962 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbc768d1-ef8f-4c8f-baca-fbde7c820793" containerName="ceilometer-central-agent" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.880018 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbc768d1-ef8f-4c8f-baca-fbde7c820793" containerName="ceilometer-notification-agent" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.881937 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.889202 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.893985 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.894233 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.894574 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.993478 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-scripts\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.993624 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.993772 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a22c5084-e53e-4529-ba7f-31468f99fe79-run-httpd\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.993866 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.994135 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9rfm\" (UniqueName: \"kubernetes.io/projected/a22c5084-e53e-4529-ba7f-31468f99fe79-kube-api-access-g9rfm\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.994223 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-config-data\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.994259 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:57 crc kubenswrapper[4766]: I1126 00:50:57.994371 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a22c5084-e53e-4529-ba7f-31468f99fe79-log-httpd\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:58 crc kubenswrapper[4766]: I1126 00:50:58.099558 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:58 crc kubenswrapper[4766]: I1126 00:50:58.099692 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9rfm\" (UniqueName: \"kubernetes.io/projected/a22c5084-e53e-4529-ba7f-31468f99fe79-kube-api-access-g9rfm\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:58 crc kubenswrapper[4766]: I1126 00:50:58.099730 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-config-data\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:58 crc kubenswrapper[4766]: I1126 00:50:58.099753 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:58 crc kubenswrapper[4766]: I1126 00:50:58.099993 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a22c5084-e53e-4529-ba7f-31468f99fe79-log-httpd\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:58 crc kubenswrapper[4766]: I1126 00:50:58.100106 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-scripts\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:58 crc kubenswrapper[4766]: I1126 00:50:58.100158 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:58 crc kubenswrapper[4766]: I1126 00:50:58.100194 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a22c5084-e53e-4529-ba7f-31468f99fe79-run-httpd\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:58 crc kubenswrapper[4766]: I1126 00:50:58.100886 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a22c5084-e53e-4529-ba7f-31468f99fe79-run-httpd\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:58 crc kubenswrapper[4766]: I1126 00:50:58.101172 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a22c5084-e53e-4529-ba7f-31468f99fe79-log-httpd\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:58 crc kubenswrapper[4766]: I1126 00:50:58.104403 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:58 crc kubenswrapper[4766]: I1126 00:50:58.104966 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-config-data\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:58 crc kubenswrapper[4766]: I1126 00:50:58.105172 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:58 crc kubenswrapper[4766]: I1126 00:50:58.105539 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:58 crc kubenswrapper[4766]: I1126 00:50:58.106419 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-scripts\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:58 crc kubenswrapper[4766]: I1126 00:50:58.115824 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9rfm\" (UniqueName: \"kubernetes.io/projected/a22c5084-e53e-4529-ba7f-31468f99fe79-kube-api-access-g9rfm\") pod \"ceilometer-0\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " pod="openstack/ceilometer-0" Nov 26 00:50:58 crc kubenswrapper[4766]: I1126 00:50:58.248707 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 00:50:59 crc kubenswrapper[4766]: I1126 00:50:59.835303 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:50:59 crc kubenswrapper[4766]: E1126 00:50:59.835846 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:50:59 crc kubenswrapper[4766]: I1126 00:50:59.844960 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbc768d1-ef8f-4c8f-baca-fbde7c820793" path="/var/lib/kubelet/pods/dbc768d1-ef8f-4c8f-baca-fbde7c820793/volumes" Nov 26 00:51:02 crc kubenswrapper[4766]: I1126 00:51:02.550391 4766 generic.go:334] "Generic (PLEG): container finished" podID="0527e2fe-678e-4712-ac14-3f2e1f3f0fe0" containerID="2406d6b7cfd725312646e13f515b83a477e306195d4a6f5b4240ac62783379e0" exitCode=0 Nov 26 00:51:02 crc kubenswrapper[4766]: I1126 00:51:02.550484 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0","Type":"ContainerDied","Data":"2406d6b7cfd725312646e13f515b83a477e306195d4a6f5b4240ac62783379e0"} Nov 26 00:51:02 crc kubenswrapper[4766]: I1126 00:51:02.553247 4766 generic.go:334] "Generic (PLEG): container finished" podID="8a944def-677c-4bdc-a7df-2586d0c7937f" containerID="8a31ae18469d427edc200ce8bf40458cbdf1027220aa0c2d71e7e146412710b9" exitCode=0 Nov 26 00:51:02 crc kubenswrapper[4766]: I1126 00:51:02.553273 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8a944def-677c-4bdc-a7df-2586d0c7937f","Type":"ContainerDied","Data":"8a31ae18469d427edc200ce8bf40458cbdf1027220aa0c2d71e7e146412710b9"} Nov 26 00:51:05 crc kubenswrapper[4766]: I1126 00:51:05.963066 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-7tt4r"] Nov 26 00:51:05 crc kubenswrapper[4766]: I1126 00:51:05.966316 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:05 crc kubenswrapper[4766]: I1126 00:51:05.969814 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 26 00:51:05 crc kubenswrapper[4766]: I1126 00:51:05.987983 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-7tt4r"] Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.082245 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-ovsdbserver-nb\") pod \"dnsmasq-dns-5b75489c6f-7tt4r\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.082296 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-config\") pod \"dnsmasq-dns-5b75489c6f-7tt4r\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.082355 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9hwf\" (UniqueName: \"kubernetes.io/projected/0064e89a-855d-4d53-8649-9e0745d5c463-kube-api-access-w9hwf\") pod \"dnsmasq-dns-5b75489c6f-7tt4r\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.082382 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-dns-swift-storage-0\") pod \"dnsmasq-dns-5b75489c6f-7tt4r\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.082606 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b75489c6f-7tt4r\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.082723 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-ovsdbserver-sb\") pod \"dnsmasq-dns-5b75489c6f-7tt4r\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.082780 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-dns-svc\") pod \"dnsmasq-dns-5b75489c6f-7tt4r\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.184684 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-dns-svc\") pod \"dnsmasq-dns-5b75489c6f-7tt4r\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.184769 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-ovsdbserver-nb\") pod \"dnsmasq-dns-5b75489c6f-7tt4r\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.184803 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-config\") pod \"dnsmasq-dns-5b75489c6f-7tt4r\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.184864 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9hwf\" (UniqueName: \"kubernetes.io/projected/0064e89a-855d-4d53-8649-9e0745d5c463-kube-api-access-w9hwf\") pod \"dnsmasq-dns-5b75489c6f-7tt4r\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.184890 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-dns-swift-storage-0\") pod \"dnsmasq-dns-5b75489c6f-7tt4r\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.184947 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b75489c6f-7tt4r\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.184980 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-ovsdbserver-sb\") pod \"dnsmasq-dns-5b75489c6f-7tt4r\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.185832 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-ovsdbserver-sb\") pod \"dnsmasq-dns-5b75489c6f-7tt4r\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.187597 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-dns-swift-storage-0\") pod \"dnsmasq-dns-5b75489c6f-7tt4r\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.187639 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b75489c6f-7tt4r\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.187773 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-dns-svc\") pod \"dnsmasq-dns-5b75489c6f-7tt4r\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.188374 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-config\") pod \"dnsmasq-dns-5b75489c6f-7tt4r\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.188882 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-ovsdbserver-nb\") pod \"dnsmasq-dns-5b75489c6f-7tt4r\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.211666 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9hwf\" (UniqueName: \"kubernetes.io/projected/0064e89a-855d-4d53-8649-9e0745d5c463-kube-api-access-w9hwf\") pod \"dnsmasq-dns-5b75489c6f-7tt4r\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:06 crc kubenswrapper[4766]: I1126 00:51:06.287314 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.677159 4766 scope.go:117] "RemoveContainer" containerID="5715943ffc4af11e634fdc21dcee90a4abe8b3afe0a72df09d8a87d8e626eeec" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.804808 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.817047 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.843624 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8a944def-677c-4bdc-a7df-2586d0c7937f-plugins-conf\") pod \"8a944def-677c-4bdc-a7df-2586d0c7937f\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.843676 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a944def-677c-4bdc-a7df-2586d0c7937f-config-data\") pod \"8a944def-677c-4bdc-a7df-2586d0c7937f\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.843695 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8a944def-677c-4bdc-a7df-2586d0c7937f-server-conf\") pod \"8a944def-677c-4bdc-a7df-2586d0c7937f\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.843717 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-tls\") pod \"8a944def-677c-4bdc-a7df-2586d0c7937f\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.843735 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"8a944def-677c-4bdc-a7df-2586d0c7937f\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.843856 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-confd\") pod \"8a944def-677c-4bdc-a7df-2586d0c7937f\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.843928 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8a944def-677c-4bdc-a7df-2586d0c7937f-pod-info\") pod \"8a944def-677c-4bdc-a7df-2586d0c7937f\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.843976 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dpv8\" (UniqueName: \"kubernetes.io/projected/8a944def-677c-4bdc-a7df-2586d0c7937f-kube-api-access-5dpv8\") pod \"8a944def-677c-4bdc-a7df-2586d0c7937f\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.843998 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-plugins\") pod \"8a944def-677c-4bdc-a7df-2586d0c7937f\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.844034 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8a944def-677c-4bdc-a7df-2586d0c7937f-erlang-cookie-secret\") pod \"8a944def-677c-4bdc-a7df-2586d0c7937f\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.844086 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-erlang-cookie\") pod \"8a944def-677c-4bdc-a7df-2586d0c7937f\" (UID: \"8a944def-677c-4bdc-a7df-2586d0c7937f\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.845174 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "8a944def-677c-4bdc-a7df-2586d0c7937f" (UID: "8a944def-677c-4bdc-a7df-2586d0c7937f"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.855231 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a944def-677c-4bdc-a7df-2586d0c7937f-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "8a944def-677c-4bdc-a7df-2586d0c7937f" (UID: "8a944def-677c-4bdc-a7df-2586d0c7937f"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.874037 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "8a944def-677c-4bdc-a7df-2586d0c7937f" (UID: "8a944def-677c-4bdc-a7df-2586d0c7937f"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.881774 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "8a944def-677c-4bdc-a7df-2586d0c7937f" (UID: "8a944def-677c-4bdc-a7df-2586d0c7937f"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.891504 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/8a944def-677c-4bdc-a7df-2586d0c7937f-pod-info" (OuterVolumeSpecName: "pod-info") pod "8a944def-677c-4bdc-a7df-2586d0c7937f" (UID: "8a944def-677c-4bdc-a7df-2586d0c7937f"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.891987 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a944def-677c-4bdc-a7df-2586d0c7937f-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "8a944def-677c-4bdc-a7df-2586d0c7937f" (UID: "8a944def-677c-4bdc-a7df-2586d0c7937f"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.894825 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "8a944def-677c-4bdc-a7df-2586d0c7937f" (UID: "8a944def-677c-4bdc-a7df-2586d0c7937f"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.899012 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a944def-677c-4bdc-a7df-2586d0c7937f-config-data" (OuterVolumeSpecName: "config-data") pod "8a944def-677c-4bdc-a7df-2586d0c7937f" (UID: "8a944def-677c-4bdc-a7df-2586d0c7937f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.905732 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a944def-677c-4bdc-a7df-2586d0c7937f-kube-api-access-5dpv8" (OuterVolumeSpecName: "kube-api-access-5dpv8") pod "8a944def-677c-4bdc-a7df-2586d0c7937f" (UID: "8a944def-677c-4bdc-a7df-2586d0c7937f"). InnerVolumeSpecName "kube-api-access-5dpv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.966628 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.966768 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-plugins\") pod \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.966800 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-confd\") pod \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.966853 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-pod-info\") pod \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.966882 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-tls\") pod \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.966968 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-config-data\") pod \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.967013 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-erlang-cookie\") pod \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.967041 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqrq7\" (UniqueName: \"kubernetes.io/projected/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-kube-api-access-dqrq7\") pod \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.967064 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-erlang-cookie-secret\") pod \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.967078 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-plugins-conf\") pod \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.967119 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-server-conf\") pod \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\" (UID: \"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0\") " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.968186 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dpv8\" (UniqueName: \"kubernetes.io/projected/8a944def-677c-4bdc-a7df-2586d0c7937f-kube-api-access-5dpv8\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.968202 4766 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.968212 4766 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8a944def-677c-4bdc-a7df-2586d0c7937f-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.968221 4766 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.968231 4766 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8a944def-677c-4bdc-a7df-2586d0c7937f-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.968242 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a944def-677c-4bdc-a7df-2586d0c7937f-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.968250 4766 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.968268 4766 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.968280 4766 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8a944def-677c-4bdc-a7df-2586d0c7937f-pod-info\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.972400 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0" (UID: "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.973254 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0" (UID: "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.973834 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0" (UID: "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.976037 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0" (UID: "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:51:08 crc kubenswrapper[4766]: I1126 00:51:08.985164 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0" (UID: "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.004099 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-pod-info" (OuterVolumeSpecName: "pod-info") pod "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0" (UID: "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.031738 4766 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.031374 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-kube-api-access-dqrq7" (OuterVolumeSpecName: "kube-api-access-dqrq7") pod "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0" (UID: "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0"). InnerVolumeSpecName "kube-api-access-dqrq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.056775 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-config-data" (OuterVolumeSpecName: "config-data") pod "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0" (UID: "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.062455 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0" (UID: "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.070821 4766 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.070853 4766 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.070863 4766 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-pod-info\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.070871 4766 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.070882 4766 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.070891 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.070901 4766 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.070911 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqrq7\" (UniqueName: \"kubernetes.io/projected/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-kube-api-access-dqrq7\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.070918 4766 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.070926 4766 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.101954 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a944def-677c-4bdc-a7df-2586d0c7937f-server-conf" (OuterVolumeSpecName: "server-conf") pod "8a944def-677c-4bdc-a7df-2586d0c7937f" (UID: "8a944def-677c-4bdc-a7df-2586d0c7937f"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.117436 4766 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.168105 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-server-conf" (OuterVolumeSpecName: "server-conf") pod "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0" (UID: "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.173260 4766 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.173292 4766 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8a944def-677c-4bdc-a7df-2586d0c7937f-server-conf\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.173303 4766 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-server-conf\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.180094 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "8a944def-677c-4bdc-a7df-2586d0c7937f" (UID: "8a944def-677c-4bdc-a7df-2586d0c7937f"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.188029 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0" (UID: "0527e2fe-678e-4712-ac14-3f2e1f3f0fe0"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.276067 4766 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8a944def-677c-4bdc-a7df-2586d0c7937f-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.276095 4766 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.409437 4766 scope.go:117] "RemoveContainer" containerID="4b5c5b7cc746ea7f62128481c01f60c0ad13e847bbbc90f3e1d919a11559ccac" Nov 26 00:51:09 crc kubenswrapper[4766]: E1126 00:51:09.425998 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Nov 26 00:51:09 crc kubenswrapper[4766]: E1126 00:51:09.426062 4766 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Nov 26 00:51:09 crc kubenswrapper[4766]: E1126 00:51:09.426222 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bzzds,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-8vfjx_openstack(75040a0b-cc48-4e77-9956-ca849299ff7b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 00:51:09 crc kubenswrapper[4766]: E1126 00:51:09.427395 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-8vfjx" podUID="75040a0b-cc48-4e77-9956-ca849299ff7b" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.458790 4766 scope.go:117] "RemoveContainer" containerID="efcafb1d3b6ab9a688d80751a6a8e50d35e35ca981bb7ae8b6994d88b053d737" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.515540 4766 scope.go:117] "RemoveContainer" containerID="d3d64f79241bbd8529f11a9ada179a12bb7068bb63af1244e64a039b89e414ac" Nov 26 00:51:09 crc kubenswrapper[4766]: E1126 00:51:09.516341 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3d64f79241bbd8529f11a9ada179a12bb7068bb63af1244e64a039b89e414ac\": container with ID starting with d3d64f79241bbd8529f11a9ada179a12bb7068bb63af1244e64a039b89e414ac not found: ID does not exist" containerID="d3d64f79241bbd8529f11a9ada179a12bb7068bb63af1244e64a039b89e414ac" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.516380 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3d64f79241bbd8529f11a9ada179a12bb7068bb63af1244e64a039b89e414ac"} err="failed to get container status \"d3d64f79241bbd8529f11a9ada179a12bb7068bb63af1244e64a039b89e414ac\": rpc error: code = NotFound desc = could not find container \"d3d64f79241bbd8529f11a9ada179a12bb7068bb63af1244e64a039b89e414ac\": container with ID starting with d3d64f79241bbd8529f11a9ada179a12bb7068bb63af1244e64a039b89e414ac not found: ID does not exist" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.516405 4766 scope.go:117] "RemoveContainer" containerID="5715943ffc4af11e634fdc21dcee90a4abe8b3afe0a72df09d8a87d8e626eeec" Nov 26 00:51:09 crc kubenswrapper[4766]: E1126 00:51:09.520253 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5715943ffc4af11e634fdc21dcee90a4abe8b3afe0a72df09d8a87d8e626eeec\": container with ID starting with 5715943ffc4af11e634fdc21dcee90a4abe8b3afe0a72df09d8a87d8e626eeec not found: ID does not exist" containerID="5715943ffc4af11e634fdc21dcee90a4abe8b3afe0a72df09d8a87d8e626eeec" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.520309 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5715943ffc4af11e634fdc21dcee90a4abe8b3afe0a72df09d8a87d8e626eeec"} err="failed to get container status \"5715943ffc4af11e634fdc21dcee90a4abe8b3afe0a72df09d8a87d8e626eeec\": rpc error: code = NotFound desc = could not find container \"5715943ffc4af11e634fdc21dcee90a4abe8b3afe0a72df09d8a87d8e626eeec\": container with ID starting with 5715943ffc4af11e634fdc21dcee90a4abe8b3afe0a72df09d8a87d8e626eeec not found: ID does not exist" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.520329 4766 scope.go:117] "RemoveContainer" containerID="4b5c5b7cc746ea7f62128481c01f60c0ad13e847bbbc90f3e1d919a11559ccac" Nov 26 00:51:09 crc kubenswrapper[4766]: E1126 00:51:09.520642 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b5c5b7cc746ea7f62128481c01f60c0ad13e847bbbc90f3e1d919a11559ccac\": container with ID starting with 4b5c5b7cc746ea7f62128481c01f60c0ad13e847bbbc90f3e1d919a11559ccac not found: ID does not exist" containerID="4b5c5b7cc746ea7f62128481c01f60c0ad13e847bbbc90f3e1d919a11559ccac" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.520683 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b5c5b7cc746ea7f62128481c01f60c0ad13e847bbbc90f3e1d919a11559ccac"} err="failed to get container status \"4b5c5b7cc746ea7f62128481c01f60c0ad13e847bbbc90f3e1d919a11559ccac\": rpc error: code = NotFound desc = could not find container \"4b5c5b7cc746ea7f62128481c01f60c0ad13e847bbbc90f3e1d919a11559ccac\": container with ID starting with 4b5c5b7cc746ea7f62128481c01f60c0ad13e847bbbc90f3e1d919a11559ccac not found: ID does not exist" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.520703 4766 scope.go:117] "RemoveContainer" containerID="efcafb1d3b6ab9a688d80751a6a8e50d35e35ca981bb7ae8b6994d88b053d737" Nov 26 00:51:09 crc kubenswrapper[4766]: E1126 00:51:09.521014 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efcafb1d3b6ab9a688d80751a6a8e50d35e35ca981bb7ae8b6994d88b053d737\": container with ID starting with efcafb1d3b6ab9a688d80751a6a8e50d35e35ca981bb7ae8b6994d88b053d737 not found: ID does not exist" containerID="efcafb1d3b6ab9a688d80751a6a8e50d35e35ca981bb7ae8b6994d88b053d737" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.521036 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efcafb1d3b6ab9a688d80751a6a8e50d35e35ca981bb7ae8b6994d88b053d737"} err="failed to get container status \"efcafb1d3b6ab9a688d80751a6a8e50d35e35ca981bb7ae8b6994d88b053d737\": rpc error: code = NotFound desc = could not find container \"efcafb1d3b6ab9a688d80751a6a8e50d35e35ca981bb7ae8b6994d88b053d737\": container with ID starting with efcafb1d3b6ab9a688d80751a6a8e50d35e35ca981bb7ae8b6994d88b053d737 not found: ID does not exist" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.637502 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="8a944def-677c-4bdc-a7df-2586d0c7937f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.120:5671: i/o timeout" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.655108 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.654962 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8a944def-677c-4bdc-a7df-2586d0c7937f","Type":"ContainerDied","Data":"ea70652be2b194e5c24fcf97542f3dab2c21f84e848aa50dd6610284c7b5f461"} Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.655418 4766 scope.go:117] "RemoveContainer" containerID="8a31ae18469d427edc200ce8bf40458cbdf1027220aa0c2d71e7e146412710b9" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.661316 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.663907 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0527e2fe-678e-4712-ac14-3f2e1f3f0fe0","Type":"ContainerDied","Data":"40456bb04133e9b6fce53ed7c23091720a5c5fed300ddcb1d404ce30ba23c0af"} Nov 26 00:51:09 crc kubenswrapper[4766]: E1126 00:51:09.665385 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-8vfjx" podUID="75040a0b-cc48-4e77-9956-ca849299ff7b" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.781101 4766 scope.go:117] "RemoveContainer" containerID="ce25f81214edc08620b9866c07d4fdf518d951e03f1f623a3eeb20ca690ef230" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.789425 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.823604 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.835909 4766 scope.go:117] "RemoveContainer" containerID="2406d6b7cfd725312646e13f515b83a477e306195d4a6f5b4240ac62783379e0" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.854576 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0527e2fe-678e-4712-ac14-3f2e1f3f0fe0" path="/var/lib/kubelet/pods/0527e2fe-678e-4712-ac14-3f2e1f3f0fe0/volumes" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.855416 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.857405 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.880527 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 26 00:51:09 crc kubenswrapper[4766]: E1126 00:51:09.881047 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a944def-677c-4bdc-a7df-2586d0c7937f" containerName="setup-container" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.881067 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a944def-677c-4bdc-a7df-2586d0c7937f" containerName="setup-container" Nov 26 00:51:09 crc kubenswrapper[4766]: E1126 00:51:09.881078 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0527e2fe-678e-4712-ac14-3f2e1f3f0fe0" containerName="rabbitmq" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.881084 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0527e2fe-678e-4712-ac14-3f2e1f3f0fe0" containerName="rabbitmq" Nov 26 00:51:09 crc kubenswrapper[4766]: E1126 00:51:09.881102 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0527e2fe-678e-4712-ac14-3f2e1f3f0fe0" containerName="setup-container" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.881109 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0527e2fe-678e-4712-ac14-3f2e1f3f0fe0" containerName="setup-container" Nov 26 00:51:09 crc kubenswrapper[4766]: E1126 00:51:09.881121 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a944def-677c-4bdc-a7df-2586d0c7937f" containerName="rabbitmq" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.881127 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a944def-677c-4bdc-a7df-2586d0c7937f" containerName="rabbitmq" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.881368 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="0527e2fe-678e-4712-ac14-3f2e1f3f0fe0" containerName="rabbitmq" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.881401 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a944def-677c-4bdc-a7df-2586d0c7937f" containerName="rabbitmq" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.883807 4766 scope.go:117] "RemoveContainer" containerID="84d5f79a02963d431e4b2b22e8c479a860d1676c7ccd73a598d1c7c0726230be" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.907427 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.911585 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.911932 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.911644 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.911787 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.911863 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.911904 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.912009 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-llt2n" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.912284 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.939378 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="0527e2fe-678e-4712-ac14-3f2e1f3f0fe0" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.121:5671: i/o timeout" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.940263 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.957544 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.962792 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.965366 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7g6b2" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.965502 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.965389 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.965755 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.965936 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.966637 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.967335 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.969840 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.980342 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-7tt4r"] Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.990765 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.990821 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.990884 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.990905 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.990935 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.990954 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdjlj\" (UniqueName: \"kubernetes.io/projected/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-kube-api-access-gdjlj\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.990988 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.991056 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.991091 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.991110 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:09 crc kubenswrapper[4766]: I1126 00:51:09.991125 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.092846 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.092895 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/797e86ef-8e0d-499e-859f-6f6732826d7e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.092936 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.092988 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.093008 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/797e86ef-8e0d-499e-859f-6f6732826d7e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.093030 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.093055 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.093075 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdjlj\" (UniqueName: \"kubernetes.io/projected/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-kube-api-access-gdjlj\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.093092 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.093116 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/797e86ef-8e0d-499e-859f-6f6732826d7e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.093153 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.093170 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/797e86ef-8e0d-499e-859f-6f6732826d7e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.093208 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/797e86ef-8e0d-499e-859f-6f6732826d7e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.093233 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/797e86ef-8e0d-499e-859f-6f6732826d7e-config-data\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.093252 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/797e86ef-8e0d-499e-859f-6f6732826d7e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.093271 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/797e86ef-8e0d-499e-859f-6f6732826d7e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.093289 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.093308 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/797e86ef-8e0d-499e-859f-6f6732826d7e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.093342 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.093363 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.093379 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.093411 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wpn8\" (UniqueName: \"kubernetes.io/projected/797e86ef-8e0d-499e-859f-6f6732826d7e-kube-api-access-8wpn8\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.094515 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.095086 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.095327 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.096327 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.096710 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.097057 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.099482 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.099502 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.100049 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.100759 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.112123 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdjlj\" (UniqueName: \"kubernetes.io/projected/64400ff7-9c3a-42c3-b1fc-2f9696f1206c-kube-api-access-gdjlj\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.138560 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"64400ff7-9c3a-42c3-b1fc-2f9696f1206c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.195446 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wpn8\" (UniqueName: \"kubernetes.io/projected/797e86ef-8e0d-499e-859f-6f6732826d7e-kube-api-access-8wpn8\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.195506 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/797e86ef-8e0d-499e-859f-6f6732826d7e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.195574 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/797e86ef-8e0d-499e-859f-6f6732826d7e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.195607 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.195628 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/797e86ef-8e0d-499e-859f-6f6732826d7e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.195667 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/797e86ef-8e0d-499e-859f-6f6732826d7e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.195702 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/797e86ef-8e0d-499e-859f-6f6732826d7e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.195728 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/797e86ef-8e0d-499e-859f-6f6732826d7e-config-data\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.195747 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/797e86ef-8e0d-499e-859f-6f6732826d7e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.195767 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/797e86ef-8e0d-499e-859f-6f6732826d7e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.196128 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/797e86ef-8e0d-499e-859f-6f6732826d7e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.196225 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.196270 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/797e86ef-8e0d-499e-859f-6f6732826d7e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.196854 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/797e86ef-8e0d-499e-859f-6f6732826d7e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.197213 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/797e86ef-8e0d-499e-859f-6f6732826d7e-config-data\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.197380 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/797e86ef-8e0d-499e-859f-6f6732826d7e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.197749 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/797e86ef-8e0d-499e-859f-6f6732826d7e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.199390 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/797e86ef-8e0d-499e-859f-6f6732826d7e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.199403 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/797e86ef-8e0d-499e-859f-6f6732826d7e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.203217 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/797e86ef-8e0d-499e-859f-6f6732826d7e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.209400 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/797e86ef-8e0d-499e-859f-6f6732826d7e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.213758 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wpn8\" (UniqueName: \"kubernetes.io/projected/797e86ef-8e0d-499e-859f-6f6732826d7e-kube-api-access-8wpn8\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.238548 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"797e86ef-8e0d-499e-859f-6f6732826d7e\") " pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.349782 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.449439 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.679205 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" event={"ID":"0064e89a-855d-4d53-8649-9e0745d5c463","Type":"ContainerStarted","Data":"d76e0de459dca0f5f847a69bd9f9eeb3453f0e89cb6f01fbf417f5cad165e3c9"} Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.679256 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" event={"ID":"0064e89a-855d-4d53-8649-9e0745d5c463","Type":"ContainerStarted","Data":"8f184287c553163106bab689b065fbaa92c4d8d0ce423b7d67d0bfc704489046"} Nov 26 00:51:10 crc kubenswrapper[4766]: I1126 00:51:10.682305 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a22c5084-e53e-4529-ba7f-31468f99fe79","Type":"ContainerStarted","Data":"dbc81db846693815c88a8535746e7cfb37e6bfd385b685d2151dc4dab4d704be"} Nov 26 00:51:11 crc kubenswrapper[4766]: I1126 00:51:11.835459 4766 generic.go:334] "Generic (PLEG): container finished" podID="0064e89a-855d-4d53-8649-9e0745d5c463" containerID="d76e0de459dca0f5f847a69bd9f9eeb3453f0e89cb6f01fbf417f5cad165e3c9" exitCode=0 Nov 26 00:51:11 crc kubenswrapper[4766]: I1126 00:51:11.882348 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a944def-677c-4bdc-a7df-2586d0c7937f" path="/var/lib/kubelet/pods/8a944def-677c-4bdc-a7df-2586d0c7937f/volumes" Nov 26 00:51:11 crc kubenswrapper[4766]: I1126 00:51:11.891544 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" event={"ID":"0064e89a-855d-4d53-8649-9e0745d5c463","Type":"ContainerDied","Data":"d76e0de459dca0f5f847a69bd9f9eeb3453f0e89cb6f01fbf417f5cad165e3c9"} Nov 26 00:51:12 crc kubenswrapper[4766]: I1126 00:51:12.133403 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 26 00:51:12 crc kubenswrapper[4766]: W1126 00:51:12.140779 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64400ff7_9c3a_42c3_b1fc_2f9696f1206c.slice/crio-0242cae3bb042e7c1a0ae8f419e4eeef70d1e82d8083fa87eedbb6db96cbe7e5 WatchSource:0}: Error finding container 0242cae3bb042e7c1a0ae8f419e4eeef70d1e82d8083fa87eedbb6db96cbe7e5: Status 404 returned error can't find the container with id 0242cae3bb042e7c1a0ae8f419e4eeef70d1e82d8083fa87eedbb6db96cbe7e5 Nov 26 00:51:12 crc kubenswrapper[4766]: I1126 00:51:12.347400 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 26 00:51:12 crc kubenswrapper[4766]: W1126 00:51:12.349584 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod797e86ef_8e0d_499e_859f_6f6732826d7e.slice/crio-77ca21ab1e871e987ace0c46eadf0066a032173e223c2089cd606f7426c2ddbb WatchSource:0}: Error finding container 77ca21ab1e871e987ace0c46eadf0066a032173e223c2089cd606f7426c2ddbb: Status 404 returned error can't find the container with id 77ca21ab1e871e987ace0c46eadf0066a032173e223c2089cd606f7426c2ddbb Nov 26 00:51:12 crc kubenswrapper[4766]: I1126 00:51:12.827495 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:51:12 crc kubenswrapper[4766]: E1126 00:51:12.827826 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:51:12 crc kubenswrapper[4766]: I1126 00:51:12.860557 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"797e86ef-8e0d-499e-859f-6f6732826d7e","Type":"ContainerStarted","Data":"77ca21ab1e871e987ace0c46eadf0066a032173e223c2089cd606f7426c2ddbb"} Nov 26 00:51:12 crc kubenswrapper[4766]: I1126 00:51:12.866509 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"64400ff7-9c3a-42c3-b1fc-2f9696f1206c","Type":"ContainerStarted","Data":"0242cae3bb042e7c1a0ae8f419e4eeef70d1e82d8083fa87eedbb6db96cbe7e5"} Nov 26 00:51:12 crc kubenswrapper[4766]: I1126 00:51:12.868510 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" event={"ID":"0064e89a-855d-4d53-8649-9e0745d5c463","Type":"ContainerStarted","Data":"b50f6b68ef7344f2fbd7c4e8f7fb05f97059425bc31e69b389dbd83600f65ccf"} Nov 26 00:51:12 crc kubenswrapper[4766]: I1126 00:51:12.869468 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:12 crc kubenswrapper[4766]: I1126 00:51:12.906839 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" podStartSLOduration=7.90681963 podStartE2EDuration="7.90681963s" podCreationTimestamp="2025-11-26 00:51:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:51:12.894670377 +0000 UTC m=+1653.743440817" watchObservedRunningTime="2025-11-26 00:51:12.90681963 +0000 UTC m=+1653.755590060" Nov 26 00:51:15 crc kubenswrapper[4766]: I1126 00:51:15.900753 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"64400ff7-9c3a-42c3-b1fc-2f9696f1206c","Type":"ContainerStarted","Data":"b6b967648cb30b946648f540cb66b304c9ea44a740abf811907c71add66c3c7f"} Nov 26 00:51:15 crc kubenswrapper[4766]: I1126 00:51:15.909830 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"797e86ef-8e0d-499e-859f-6f6732826d7e","Type":"ContainerStarted","Data":"ac4e43dceb51e8d670e672114cea611a1f004f02b356c30a9b2f9b59f7088978"} Nov 26 00:51:16 crc kubenswrapper[4766]: I1126 00:51:16.923781 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a22c5084-e53e-4529-ba7f-31468f99fe79","Type":"ContainerStarted","Data":"13db0c5081c86540d776fae42cf51d65f73dfc346caeeb8e9af07a15742c3a49"} Nov 26 00:51:16 crc kubenswrapper[4766]: I1126 00:51:16.924306 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a22c5084-e53e-4529-ba7f-31468f99fe79","Type":"ContainerStarted","Data":"98755bd910987eb515252f2f2e9abe4a387ffd403a0971db75e38ef549222632"} Nov 26 00:51:17 crc kubenswrapper[4766]: I1126 00:51:17.947811 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a22c5084-e53e-4529-ba7f-31468f99fe79","Type":"ContainerStarted","Data":"3fd7408fd83a3c1fe26d8c321a6fb0dcf192d0be1a723c7318ac2243a5451575"} Nov 26 00:51:18 crc kubenswrapper[4766]: I1126 00:51:18.972205 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a22c5084-e53e-4529-ba7f-31468f99fe79","Type":"ContainerStarted","Data":"c35f2a49b9d1b4c98a9e3c26b6539ac9cb5060fa828c620187f1bfd114411294"} Nov 26 00:51:18 crc kubenswrapper[4766]: I1126 00:51:18.972747 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 26 00:51:19 crc kubenswrapper[4766]: I1126 00:51:19.001997 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=13.298439279 podStartE2EDuration="22.001975677s" podCreationTimestamp="2025-11-26 00:50:57 +0000 UTC" firstStartedPulling="2025-11-26 00:51:09.891414278 +0000 UTC m=+1650.740184698" lastFinishedPulling="2025-11-26 00:51:18.594950626 +0000 UTC m=+1659.443721096" observedRunningTime="2025-11-26 00:51:18.9980962 +0000 UTC m=+1659.846866640" watchObservedRunningTime="2025-11-26 00:51:19.001975677 +0000 UTC m=+1659.850746117" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.289818 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.432899 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-ff4gf"] Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.433350 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" podUID="90d82ce9-7909-40ac-9c8f-b15b6f68f24f" containerName="dnsmasq-dns" containerID="cri-o://e0fc6733b38804513dacda343aa208bb447365560aaa4baf35e5bbe9390d787a" gracePeriod=10 Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.560483 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9"] Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.562492 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.586285 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9"] Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.737289 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpxqs\" (UniqueName: \"kubernetes.io/projected/f586d76c-85fe-40dc-8158-52a1747703d7-kube-api-access-jpxqs\") pod \"dnsmasq-dns-5cf7b6cbf7-jxdn9\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.737500 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-openstack-edpm-ipam\") pod \"dnsmasq-dns-5cf7b6cbf7-jxdn9\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.737587 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-dns-svc\") pod \"dnsmasq-dns-5cf7b6cbf7-jxdn9\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.737664 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-dns-swift-storage-0\") pod \"dnsmasq-dns-5cf7b6cbf7-jxdn9\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.738232 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-config\") pod \"dnsmasq-dns-5cf7b6cbf7-jxdn9\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.738486 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-ovsdbserver-nb\") pod \"dnsmasq-dns-5cf7b6cbf7-jxdn9\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.738559 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-ovsdbserver-sb\") pod \"dnsmasq-dns-5cf7b6cbf7-jxdn9\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.843818 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-config\") pod \"dnsmasq-dns-5cf7b6cbf7-jxdn9\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.843929 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-ovsdbserver-nb\") pod \"dnsmasq-dns-5cf7b6cbf7-jxdn9\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.843951 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-ovsdbserver-sb\") pod \"dnsmasq-dns-5cf7b6cbf7-jxdn9\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.844021 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpxqs\" (UniqueName: \"kubernetes.io/projected/f586d76c-85fe-40dc-8158-52a1747703d7-kube-api-access-jpxqs\") pod \"dnsmasq-dns-5cf7b6cbf7-jxdn9\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.844075 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-openstack-edpm-ipam\") pod \"dnsmasq-dns-5cf7b6cbf7-jxdn9\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.844093 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-dns-swift-storage-0\") pod \"dnsmasq-dns-5cf7b6cbf7-jxdn9\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.844110 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-dns-svc\") pod \"dnsmasq-dns-5cf7b6cbf7-jxdn9\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.845054 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-dns-svc\") pod \"dnsmasq-dns-5cf7b6cbf7-jxdn9\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.845968 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-config\") pod \"dnsmasq-dns-5cf7b6cbf7-jxdn9\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.845856 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-ovsdbserver-nb\") pod \"dnsmasq-dns-5cf7b6cbf7-jxdn9\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.846245 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-ovsdbserver-sb\") pod \"dnsmasq-dns-5cf7b6cbf7-jxdn9\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.849768 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-dns-swift-storage-0\") pod \"dnsmasq-dns-5cf7b6cbf7-jxdn9\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.855561 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-openstack-edpm-ipam\") pod \"dnsmasq-dns-5cf7b6cbf7-jxdn9\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.869889 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpxqs\" (UniqueName: \"kubernetes.io/projected/f586d76c-85fe-40dc-8158-52a1747703d7-kube-api-access-jpxqs\") pod \"dnsmasq-dns-5cf7b6cbf7-jxdn9\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:21 crc kubenswrapper[4766]: I1126 00:51:21.919286 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.027078 4766 generic.go:334] "Generic (PLEG): container finished" podID="90d82ce9-7909-40ac-9c8f-b15b6f68f24f" containerID="e0fc6733b38804513dacda343aa208bb447365560aaa4baf35e5bbe9390d787a" exitCode=0 Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.027120 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" event={"ID":"90d82ce9-7909-40ac-9c8f-b15b6f68f24f","Type":"ContainerDied","Data":"e0fc6733b38804513dacda343aa208bb447365560aaa4baf35e5bbe9390d787a"} Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.027148 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" event={"ID":"90d82ce9-7909-40ac-9c8f-b15b6f68f24f","Type":"ContainerDied","Data":"0e66e7a89154eecc5b3e106b8cbb55b12a241e065600bc55833aaef92382d7f2"} Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.027159 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e66e7a89154eecc5b3e106b8cbb55b12a241e065600bc55833aaef92382d7f2" Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.115316 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.252604 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-dns-svc\") pod \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.253006 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-ovsdbserver-nb\") pod \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.253060 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-dns-swift-storage-0\") pod \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.253099 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-config\") pod \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.253216 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqp58\" (UniqueName: \"kubernetes.io/projected/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-kube-api-access-dqp58\") pod \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.253243 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-ovsdbserver-sb\") pod \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\" (UID: \"90d82ce9-7909-40ac-9c8f-b15b6f68f24f\") " Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.267201 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-kube-api-access-dqp58" (OuterVolumeSpecName: "kube-api-access-dqp58") pod "90d82ce9-7909-40ac-9c8f-b15b6f68f24f" (UID: "90d82ce9-7909-40ac-9c8f-b15b6f68f24f"). InnerVolumeSpecName "kube-api-access-dqp58". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.328275 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "90d82ce9-7909-40ac-9c8f-b15b6f68f24f" (UID: "90d82ce9-7909-40ac-9c8f-b15b6f68f24f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.336358 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "90d82ce9-7909-40ac-9c8f-b15b6f68f24f" (UID: "90d82ce9-7909-40ac-9c8f-b15b6f68f24f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.344091 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "90d82ce9-7909-40ac-9c8f-b15b6f68f24f" (UID: "90d82ce9-7909-40ac-9c8f-b15b6f68f24f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.355196 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqp58\" (UniqueName: \"kubernetes.io/projected/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-kube-api-access-dqp58\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.355235 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.355249 4766 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.355258 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.355786 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-config" (OuterVolumeSpecName: "config") pod "90d82ce9-7909-40ac-9c8f-b15b6f68f24f" (UID: "90d82ce9-7909-40ac-9c8f-b15b6f68f24f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.357835 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "90d82ce9-7909-40ac-9c8f-b15b6f68f24f" (UID: "90d82ce9-7909-40ac-9c8f-b15b6f68f24f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:51:22 crc kubenswrapper[4766]: W1126 00:51:22.376708 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf586d76c_85fe_40dc_8158_52a1747703d7.slice/crio-904092226a87448a7c622cf5243115d77ef33901904e7b56b2ead915230935d5 WatchSource:0}: Error finding container 904092226a87448a7c622cf5243115d77ef33901904e7b56b2ead915230935d5: Status 404 returned error can't find the container with id 904092226a87448a7c622cf5243115d77ef33901904e7b56b2ead915230935d5 Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.380872 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9"] Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.457586 4766 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:22 crc kubenswrapper[4766]: I1126 00:51:22.457620 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90d82ce9-7909-40ac-9c8f-b15b6f68f24f-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:23 crc kubenswrapper[4766]: I1126 00:51:23.038713 4766 generic.go:334] "Generic (PLEG): container finished" podID="f586d76c-85fe-40dc-8158-52a1747703d7" containerID="394dc6058d6729712a339c628479c02ab45716008c538eead220353c77dcf99b" exitCode=0 Nov 26 00:51:23 crc kubenswrapper[4766]: I1126 00:51:23.038964 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" event={"ID":"f586d76c-85fe-40dc-8158-52a1747703d7","Type":"ContainerDied","Data":"394dc6058d6729712a339c628479c02ab45716008c538eead220353c77dcf99b"} Nov 26 00:51:23 crc kubenswrapper[4766]: I1126 00:51:23.038989 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" event={"ID":"f586d76c-85fe-40dc-8158-52a1747703d7","Type":"ContainerStarted","Data":"904092226a87448a7c622cf5243115d77ef33901904e7b56b2ead915230935d5"} Nov 26 00:51:23 crc kubenswrapper[4766]: I1126 00:51:23.043203 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" Nov 26 00:51:23 crc kubenswrapper[4766]: I1126 00:51:23.043205 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-8vfjx" event={"ID":"75040a0b-cc48-4e77-9956-ca849299ff7b","Type":"ContainerStarted","Data":"17a7b3c12e36d0510101c272211eeed5cbabed5ed57c708c65189e195d9a53d8"} Nov 26 00:51:23 crc kubenswrapper[4766]: I1126 00:51:23.103372 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-8vfjx" podStartSLOduration=2.983440034 podStartE2EDuration="35.103352538s" podCreationTimestamp="2025-11-26 00:50:48 +0000 UTC" firstStartedPulling="2025-11-26 00:50:49.880032241 +0000 UTC m=+1630.728802671" lastFinishedPulling="2025-11-26 00:51:21.999944745 +0000 UTC m=+1662.848715175" observedRunningTime="2025-11-26 00:51:23.097228225 +0000 UTC m=+1663.945998665" watchObservedRunningTime="2025-11-26 00:51:23.103352538 +0000 UTC m=+1663.952122968" Nov 26 00:51:23 crc kubenswrapper[4766]: I1126 00:51:23.258368 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-ff4gf"] Nov 26 00:51:23 crc kubenswrapper[4766]: I1126 00:51:23.270618 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-ff4gf"] Nov 26 00:51:23 crc kubenswrapper[4766]: I1126 00:51:23.849758 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90d82ce9-7909-40ac-9c8f-b15b6f68f24f" path="/var/lib/kubelet/pods/90d82ce9-7909-40ac-9c8f-b15b6f68f24f/volumes" Nov 26 00:51:24 crc kubenswrapper[4766]: I1126 00:51:24.061574 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" event={"ID":"f586d76c-85fe-40dc-8158-52a1747703d7","Type":"ContainerStarted","Data":"38c1d8e99a74fa7e315f5cf2480050f5585758a433f3aec3b382768482a5866c"} Nov 26 00:51:24 crc kubenswrapper[4766]: I1126 00:51:24.061836 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:24 crc kubenswrapper[4766]: I1126 00:51:24.097605 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" podStartSLOduration=3.09758343 podStartE2EDuration="3.09758343s" podCreationTimestamp="2025-11-26 00:51:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:51:24.091201611 +0000 UTC m=+1664.939972061" watchObservedRunningTime="2025-11-26 00:51:24.09758343 +0000 UTC m=+1664.946353870" Nov 26 00:51:25 crc kubenswrapper[4766]: I1126 00:51:25.074672 4766 generic.go:334] "Generic (PLEG): container finished" podID="75040a0b-cc48-4e77-9956-ca849299ff7b" containerID="17a7b3c12e36d0510101c272211eeed5cbabed5ed57c708c65189e195d9a53d8" exitCode=0 Nov 26 00:51:25 crc kubenswrapper[4766]: I1126 00:51:25.074820 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-8vfjx" event={"ID":"75040a0b-cc48-4e77-9956-ca849299ff7b","Type":"ContainerDied","Data":"17a7b3c12e36d0510101c272211eeed5cbabed5ed57c708c65189e195d9a53d8"} Nov 26 00:51:26 crc kubenswrapper[4766]: I1126 00:51:26.562583 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-8vfjx" Nov 26 00:51:26 crc kubenswrapper[4766]: I1126 00:51:26.631972 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75040a0b-cc48-4e77-9956-ca849299ff7b-combined-ca-bundle\") pod \"75040a0b-cc48-4e77-9956-ca849299ff7b\" (UID: \"75040a0b-cc48-4e77-9956-ca849299ff7b\") " Nov 26 00:51:26 crc kubenswrapper[4766]: I1126 00:51:26.632249 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzzds\" (UniqueName: \"kubernetes.io/projected/75040a0b-cc48-4e77-9956-ca849299ff7b-kube-api-access-bzzds\") pod \"75040a0b-cc48-4e77-9956-ca849299ff7b\" (UID: \"75040a0b-cc48-4e77-9956-ca849299ff7b\") " Nov 26 00:51:26 crc kubenswrapper[4766]: I1126 00:51:26.632280 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75040a0b-cc48-4e77-9956-ca849299ff7b-config-data\") pod \"75040a0b-cc48-4e77-9956-ca849299ff7b\" (UID: \"75040a0b-cc48-4e77-9956-ca849299ff7b\") " Nov 26 00:51:26 crc kubenswrapper[4766]: I1126 00:51:26.643332 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75040a0b-cc48-4e77-9956-ca849299ff7b-kube-api-access-bzzds" (OuterVolumeSpecName: "kube-api-access-bzzds") pod "75040a0b-cc48-4e77-9956-ca849299ff7b" (UID: "75040a0b-cc48-4e77-9956-ca849299ff7b"). InnerVolumeSpecName "kube-api-access-bzzds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:51:26 crc kubenswrapper[4766]: I1126 00:51:26.676932 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75040a0b-cc48-4e77-9956-ca849299ff7b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "75040a0b-cc48-4e77-9956-ca849299ff7b" (UID: "75040a0b-cc48-4e77-9956-ca849299ff7b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:26 crc kubenswrapper[4766]: I1126 00:51:26.734841 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzzds\" (UniqueName: \"kubernetes.io/projected/75040a0b-cc48-4e77-9956-ca849299ff7b-kube-api-access-bzzds\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:26 crc kubenswrapper[4766]: I1126 00:51:26.734888 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75040a0b-cc48-4e77-9956-ca849299ff7b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:26 crc kubenswrapper[4766]: I1126 00:51:26.775705 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75040a0b-cc48-4e77-9956-ca849299ff7b-config-data" (OuterVolumeSpecName: "config-data") pod "75040a0b-cc48-4e77-9956-ca849299ff7b" (UID: "75040a0b-cc48-4e77-9956-ca849299ff7b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:26 crc kubenswrapper[4766]: I1126 00:51:26.812080 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-f84f9ccf-ff4gf" podUID="90d82ce9-7909-40ac-9c8f-b15b6f68f24f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.238:5353: i/o timeout" Nov 26 00:51:26 crc kubenswrapper[4766]: I1126 00:51:26.836638 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75040a0b-cc48-4e77-9956-ca849299ff7b-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:27 crc kubenswrapper[4766]: I1126 00:51:27.112074 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-8vfjx" event={"ID":"75040a0b-cc48-4e77-9956-ca849299ff7b","Type":"ContainerDied","Data":"7d971d0341e5ea2c8acb2a1e9a8a1b1e72e08606994c9a09824a8575c0e661d7"} Nov 26 00:51:27 crc kubenswrapper[4766]: I1126 00:51:27.112134 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d971d0341e5ea2c8acb2a1e9a8a1b1e72e08606994c9a09824a8575c0e661d7" Nov 26 00:51:27 crc kubenswrapper[4766]: I1126 00:51:27.112226 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-8vfjx" Nov 26 00:51:27 crc kubenswrapper[4766]: I1126 00:51:27.828617 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:51:27 crc kubenswrapper[4766]: E1126 00:51:27.829384 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.268868 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.738924 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-6ff8c446-6wzh8"] Nov 26 00:51:28 crc kubenswrapper[4766]: E1126 00:51:28.739781 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75040a0b-cc48-4e77-9956-ca849299ff7b" containerName="heat-db-sync" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.739887 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="75040a0b-cc48-4e77-9956-ca849299ff7b" containerName="heat-db-sync" Nov 26 00:51:28 crc kubenswrapper[4766]: E1126 00:51:28.739998 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90d82ce9-7909-40ac-9c8f-b15b6f68f24f" containerName="dnsmasq-dns" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.740077 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="90d82ce9-7909-40ac-9c8f-b15b6f68f24f" containerName="dnsmasq-dns" Nov 26 00:51:28 crc kubenswrapper[4766]: E1126 00:51:28.740165 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90d82ce9-7909-40ac-9c8f-b15b6f68f24f" containerName="init" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.740244 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="90d82ce9-7909-40ac-9c8f-b15b6f68f24f" containerName="init" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.740833 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="75040a0b-cc48-4e77-9956-ca849299ff7b" containerName="heat-db-sync" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.740946 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="90d82ce9-7909-40ac-9c8f-b15b6f68f24f" containerName="dnsmasq-dns" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.742124 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6ff8c446-6wzh8" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.779891 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/775687b1-1657-4115-b482-f7e418e77727-combined-ca-bundle\") pod \"heat-engine-6ff8c446-6wzh8\" (UID: \"775687b1-1657-4115-b482-f7e418e77727\") " pod="openstack/heat-engine-6ff8c446-6wzh8" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.779965 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmj6j\" (UniqueName: \"kubernetes.io/projected/775687b1-1657-4115-b482-f7e418e77727-kube-api-access-fmj6j\") pod \"heat-engine-6ff8c446-6wzh8\" (UID: \"775687b1-1657-4115-b482-f7e418e77727\") " pod="openstack/heat-engine-6ff8c446-6wzh8" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.780012 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/775687b1-1657-4115-b482-f7e418e77727-config-data-custom\") pod \"heat-engine-6ff8c446-6wzh8\" (UID: \"775687b1-1657-4115-b482-f7e418e77727\") " pod="openstack/heat-engine-6ff8c446-6wzh8" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.780078 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/775687b1-1657-4115-b482-f7e418e77727-config-data\") pod \"heat-engine-6ff8c446-6wzh8\" (UID: \"775687b1-1657-4115-b482-f7e418e77727\") " pod="openstack/heat-engine-6ff8c446-6wzh8" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.780420 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6ff8c446-6wzh8"] Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.807061 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-7f6d8f97f4-szt84"] Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.809005 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.818940 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-5ff994f754-l2ctn"] Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.821355 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.886493 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/775687b1-1657-4115-b482-f7e418e77727-config-data-custom\") pod \"heat-engine-6ff8c446-6wzh8\" (UID: \"775687b1-1657-4115-b482-f7e418e77727\") " pod="openstack/heat-engine-6ff8c446-6wzh8" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.887221 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/775687b1-1657-4115-b482-f7e418e77727-config-data\") pod \"heat-engine-6ff8c446-6wzh8\" (UID: \"775687b1-1657-4115-b482-f7e418e77727\") " pod="openstack/heat-engine-6ff8c446-6wzh8" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.887278 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65159db0-64b4-4cce-ba5e-aa6f889f7f58-config-data-custom\") pod \"heat-api-7f6d8f97f4-szt84\" (UID: \"65159db0-64b4-4cce-ba5e-aa6f889f7f58\") " pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.887312 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f65a888-def8-495b-915a-0b6556aff6de-config-data-custom\") pod \"heat-cfnapi-5ff994f754-l2ctn\" (UID: \"5f65a888-def8-495b-915a-0b6556aff6de\") " pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.887379 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65159db0-64b4-4cce-ba5e-aa6f889f7f58-combined-ca-bundle\") pod \"heat-api-7f6d8f97f4-szt84\" (UID: \"65159db0-64b4-4cce-ba5e-aa6f889f7f58\") " pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.892807 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65159db0-64b4-4cce-ba5e-aa6f889f7f58-internal-tls-certs\") pod \"heat-api-7f6d8f97f4-szt84\" (UID: \"65159db0-64b4-4cce-ba5e-aa6f889f7f58\") " pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.892942 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44zkd\" (UniqueName: \"kubernetes.io/projected/5f65a888-def8-495b-915a-0b6556aff6de-kube-api-access-44zkd\") pod \"heat-cfnapi-5ff994f754-l2ctn\" (UID: \"5f65a888-def8-495b-915a-0b6556aff6de\") " pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.892975 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f65a888-def8-495b-915a-0b6556aff6de-internal-tls-certs\") pod \"heat-cfnapi-5ff994f754-l2ctn\" (UID: \"5f65a888-def8-495b-915a-0b6556aff6de\") " pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.893129 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f65a888-def8-495b-915a-0b6556aff6de-public-tls-certs\") pod \"heat-cfnapi-5ff994f754-l2ctn\" (UID: \"5f65a888-def8-495b-915a-0b6556aff6de\") " pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.893222 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2rjb\" (UniqueName: \"kubernetes.io/projected/65159db0-64b4-4cce-ba5e-aa6f889f7f58-kube-api-access-q2rjb\") pod \"heat-api-7f6d8f97f4-szt84\" (UID: \"65159db0-64b4-4cce-ba5e-aa6f889f7f58\") " pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.893271 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65159db0-64b4-4cce-ba5e-aa6f889f7f58-public-tls-certs\") pod \"heat-api-7f6d8f97f4-szt84\" (UID: \"65159db0-64b4-4cce-ba5e-aa6f889f7f58\") " pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.893369 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/775687b1-1657-4115-b482-f7e418e77727-combined-ca-bundle\") pod \"heat-engine-6ff8c446-6wzh8\" (UID: \"775687b1-1657-4115-b482-f7e418e77727\") " pod="openstack/heat-engine-6ff8c446-6wzh8" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.893408 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f65a888-def8-495b-915a-0b6556aff6de-config-data\") pod \"heat-cfnapi-5ff994f754-l2ctn\" (UID: \"5f65a888-def8-495b-915a-0b6556aff6de\") " pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.893438 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65159db0-64b4-4cce-ba5e-aa6f889f7f58-config-data\") pod \"heat-api-7f6d8f97f4-szt84\" (UID: \"65159db0-64b4-4cce-ba5e-aa6f889f7f58\") " pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.893474 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmj6j\" (UniqueName: \"kubernetes.io/projected/775687b1-1657-4115-b482-f7e418e77727-kube-api-access-fmj6j\") pod \"heat-engine-6ff8c446-6wzh8\" (UID: \"775687b1-1657-4115-b482-f7e418e77727\") " pod="openstack/heat-engine-6ff8c446-6wzh8" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.893544 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f65a888-def8-495b-915a-0b6556aff6de-combined-ca-bundle\") pod \"heat-cfnapi-5ff994f754-l2ctn\" (UID: \"5f65a888-def8-495b-915a-0b6556aff6de\") " pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.900996 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/775687b1-1657-4115-b482-f7e418e77727-config-data-custom\") pod \"heat-engine-6ff8c446-6wzh8\" (UID: \"775687b1-1657-4115-b482-f7e418e77727\") " pod="openstack/heat-engine-6ff8c446-6wzh8" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.901419 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/775687b1-1657-4115-b482-f7e418e77727-combined-ca-bundle\") pod \"heat-engine-6ff8c446-6wzh8\" (UID: \"775687b1-1657-4115-b482-f7e418e77727\") " pod="openstack/heat-engine-6ff8c446-6wzh8" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.926309 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/775687b1-1657-4115-b482-f7e418e77727-config-data\") pod \"heat-engine-6ff8c446-6wzh8\" (UID: \"775687b1-1657-4115-b482-f7e418e77727\") " pod="openstack/heat-engine-6ff8c446-6wzh8" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.927948 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmj6j\" (UniqueName: \"kubernetes.io/projected/775687b1-1657-4115-b482-f7e418e77727-kube-api-access-fmj6j\") pod \"heat-engine-6ff8c446-6wzh8\" (UID: \"775687b1-1657-4115-b482-f7e418e77727\") " pod="openstack/heat-engine-6ff8c446-6wzh8" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.930298 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7f6d8f97f4-szt84"] Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.947912 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5ff994f754-l2ctn"] Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.996031 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44zkd\" (UniqueName: \"kubernetes.io/projected/5f65a888-def8-495b-915a-0b6556aff6de-kube-api-access-44zkd\") pod \"heat-cfnapi-5ff994f754-l2ctn\" (UID: \"5f65a888-def8-495b-915a-0b6556aff6de\") " pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.996078 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f65a888-def8-495b-915a-0b6556aff6de-internal-tls-certs\") pod \"heat-cfnapi-5ff994f754-l2ctn\" (UID: \"5f65a888-def8-495b-915a-0b6556aff6de\") " pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.996139 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f65a888-def8-495b-915a-0b6556aff6de-public-tls-certs\") pod \"heat-cfnapi-5ff994f754-l2ctn\" (UID: \"5f65a888-def8-495b-915a-0b6556aff6de\") " pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.996174 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2rjb\" (UniqueName: \"kubernetes.io/projected/65159db0-64b4-4cce-ba5e-aa6f889f7f58-kube-api-access-q2rjb\") pod \"heat-api-7f6d8f97f4-szt84\" (UID: \"65159db0-64b4-4cce-ba5e-aa6f889f7f58\") " pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.996194 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65159db0-64b4-4cce-ba5e-aa6f889f7f58-public-tls-certs\") pod \"heat-api-7f6d8f97f4-szt84\" (UID: \"65159db0-64b4-4cce-ba5e-aa6f889f7f58\") " pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.996232 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f65a888-def8-495b-915a-0b6556aff6de-config-data\") pod \"heat-cfnapi-5ff994f754-l2ctn\" (UID: \"5f65a888-def8-495b-915a-0b6556aff6de\") " pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.996251 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65159db0-64b4-4cce-ba5e-aa6f889f7f58-config-data\") pod \"heat-api-7f6d8f97f4-szt84\" (UID: \"65159db0-64b4-4cce-ba5e-aa6f889f7f58\") " pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.996283 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f65a888-def8-495b-915a-0b6556aff6de-combined-ca-bundle\") pod \"heat-cfnapi-5ff994f754-l2ctn\" (UID: \"5f65a888-def8-495b-915a-0b6556aff6de\") " pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.996333 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65159db0-64b4-4cce-ba5e-aa6f889f7f58-config-data-custom\") pod \"heat-api-7f6d8f97f4-szt84\" (UID: \"65159db0-64b4-4cce-ba5e-aa6f889f7f58\") " pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.996352 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f65a888-def8-495b-915a-0b6556aff6de-config-data-custom\") pod \"heat-cfnapi-5ff994f754-l2ctn\" (UID: \"5f65a888-def8-495b-915a-0b6556aff6de\") " pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.996384 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65159db0-64b4-4cce-ba5e-aa6f889f7f58-combined-ca-bundle\") pod \"heat-api-7f6d8f97f4-szt84\" (UID: \"65159db0-64b4-4cce-ba5e-aa6f889f7f58\") " pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:28 crc kubenswrapper[4766]: I1126 00:51:28.996421 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65159db0-64b4-4cce-ba5e-aa6f889f7f58-internal-tls-certs\") pod \"heat-api-7f6d8f97f4-szt84\" (UID: \"65159db0-64b4-4cce-ba5e-aa6f889f7f58\") " pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.000127 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65159db0-64b4-4cce-ba5e-aa6f889f7f58-internal-tls-certs\") pod \"heat-api-7f6d8f97f4-szt84\" (UID: \"65159db0-64b4-4cce-ba5e-aa6f889f7f58\") " pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.000805 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65159db0-64b4-4cce-ba5e-aa6f889f7f58-combined-ca-bundle\") pod \"heat-api-7f6d8f97f4-szt84\" (UID: \"65159db0-64b4-4cce-ba5e-aa6f889f7f58\") " pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.001534 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65159db0-64b4-4cce-ba5e-aa6f889f7f58-public-tls-certs\") pod \"heat-api-7f6d8f97f4-szt84\" (UID: \"65159db0-64b4-4cce-ba5e-aa6f889f7f58\") " pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.001974 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f65a888-def8-495b-915a-0b6556aff6de-config-data-custom\") pod \"heat-cfnapi-5ff994f754-l2ctn\" (UID: \"5f65a888-def8-495b-915a-0b6556aff6de\") " pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.002504 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f65a888-def8-495b-915a-0b6556aff6de-config-data\") pod \"heat-cfnapi-5ff994f754-l2ctn\" (UID: \"5f65a888-def8-495b-915a-0b6556aff6de\") " pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.003066 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f65a888-def8-495b-915a-0b6556aff6de-public-tls-certs\") pod \"heat-cfnapi-5ff994f754-l2ctn\" (UID: \"5f65a888-def8-495b-915a-0b6556aff6de\") " pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.003341 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65159db0-64b4-4cce-ba5e-aa6f889f7f58-config-data\") pod \"heat-api-7f6d8f97f4-szt84\" (UID: \"65159db0-64b4-4cce-ba5e-aa6f889f7f58\") " pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.003435 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f65a888-def8-495b-915a-0b6556aff6de-combined-ca-bundle\") pod \"heat-cfnapi-5ff994f754-l2ctn\" (UID: \"5f65a888-def8-495b-915a-0b6556aff6de\") " pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.009977 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f65a888-def8-495b-915a-0b6556aff6de-internal-tls-certs\") pod \"heat-cfnapi-5ff994f754-l2ctn\" (UID: \"5f65a888-def8-495b-915a-0b6556aff6de\") " pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.013138 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65159db0-64b4-4cce-ba5e-aa6f889f7f58-config-data-custom\") pod \"heat-api-7f6d8f97f4-szt84\" (UID: \"65159db0-64b4-4cce-ba5e-aa6f889f7f58\") " pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.015641 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2rjb\" (UniqueName: \"kubernetes.io/projected/65159db0-64b4-4cce-ba5e-aa6f889f7f58-kube-api-access-q2rjb\") pod \"heat-api-7f6d8f97f4-szt84\" (UID: \"65159db0-64b4-4cce-ba5e-aa6f889f7f58\") " pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.019636 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44zkd\" (UniqueName: \"kubernetes.io/projected/5f65a888-def8-495b-915a-0b6556aff6de-kube-api-access-44zkd\") pod \"heat-cfnapi-5ff994f754-l2ctn\" (UID: \"5f65a888-def8-495b-915a-0b6556aff6de\") " pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.114120 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6ff8c446-6wzh8" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.121643 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-mfr52"] Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.138009 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-mfr52"] Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.141745 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.174881 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-tz998"] Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.176795 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-tz998" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.178806 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.186020 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.187040 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-tz998"] Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.199740 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d6086e9-30aa-4a56-9233-182c811c85c2-combined-ca-bundle\") pod \"aodh-db-sync-tz998\" (UID: \"6d6086e9-30aa-4a56-9233-182c811c85c2\") " pod="openstack/aodh-db-sync-tz998" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.200082 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqz8d\" (UniqueName: \"kubernetes.io/projected/6d6086e9-30aa-4a56-9233-182c811c85c2-kube-api-access-dqz8d\") pod \"aodh-db-sync-tz998\" (UID: \"6d6086e9-30aa-4a56-9233-182c811c85c2\") " pod="openstack/aodh-db-sync-tz998" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.200152 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d6086e9-30aa-4a56-9233-182c811c85c2-config-data\") pod \"aodh-db-sync-tz998\" (UID: \"6d6086e9-30aa-4a56-9233-182c811c85c2\") " pod="openstack/aodh-db-sync-tz998" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.200388 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d6086e9-30aa-4a56-9233-182c811c85c2-scripts\") pod \"aodh-db-sync-tz998\" (UID: \"6d6086e9-30aa-4a56-9233-182c811c85c2\") " pod="openstack/aodh-db-sync-tz998" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.302460 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d6086e9-30aa-4a56-9233-182c811c85c2-combined-ca-bundle\") pod \"aodh-db-sync-tz998\" (UID: \"6d6086e9-30aa-4a56-9233-182c811c85c2\") " pod="openstack/aodh-db-sync-tz998" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.302802 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqz8d\" (UniqueName: \"kubernetes.io/projected/6d6086e9-30aa-4a56-9233-182c811c85c2-kube-api-access-dqz8d\") pod \"aodh-db-sync-tz998\" (UID: \"6d6086e9-30aa-4a56-9233-182c811c85c2\") " pod="openstack/aodh-db-sync-tz998" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.302838 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d6086e9-30aa-4a56-9233-182c811c85c2-config-data\") pod \"aodh-db-sync-tz998\" (UID: \"6d6086e9-30aa-4a56-9233-182c811c85c2\") " pod="openstack/aodh-db-sync-tz998" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.302924 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d6086e9-30aa-4a56-9233-182c811c85c2-scripts\") pod \"aodh-db-sync-tz998\" (UID: \"6d6086e9-30aa-4a56-9233-182c811c85c2\") " pod="openstack/aodh-db-sync-tz998" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.311428 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d6086e9-30aa-4a56-9233-182c811c85c2-scripts\") pod \"aodh-db-sync-tz998\" (UID: \"6d6086e9-30aa-4a56-9233-182c811c85c2\") " pod="openstack/aodh-db-sync-tz998" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.311746 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d6086e9-30aa-4a56-9233-182c811c85c2-combined-ca-bundle\") pod \"aodh-db-sync-tz998\" (UID: \"6d6086e9-30aa-4a56-9233-182c811c85c2\") " pod="openstack/aodh-db-sync-tz998" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.314147 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d6086e9-30aa-4a56-9233-182c811c85c2-config-data\") pod \"aodh-db-sync-tz998\" (UID: \"6d6086e9-30aa-4a56-9233-182c811c85c2\") " pod="openstack/aodh-db-sync-tz998" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.342553 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqz8d\" (UniqueName: \"kubernetes.io/projected/6d6086e9-30aa-4a56-9233-182c811c85c2-kube-api-access-dqz8d\") pod \"aodh-db-sync-tz998\" (UID: \"6d6086e9-30aa-4a56-9233-182c811c85c2\") " pod="openstack/aodh-db-sync-tz998" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.523006 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-tz998" Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.664125 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6ff8c446-6wzh8"] Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.796608 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5ff994f754-l2ctn"] Nov 26 00:51:29 crc kubenswrapper[4766]: W1126 00:51:29.802044 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f65a888_def8_495b_915a_0b6556aff6de.slice/crio-459077c9c650e3619870ac95b98324386ad5daa3c3e01ff8cd1c539e662cff8a WatchSource:0}: Error finding container 459077c9c650e3619870ac95b98324386ad5daa3c3e01ff8cd1c539e662cff8a: Status 404 returned error can't find the container with id 459077c9c650e3619870ac95b98324386ad5daa3c3e01ff8cd1c539e662cff8a Nov 26 00:51:29 crc kubenswrapper[4766]: W1126 00:51:29.805265 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65159db0_64b4_4cce_ba5e_aa6f889f7f58.slice/crio-e1b97008764e9144c1a35544920de021619335d470a54c89fd510b34f610d4b2 WatchSource:0}: Error finding container e1b97008764e9144c1a35544920de021619335d470a54c89fd510b34f610d4b2: Status 404 returned error can't find the container with id e1b97008764e9144c1a35544920de021619335d470a54c89fd510b34f610d4b2 Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.811703 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7f6d8f97f4-szt84"] Nov 26 00:51:29 crc kubenswrapper[4766]: I1126 00:51:29.842347 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="638a59e9-68c3-43cf-ae3b-4510e7df42ec" path="/var/lib/kubelet/pods/638a59e9-68c3-43cf-ae3b-4510e7df42ec/volumes" Nov 26 00:51:30 crc kubenswrapper[4766]: I1126 00:51:30.079251 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-tz998"] Nov 26 00:51:30 crc kubenswrapper[4766]: I1126 00:51:30.168664 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5ff994f754-l2ctn" event={"ID":"5f65a888-def8-495b-915a-0b6556aff6de","Type":"ContainerStarted","Data":"459077c9c650e3619870ac95b98324386ad5daa3c3e01ff8cd1c539e662cff8a"} Nov 26 00:51:30 crc kubenswrapper[4766]: I1126 00:51:30.170462 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6ff8c446-6wzh8" event={"ID":"775687b1-1657-4115-b482-f7e418e77727","Type":"ContainerStarted","Data":"d79c97e9f2de4ea912f554ad27b282c9170aa0ce9a04e9c8ece1fc27ba0b68d1"} Nov 26 00:51:30 crc kubenswrapper[4766]: I1126 00:51:30.170508 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6ff8c446-6wzh8" event={"ID":"775687b1-1657-4115-b482-f7e418e77727","Type":"ContainerStarted","Data":"587dbc7d021c103fc2b25d2154d971fc60cdf4654e7d90d81e501201ffe1b7ae"} Nov 26 00:51:30 crc kubenswrapper[4766]: I1126 00:51:30.170589 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-6ff8c446-6wzh8" Nov 26 00:51:30 crc kubenswrapper[4766]: I1126 00:51:30.172909 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7f6d8f97f4-szt84" event={"ID":"65159db0-64b4-4cce-ba5e-aa6f889f7f58","Type":"ContainerStarted","Data":"e1b97008764e9144c1a35544920de021619335d470a54c89fd510b34f610d4b2"} Nov 26 00:51:30 crc kubenswrapper[4766]: I1126 00:51:30.174489 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-tz998" event={"ID":"6d6086e9-30aa-4a56-9233-182c811c85c2","Type":"ContainerStarted","Data":"ab37600bc642843d41a21ba874ea2168fd3a8e19f0fa0bf45ab2883f0381bf7e"} Nov 26 00:51:30 crc kubenswrapper[4766]: I1126 00:51:30.189442 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-6ff8c446-6wzh8" podStartSLOduration=2.189423986 podStartE2EDuration="2.189423986s" podCreationTimestamp="2025-11-26 00:51:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:51:30.184150644 +0000 UTC m=+1671.032921084" watchObservedRunningTime="2025-11-26 00:51:30.189423986 +0000 UTC m=+1671.038194416" Nov 26 00:51:31 crc kubenswrapper[4766]: I1126 00:51:31.922754 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.001351 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-7tt4r"] Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.001574 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" podUID="0064e89a-855d-4d53-8649-9e0745d5c463" containerName="dnsmasq-dns" containerID="cri-o://b50f6b68ef7344f2fbd7c4e8f7fb05f97059425bc31e69b389dbd83600f65ccf" gracePeriod=10 Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.216939 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5ff994f754-l2ctn" event={"ID":"5f65a888-def8-495b-915a-0b6556aff6de","Type":"ContainerStarted","Data":"4dd65246b15d49307c8ff440f42c271ee559b303ba689e0bf99cdb02a36c6644"} Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.218416 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.233236 4766 generic.go:334] "Generic (PLEG): container finished" podID="0064e89a-855d-4d53-8649-9e0745d5c463" containerID="b50f6b68ef7344f2fbd7c4e8f7fb05f97059425bc31e69b389dbd83600f65ccf" exitCode=0 Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.233279 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" event={"ID":"0064e89a-855d-4d53-8649-9e0745d5c463","Type":"ContainerDied","Data":"b50f6b68ef7344f2fbd7c4e8f7fb05f97059425bc31e69b389dbd83600f65ccf"} Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.237249 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-5ff994f754-l2ctn" podStartSLOduration=2.2615198899999998 podStartE2EDuration="4.237234408s" podCreationTimestamp="2025-11-26 00:51:28 +0000 UTC" firstStartedPulling="2025-11-26 00:51:29.806660818 +0000 UTC m=+1670.655431248" lastFinishedPulling="2025-11-26 00:51:31.782375336 +0000 UTC m=+1672.631145766" observedRunningTime="2025-11-26 00:51:32.23369239 +0000 UTC m=+1673.082462820" watchObservedRunningTime="2025-11-26 00:51:32.237234408 +0000 UTC m=+1673.086004838" Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.504970 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.697031 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9hwf\" (UniqueName: \"kubernetes.io/projected/0064e89a-855d-4d53-8649-9e0745d5c463-kube-api-access-w9hwf\") pod \"0064e89a-855d-4d53-8649-9e0745d5c463\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.697308 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-ovsdbserver-sb\") pod \"0064e89a-855d-4d53-8649-9e0745d5c463\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.697449 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-dns-svc\") pod \"0064e89a-855d-4d53-8649-9e0745d5c463\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.697498 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-config\") pod \"0064e89a-855d-4d53-8649-9e0745d5c463\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.697543 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-ovsdbserver-nb\") pod \"0064e89a-855d-4d53-8649-9e0745d5c463\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.697596 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-openstack-edpm-ipam\") pod \"0064e89a-855d-4d53-8649-9e0745d5c463\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.697691 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-dns-swift-storage-0\") pod \"0064e89a-855d-4d53-8649-9e0745d5c463\" (UID: \"0064e89a-855d-4d53-8649-9e0745d5c463\") " Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.712949 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0064e89a-855d-4d53-8649-9e0745d5c463-kube-api-access-w9hwf" (OuterVolumeSpecName: "kube-api-access-w9hwf") pod "0064e89a-855d-4d53-8649-9e0745d5c463" (UID: "0064e89a-855d-4d53-8649-9e0745d5c463"). InnerVolumeSpecName "kube-api-access-w9hwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.762260 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0064e89a-855d-4d53-8649-9e0745d5c463" (UID: "0064e89a-855d-4d53-8649-9e0745d5c463"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.779750 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "0064e89a-855d-4d53-8649-9e0745d5c463" (UID: "0064e89a-855d-4d53-8649-9e0745d5c463"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.780602 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0064e89a-855d-4d53-8649-9e0745d5c463" (UID: "0064e89a-855d-4d53-8649-9e0745d5c463"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.794617 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0064e89a-855d-4d53-8649-9e0745d5c463" (UID: "0064e89a-855d-4d53-8649-9e0745d5c463"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.796948 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0064e89a-855d-4d53-8649-9e0745d5c463" (UID: "0064e89a-855d-4d53-8649-9e0745d5c463"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.800936 4766 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.800964 4766 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.800975 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9hwf\" (UniqueName: \"kubernetes.io/projected/0064e89a-855d-4d53-8649-9e0745d5c463-kube-api-access-w9hwf\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.800985 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.800996 4766 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.801011 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.806477 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-config" (OuterVolumeSpecName: "config") pod "0064e89a-855d-4d53-8649-9e0745d5c463" (UID: "0064e89a-855d-4d53-8649-9e0745d5c463"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 00:51:32 crc kubenswrapper[4766]: I1126 00:51:32.903430 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0064e89a-855d-4d53-8649-9e0745d5c463-config\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:33 crc kubenswrapper[4766]: I1126 00:51:33.252104 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7f6d8f97f4-szt84" event={"ID":"65159db0-64b4-4cce-ba5e-aa6f889f7f58","Type":"ContainerStarted","Data":"ec98724918b127fd368b4fbaaa0269851a56492fba241d4110d3f23f7ba60387"} Nov 26 00:51:33 crc kubenswrapper[4766]: I1126 00:51:33.252465 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:33 crc kubenswrapper[4766]: I1126 00:51:33.256086 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" event={"ID":"0064e89a-855d-4d53-8649-9e0745d5c463","Type":"ContainerDied","Data":"8f184287c553163106bab689b065fbaa92c4d8d0ce423b7d67d0bfc704489046"} Nov 26 00:51:33 crc kubenswrapper[4766]: I1126 00:51:33.256155 4766 scope.go:117] "RemoveContainer" containerID="b50f6b68ef7344f2fbd7c4e8f7fb05f97059425bc31e69b389dbd83600f65ccf" Nov 26 00:51:33 crc kubenswrapper[4766]: I1126 00:51:33.256212 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-7tt4r" Nov 26 00:51:33 crc kubenswrapper[4766]: I1126 00:51:33.283295 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-7f6d8f97f4-szt84" podStartSLOduration=3.305023361 podStartE2EDuration="5.283268812s" podCreationTimestamp="2025-11-26 00:51:28 +0000 UTC" firstStartedPulling="2025-11-26 00:51:29.808133125 +0000 UTC m=+1670.656903555" lastFinishedPulling="2025-11-26 00:51:31.786378576 +0000 UTC m=+1672.635149006" observedRunningTime="2025-11-26 00:51:33.273981681 +0000 UTC m=+1674.122752111" watchObservedRunningTime="2025-11-26 00:51:33.283268812 +0000 UTC m=+1674.132039272" Nov 26 00:51:33 crc kubenswrapper[4766]: I1126 00:51:33.308850 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-7tt4r"] Nov 26 00:51:33 crc kubenswrapper[4766]: I1126 00:51:33.322554 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-7tt4r"] Nov 26 00:51:33 crc kubenswrapper[4766]: I1126 00:51:33.846940 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0064e89a-855d-4d53-8649-9e0745d5c463" path="/var/lib/kubelet/pods/0064e89a-855d-4d53-8649-9e0745d5c463/volumes" Nov 26 00:51:35 crc kubenswrapper[4766]: I1126 00:51:35.548324 4766 scope.go:117] "RemoveContainer" containerID="d76e0de459dca0f5f847a69bd9f9eeb3453f0e89cb6f01fbf417f5cad165e3c9" Nov 26 00:51:36 crc kubenswrapper[4766]: I1126 00:51:36.295706 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-tz998" event={"ID":"6d6086e9-30aa-4a56-9233-182c811c85c2","Type":"ContainerStarted","Data":"b0221c71033eb66079348bb9d44485237f3dfeeeb515394898f7d0d95d0b38bd"} Nov 26 00:51:36 crc kubenswrapper[4766]: I1126 00:51:36.315340 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-tz998" podStartSLOduration=1.806433588 podStartE2EDuration="7.315324369s" podCreationTimestamp="2025-11-26 00:51:29 +0000 UTC" firstStartedPulling="2025-11-26 00:51:30.131970954 +0000 UTC m=+1670.980741384" lastFinishedPulling="2025-11-26 00:51:35.640861715 +0000 UTC m=+1676.489632165" observedRunningTime="2025-11-26 00:51:36.312744845 +0000 UTC m=+1677.161515315" watchObservedRunningTime="2025-11-26 00:51:36.315324369 +0000 UTC m=+1677.164094789" Nov 26 00:51:38 crc kubenswrapper[4766]: I1126 00:51:38.319585 4766 generic.go:334] "Generic (PLEG): container finished" podID="6d6086e9-30aa-4a56-9233-182c811c85c2" containerID="b0221c71033eb66079348bb9d44485237f3dfeeeb515394898f7d0d95d0b38bd" exitCode=0 Nov 26 00:51:38 crc kubenswrapper[4766]: I1126 00:51:38.319691 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-tz998" event={"ID":"6d6086e9-30aa-4a56-9233-182c811c85c2","Type":"ContainerDied","Data":"b0221c71033eb66079348bb9d44485237f3dfeeeb515394898f7d0d95d0b38bd"} Nov 26 00:51:39 crc kubenswrapper[4766]: I1126 00:51:39.160120 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-6ff8c446-6wzh8" Nov 26 00:51:39 crc kubenswrapper[4766]: I1126 00:51:39.207457 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-668d6b6d67-r2rrd"] Nov 26 00:51:39 crc kubenswrapper[4766]: I1126 00:51:39.207685 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-668d6b6d67-r2rrd" podUID="c0d62d37-c9d8-46f3-8fcd-802578c573fb" containerName="heat-engine" containerID="cri-o://c5c61611bbcbcf3f57f04c4c93929a9c73ad6b0ae1d455e0599b25ed4f3f2da3" gracePeriod=60 Nov 26 00:51:39 crc kubenswrapper[4766]: I1126 00:51:39.726178 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-tz998" Nov 26 00:51:39 crc kubenswrapper[4766]: I1126 00:51:39.876954 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d6086e9-30aa-4a56-9233-182c811c85c2-scripts\") pod \"6d6086e9-30aa-4a56-9233-182c811c85c2\" (UID: \"6d6086e9-30aa-4a56-9233-182c811c85c2\") " Nov 26 00:51:39 crc kubenswrapper[4766]: I1126 00:51:39.877050 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d6086e9-30aa-4a56-9233-182c811c85c2-combined-ca-bundle\") pod \"6d6086e9-30aa-4a56-9233-182c811c85c2\" (UID: \"6d6086e9-30aa-4a56-9233-182c811c85c2\") " Nov 26 00:51:39 crc kubenswrapper[4766]: I1126 00:51:39.877600 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqz8d\" (UniqueName: \"kubernetes.io/projected/6d6086e9-30aa-4a56-9233-182c811c85c2-kube-api-access-dqz8d\") pod \"6d6086e9-30aa-4a56-9233-182c811c85c2\" (UID: \"6d6086e9-30aa-4a56-9233-182c811c85c2\") " Nov 26 00:51:39 crc kubenswrapper[4766]: I1126 00:51:39.877688 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d6086e9-30aa-4a56-9233-182c811c85c2-config-data\") pod \"6d6086e9-30aa-4a56-9233-182c811c85c2\" (UID: \"6d6086e9-30aa-4a56-9233-182c811c85c2\") " Nov 26 00:51:39 crc kubenswrapper[4766]: I1126 00:51:39.883820 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d6086e9-30aa-4a56-9233-182c811c85c2-kube-api-access-dqz8d" (OuterVolumeSpecName: "kube-api-access-dqz8d") pod "6d6086e9-30aa-4a56-9233-182c811c85c2" (UID: "6d6086e9-30aa-4a56-9233-182c811c85c2"). InnerVolumeSpecName "kube-api-access-dqz8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:51:39 crc kubenswrapper[4766]: I1126 00:51:39.919453 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d6086e9-30aa-4a56-9233-182c811c85c2-scripts" (OuterVolumeSpecName: "scripts") pod "6d6086e9-30aa-4a56-9233-182c811c85c2" (UID: "6d6086e9-30aa-4a56-9233-182c811c85c2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:39 crc kubenswrapper[4766]: I1126 00:51:39.925135 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d6086e9-30aa-4a56-9233-182c811c85c2-config-data" (OuterVolumeSpecName: "config-data") pod "6d6086e9-30aa-4a56-9233-182c811c85c2" (UID: "6d6086e9-30aa-4a56-9233-182c811c85c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:39 crc kubenswrapper[4766]: I1126 00:51:39.940740 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d6086e9-30aa-4a56-9233-182c811c85c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d6086e9-30aa-4a56-9233-182c811c85c2" (UID: "6d6086e9-30aa-4a56-9233-182c811c85c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:39 crc kubenswrapper[4766]: I1126 00:51:39.981882 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d6086e9-30aa-4a56-9233-182c811c85c2-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:39 crc kubenswrapper[4766]: I1126 00:51:39.981939 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d6086e9-30aa-4a56-9233-182c811c85c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:39 crc kubenswrapper[4766]: I1126 00:51:39.981962 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqz8d\" (UniqueName: \"kubernetes.io/projected/6d6086e9-30aa-4a56-9233-182c811c85c2-kube-api-access-dqz8d\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:39 crc kubenswrapper[4766]: I1126 00:51:39.981980 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d6086e9-30aa-4a56-9233-182c811c85c2-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.060009 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt"] Nov 26 00:51:40 crc kubenswrapper[4766]: E1126 00:51:40.061868 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0064e89a-855d-4d53-8649-9e0745d5c463" containerName="dnsmasq-dns" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.061968 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0064e89a-855d-4d53-8649-9e0745d5c463" containerName="dnsmasq-dns" Nov 26 00:51:40 crc kubenswrapper[4766]: E1126 00:51:40.062061 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0064e89a-855d-4d53-8649-9e0745d5c463" containerName="init" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.062122 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0064e89a-855d-4d53-8649-9e0745d5c463" containerName="init" Nov 26 00:51:40 crc kubenswrapper[4766]: E1126 00:51:40.062198 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d6086e9-30aa-4a56-9233-182c811c85c2" containerName="aodh-db-sync" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.062249 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d6086e9-30aa-4a56-9233-182c811c85c2" containerName="aodh-db-sync" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.062510 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="0064e89a-855d-4d53-8649-9e0745d5c463" containerName="dnsmasq-dns" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.062592 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d6086e9-30aa-4a56-9233-182c811c85c2" containerName="aodh-db-sync" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.063460 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.065496 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.065825 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.066046 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.070214 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.075676 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt"] Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.201773 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt\" (UID: \"7e4f73e8-58b4-47c3-83c6-75cc10ae1866\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.201809 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt\" (UID: \"7e4f73e8-58b4-47c3-83c6-75cc10ae1866\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.201857 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt\" (UID: \"7e4f73e8-58b4-47c3-83c6-75cc10ae1866\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.201961 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bzvz\" (UniqueName: \"kubernetes.io/projected/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-kube-api-access-7bzvz\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt\" (UID: \"7e4f73e8-58b4-47c3-83c6-75cc10ae1866\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.304010 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bzvz\" (UniqueName: \"kubernetes.io/projected/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-kube-api-access-7bzvz\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt\" (UID: \"7e4f73e8-58b4-47c3-83c6-75cc10ae1866\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.304338 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt\" (UID: \"7e4f73e8-58b4-47c3-83c6-75cc10ae1866\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.304360 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt\" (UID: \"7e4f73e8-58b4-47c3-83c6-75cc10ae1866\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.304406 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt\" (UID: \"7e4f73e8-58b4-47c3-83c6-75cc10ae1866\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.308428 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt\" (UID: \"7e4f73e8-58b4-47c3-83c6-75cc10ae1866\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.310697 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt\" (UID: \"7e4f73e8-58b4-47c3-83c6-75cc10ae1866\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.315984 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt\" (UID: \"7e4f73e8-58b4-47c3-83c6-75cc10ae1866\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.329204 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bzvz\" (UniqueName: \"kubernetes.io/projected/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-kube-api-access-7bzvz\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt\" (UID: \"7e4f73e8-58b4-47c3-83c6-75cc10ae1866\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.339472 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-tz998" event={"ID":"6d6086e9-30aa-4a56-9233-182c811c85c2","Type":"ContainerDied","Data":"ab37600bc642843d41a21ba874ea2168fd3a8e19f0fa0bf45ab2883f0381bf7e"} Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.339511 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab37600bc642843d41a21ba874ea2168fd3a8e19f0fa0bf45ab2883f0381bf7e" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.340317 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-tz998" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.404422 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.522096 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-5ff994f754-l2ctn" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.636946 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-864c6d46fc-tchw8"] Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.637370 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-864c6d46fc-tchw8" podUID="69584201-dc28-478a-a45c-ede768957725" containerName="heat-cfnapi" containerID="cri-o://97f6d0894f2afe83918ab8a0506b5257dab87c66af95e6b9a225f13af601974a" gracePeriod=60 Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.826605 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:51:40 crc kubenswrapper[4766]: E1126 00:51:40.826864 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:51:40 crc kubenswrapper[4766]: I1126 00:51:40.973763 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-7f6d8f97f4-szt84" Nov 26 00:51:41 crc kubenswrapper[4766]: I1126 00:51:41.052185 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6d74f96c45-v8sqm"] Nov 26 00:51:41 crc kubenswrapper[4766]: I1126 00:51:41.052450 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-6d74f96c45-v8sqm" podUID="d37231fe-97e9-4f01-b914-e0bd6ebfcf21" containerName="heat-api" containerID="cri-o://49046a5ad31c38fa307b5b32dc33c292711774a7a49cba0bbb076428eacad0e9" gracePeriod=60 Nov 26 00:51:41 crc kubenswrapper[4766]: I1126 00:51:41.114954 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt"] Nov 26 00:51:41 crc kubenswrapper[4766]: I1126 00:51:41.358709 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt" event={"ID":"7e4f73e8-58b4-47c3-83c6-75cc10ae1866","Type":"ContainerStarted","Data":"e0bc6b65bbd9e1cba01718db332ebfe509a5f1873c7c6c1a8d40b322c2395722"} Nov 26 00:51:43 crc kubenswrapper[4766]: I1126 00:51:43.786832 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-864c6d46fc-tchw8" podUID="69584201-dc28-478a-a45c-ede768957725" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.0.203:8000/healthcheck\": read tcp 10.217.0.2:47918->10.217.0.203:8000: read: connection reset by peer" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.203354 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.203972 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="5383cd51-7225-43df-8637-9d8216505757" containerName="aodh-api" containerID="cri-o://dbb697fcd0acc97e4a2057b818fc9792cc3251bef001e93c129eed46cab3d934" gracePeriod=30 Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.204528 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="5383cd51-7225-43df-8637-9d8216505757" containerName="aodh-listener" containerID="cri-o://f864a4e95e0b02c3766b4690228cf2871cb060db24cf9356761f03f29da7d731" gracePeriod=30 Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.204592 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="5383cd51-7225-43df-8637-9d8216505757" containerName="aodh-notifier" containerID="cri-o://3e34c59bf5bf7ca25e183e71e40495c20d81fb1424f79727c366c75530b5398e" gracePeriod=30 Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.204640 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="5383cd51-7225-43df-8637-9d8216505757" containerName="aodh-evaluator" containerID="cri-o://e70c99dab2f3e5e3dbfa4be4d4bf349e0fd90a658a331bd6182eda41ad03b53a" gracePeriod=30 Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.267154 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.274310 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-6d74f96c45-v8sqm" podUID="d37231fe-97e9-4f01-b914-e0bd6ebfcf21" containerName="heat-api" probeResult="failure" output="Get \"https://10.217.0.202:8004/healthcheck\": read tcp 10.217.0.2:36788->10.217.0.202:8004: read: connection reset by peer" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.403218 4766 generic.go:334] "Generic (PLEG): container finished" podID="d37231fe-97e9-4f01-b914-e0bd6ebfcf21" containerID="49046a5ad31c38fa307b5b32dc33c292711774a7a49cba0bbb076428eacad0e9" exitCode=0 Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.403285 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6d74f96c45-v8sqm" event={"ID":"d37231fe-97e9-4f01-b914-e0bd6ebfcf21","Type":"ContainerDied","Data":"49046a5ad31c38fa307b5b32dc33c292711774a7a49cba0bbb076428eacad0e9"} Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.404012 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-combined-ca-bundle\") pod \"69584201-dc28-478a-a45c-ede768957725\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.404091 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckrt5\" (UniqueName: \"kubernetes.io/projected/69584201-dc28-478a-a45c-ede768957725-kube-api-access-ckrt5\") pod \"69584201-dc28-478a-a45c-ede768957725\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.404164 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-public-tls-certs\") pod \"69584201-dc28-478a-a45c-ede768957725\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.404230 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-config-data-custom\") pod \"69584201-dc28-478a-a45c-ede768957725\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.404891 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-internal-tls-certs\") pod \"69584201-dc28-478a-a45c-ede768957725\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.405134 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-config-data\") pod \"69584201-dc28-478a-a45c-ede768957725\" (UID: \"69584201-dc28-478a-a45c-ede768957725\") " Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.414264 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "69584201-dc28-478a-a45c-ede768957725" (UID: "69584201-dc28-478a-a45c-ede768957725"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.414585 4766 generic.go:334] "Generic (PLEG): container finished" podID="69584201-dc28-478a-a45c-ede768957725" containerID="97f6d0894f2afe83918ab8a0506b5257dab87c66af95e6b9a225f13af601974a" exitCode=0 Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.414720 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-864c6d46fc-tchw8" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.414727 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-864c6d46fc-tchw8" event={"ID":"69584201-dc28-478a-a45c-ede768957725","Type":"ContainerDied","Data":"97f6d0894f2afe83918ab8a0506b5257dab87c66af95e6b9a225f13af601974a"} Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.414928 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-864c6d46fc-tchw8" event={"ID":"69584201-dc28-478a-a45c-ede768957725","Type":"ContainerDied","Data":"6e2942c9d25e30c668f40d5f83f7675185d231dedbbd07e031b2f9a4cf612f8c"} Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.415037 4766 scope.go:117] "RemoveContainer" containerID="97f6d0894f2afe83918ab8a0506b5257dab87c66af95e6b9a225f13af601974a" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.427603 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69584201-dc28-478a-a45c-ede768957725-kube-api-access-ckrt5" (OuterVolumeSpecName: "kube-api-access-ckrt5") pod "69584201-dc28-478a-a45c-ede768957725" (UID: "69584201-dc28-478a-a45c-ede768957725"). InnerVolumeSpecName "kube-api-access-ckrt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.462328 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69584201-dc28-478a-a45c-ede768957725" (UID: "69584201-dc28-478a-a45c-ede768957725"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.507156 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-config-data" (OuterVolumeSpecName: "config-data") pod "69584201-dc28-478a-a45c-ede768957725" (UID: "69584201-dc28-478a-a45c-ede768957725"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.510747 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.510774 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.510783 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckrt5\" (UniqueName: \"kubernetes.io/projected/69584201-dc28-478a-a45c-ede768957725-kube-api-access-ckrt5\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.510792 4766 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.523907 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "69584201-dc28-478a-a45c-ede768957725" (UID: "69584201-dc28-478a-a45c-ede768957725"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.534630 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "69584201-dc28-478a-a45c-ede768957725" (UID: "69584201-dc28-478a-a45c-ede768957725"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.545830 4766 scope.go:117] "RemoveContainer" containerID="97f6d0894f2afe83918ab8a0506b5257dab87c66af95e6b9a225f13af601974a" Nov 26 00:51:44 crc kubenswrapper[4766]: E1126 00:51:44.546171 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97f6d0894f2afe83918ab8a0506b5257dab87c66af95e6b9a225f13af601974a\": container with ID starting with 97f6d0894f2afe83918ab8a0506b5257dab87c66af95e6b9a225f13af601974a not found: ID does not exist" containerID="97f6d0894f2afe83918ab8a0506b5257dab87c66af95e6b9a225f13af601974a" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.546196 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97f6d0894f2afe83918ab8a0506b5257dab87c66af95e6b9a225f13af601974a"} err="failed to get container status \"97f6d0894f2afe83918ab8a0506b5257dab87c66af95e6b9a225f13af601974a\": rpc error: code = NotFound desc = could not find container \"97f6d0894f2afe83918ab8a0506b5257dab87c66af95e6b9a225f13af601974a\": container with ID starting with 97f6d0894f2afe83918ab8a0506b5257dab87c66af95e6b9a225f13af601974a not found: ID does not exist" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.613371 4766 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.613407 4766 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69584201-dc28-478a-a45c-ede768957725-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.705917 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.799977 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-864c6d46fc-tchw8"] Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.816320 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-config-data-custom\") pod \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.816822 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7smgn\" (UniqueName: \"kubernetes.io/projected/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-kube-api-access-7smgn\") pod \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.816910 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-config-data\") pod \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.817041 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-combined-ca-bundle\") pod \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.817078 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-public-tls-certs\") pod \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.817143 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-internal-tls-certs\") pod \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\" (UID: \"d37231fe-97e9-4f01-b914-e0bd6ebfcf21\") " Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.820273 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d37231fe-97e9-4f01-b914-e0bd6ebfcf21" (UID: "d37231fe-97e9-4f01-b914-e0bd6ebfcf21"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.823603 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-864c6d46fc-tchw8"] Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.829759 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-kube-api-access-7smgn" (OuterVolumeSpecName: "kube-api-access-7smgn") pod "d37231fe-97e9-4f01-b914-e0bd6ebfcf21" (UID: "d37231fe-97e9-4f01-b914-e0bd6ebfcf21"). InnerVolumeSpecName "kube-api-access-7smgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.908231 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d37231fe-97e9-4f01-b914-e0bd6ebfcf21" (UID: "d37231fe-97e9-4f01-b914-e0bd6ebfcf21"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.912354 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d37231fe-97e9-4f01-b914-e0bd6ebfcf21" (UID: "d37231fe-97e9-4f01-b914-e0bd6ebfcf21"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.920436 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7smgn\" (UniqueName: \"kubernetes.io/projected/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-kube-api-access-7smgn\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.920475 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.920487 4766 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.920498 4766 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.924044 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d37231fe-97e9-4f01-b914-e0bd6ebfcf21" (UID: "d37231fe-97e9-4f01-b914-e0bd6ebfcf21"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:44 crc kubenswrapper[4766]: I1126 00:51:44.934351 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-config-data" (OuterVolumeSpecName: "config-data") pod "d37231fe-97e9-4f01-b914-e0bd6ebfcf21" (UID: "d37231fe-97e9-4f01-b914-e0bd6ebfcf21"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:45 crc kubenswrapper[4766]: I1126 00:51:45.022180 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:45 crc kubenswrapper[4766]: I1126 00:51:45.022391 4766 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d37231fe-97e9-4f01-b914-e0bd6ebfcf21-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:45 crc kubenswrapper[4766]: I1126 00:51:45.426412 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6d74f96c45-v8sqm" event={"ID":"d37231fe-97e9-4f01-b914-e0bd6ebfcf21","Type":"ContainerDied","Data":"228523eac023b6c65a0a7320b9fd6ff6f49765792458a5e4d9a99994ad531fc4"} Nov 26 00:51:45 crc kubenswrapper[4766]: I1126 00:51:45.426463 4766 scope.go:117] "RemoveContainer" containerID="49046a5ad31c38fa307b5b32dc33c292711774a7a49cba0bbb076428eacad0e9" Nov 26 00:51:45 crc kubenswrapper[4766]: I1126 00:51:45.426475 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6d74f96c45-v8sqm" Nov 26 00:51:45 crc kubenswrapper[4766]: I1126 00:51:45.436118 4766 generic.go:334] "Generic (PLEG): container finished" podID="5383cd51-7225-43df-8637-9d8216505757" containerID="e70c99dab2f3e5e3dbfa4be4d4bf349e0fd90a658a331bd6182eda41ad03b53a" exitCode=0 Nov 26 00:51:45 crc kubenswrapper[4766]: I1126 00:51:45.436142 4766 generic.go:334] "Generic (PLEG): container finished" podID="5383cd51-7225-43df-8637-9d8216505757" containerID="dbb697fcd0acc97e4a2057b818fc9792cc3251bef001e93c129eed46cab3d934" exitCode=0 Nov 26 00:51:45 crc kubenswrapper[4766]: I1126 00:51:45.436155 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5383cd51-7225-43df-8637-9d8216505757","Type":"ContainerDied","Data":"e70c99dab2f3e5e3dbfa4be4d4bf349e0fd90a658a331bd6182eda41ad03b53a"} Nov 26 00:51:45 crc kubenswrapper[4766]: I1126 00:51:45.436196 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5383cd51-7225-43df-8637-9d8216505757","Type":"ContainerDied","Data":"dbb697fcd0acc97e4a2057b818fc9792cc3251bef001e93c129eed46cab3d934"} Nov 26 00:51:45 crc kubenswrapper[4766]: I1126 00:51:45.468744 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6d74f96c45-v8sqm"] Nov 26 00:51:45 crc kubenswrapper[4766]: I1126 00:51:45.488948 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-6d74f96c45-v8sqm"] Nov 26 00:51:45 crc kubenswrapper[4766]: I1126 00:51:45.839352 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69584201-dc28-478a-a45c-ede768957725" path="/var/lib/kubelet/pods/69584201-dc28-478a-a45c-ede768957725/volumes" Nov 26 00:51:45 crc kubenswrapper[4766]: I1126 00:51:45.840405 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d37231fe-97e9-4f01-b914-e0bd6ebfcf21" path="/var/lib/kubelet/pods/d37231fe-97e9-4f01-b914-e0bd6ebfcf21/volumes" Nov 26 00:51:47 crc kubenswrapper[4766]: I1126 00:51:47.462334 4766 generic.go:334] "Generic (PLEG): container finished" podID="64400ff7-9c3a-42c3-b1fc-2f9696f1206c" containerID="b6b967648cb30b946648f540cb66b304c9ea44a740abf811907c71add66c3c7f" exitCode=0 Nov 26 00:51:47 crc kubenswrapper[4766]: I1126 00:51:47.462446 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"64400ff7-9c3a-42c3-b1fc-2f9696f1206c","Type":"ContainerDied","Data":"b6b967648cb30b946648f540cb66b304c9ea44a740abf811907c71add66c3c7f"} Nov 26 00:51:47 crc kubenswrapper[4766]: I1126 00:51:47.467881 4766 generic.go:334] "Generic (PLEG): container finished" podID="5383cd51-7225-43df-8637-9d8216505757" containerID="f864a4e95e0b02c3766b4690228cf2871cb060db24cf9356761f03f29da7d731" exitCode=0 Nov 26 00:51:47 crc kubenswrapper[4766]: I1126 00:51:47.467920 4766 generic.go:334] "Generic (PLEG): container finished" podID="5383cd51-7225-43df-8637-9d8216505757" containerID="3e34c59bf5bf7ca25e183e71e40495c20d81fb1424f79727c366c75530b5398e" exitCode=0 Nov 26 00:51:47 crc kubenswrapper[4766]: I1126 00:51:47.467948 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5383cd51-7225-43df-8637-9d8216505757","Type":"ContainerDied","Data":"f864a4e95e0b02c3766b4690228cf2871cb060db24cf9356761f03f29da7d731"} Nov 26 00:51:47 crc kubenswrapper[4766]: I1126 00:51:47.467985 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5383cd51-7225-43df-8637-9d8216505757","Type":"ContainerDied","Data":"3e34c59bf5bf7ca25e183e71e40495c20d81fb1424f79727c366c75530b5398e"} Nov 26 00:51:48 crc kubenswrapper[4766]: E1126 00:51:48.294738 4766 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c5c61611bbcbcf3f57f04c4c93929a9c73ad6b0ae1d455e0599b25ed4f3f2da3" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 26 00:51:48 crc kubenswrapper[4766]: E1126 00:51:48.296928 4766 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c5c61611bbcbcf3f57f04c4c93929a9c73ad6b0ae1d455e0599b25ed4f3f2da3" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 26 00:51:48 crc kubenswrapper[4766]: E1126 00:51:48.298686 4766 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c5c61611bbcbcf3f57f04c4c93929a9c73ad6b0ae1d455e0599b25ed4f3f2da3" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 26 00:51:48 crc kubenswrapper[4766]: E1126 00:51:48.298781 4766 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-668d6b6d67-r2rrd" podUID="c0d62d37-c9d8-46f3-8fcd-802578c573fb" containerName="heat-engine" Nov 26 00:51:48 crc kubenswrapper[4766]: I1126 00:51:48.482221 4766 generic.go:334] "Generic (PLEG): container finished" podID="797e86ef-8e0d-499e-859f-6f6732826d7e" containerID="ac4e43dceb51e8d670e672114cea611a1f004f02b356c30a9b2f9b59f7088978" exitCode=0 Nov 26 00:51:48 crc kubenswrapper[4766]: I1126 00:51:48.484215 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"797e86ef-8e0d-499e-859f-6f6732826d7e","Type":"ContainerDied","Data":"ac4e43dceb51e8d670e672114cea611a1f004f02b356c30a9b2f9b59f7088978"} Nov 26 00:51:50 crc kubenswrapper[4766]: I1126 00:51:50.851806 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 26 00:51:50 crc kubenswrapper[4766]: I1126 00:51:50.995830 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-public-tls-certs\") pod \"5383cd51-7225-43df-8637-9d8216505757\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " Nov 26 00:51:50 crc kubenswrapper[4766]: I1126 00:51:50.996101 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7846\" (UniqueName: \"kubernetes.io/projected/5383cd51-7225-43df-8637-9d8216505757-kube-api-access-v7846\") pod \"5383cd51-7225-43df-8637-9d8216505757\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " Nov 26 00:51:50 crc kubenswrapper[4766]: I1126 00:51:50.996165 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-config-data\") pod \"5383cd51-7225-43df-8637-9d8216505757\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " Nov 26 00:51:50 crc kubenswrapper[4766]: I1126 00:51:50.996260 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-internal-tls-certs\") pod \"5383cd51-7225-43df-8637-9d8216505757\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " Nov 26 00:51:50 crc kubenswrapper[4766]: I1126 00:51:50.996322 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-scripts\") pod \"5383cd51-7225-43df-8637-9d8216505757\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " Nov 26 00:51:50 crc kubenswrapper[4766]: I1126 00:51:50.996390 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-combined-ca-bundle\") pod \"5383cd51-7225-43df-8637-9d8216505757\" (UID: \"5383cd51-7225-43df-8637-9d8216505757\") " Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.005797 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-scripts" (OuterVolumeSpecName: "scripts") pod "5383cd51-7225-43df-8637-9d8216505757" (UID: "5383cd51-7225-43df-8637-9d8216505757"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.005975 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5383cd51-7225-43df-8637-9d8216505757-kube-api-access-v7846" (OuterVolumeSpecName: "kube-api-access-v7846") pod "5383cd51-7225-43df-8637-9d8216505757" (UID: "5383cd51-7225-43df-8637-9d8216505757"). InnerVolumeSpecName "kube-api-access-v7846". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.070842 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5383cd51-7225-43df-8637-9d8216505757" (UID: "5383cd51-7225-43df-8637-9d8216505757"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.099297 4766 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.099332 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.099356 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7846\" (UniqueName: \"kubernetes.io/projected/5383cd51-7225-43df-8637-9d8216505757-kube-api-access-v7846\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.109798 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5383cd51-7225-43df-8637-9d8216505757" (UID: "5383cd51-7225-43df-8637-9d8216505757"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.182488 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-config-data" (OuterVolumeSpecName: "config-data") pod "5383cd51-7225-43df-8637-9d8216505757" (UID: "5383cd51-7225-43df-8637-9d8216505757"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.201803 4766 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.201835 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.214116 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5383cd51-7225-43df-8637-9d8216505757" (UID: "5383cd51-7225-43df-8637-9d8216505757"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.303542 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5383cd51-7225-43df-8637-9d8216505757-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.530344 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"64400ff7-9c3a-42c3-b1fc-2f9696f1206c","Type":"ContainerStarted","Data":"c1aae30aa702eb6fdbaebef99103bbd8ede8dc99cfff81aa416d5409d4076903"} Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.530636 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.538759 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5383cd51-7225-43df-8637-9d8216505757","Type":"ContainerDied","Data":"7954bcc8eaacd5ef1bb30a78deee221b2d59b5452715a1256b4d9292f93fe08a"} Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.538809 4766 scope.go:117] "RemoveContainer" containerID="f864a4e95e0b02c3766b4690228cf2871cb060db24cf9356761f03f29da7d731" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.538813 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.543292 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt" event={"ID":"7e4f73e8-58b4-47c3-83c6-75cc10ae1866","Type":"ContainerStarted","Data":"1b4c4650f175eb35d1916e6b05810a6a9f1a03d94cfb4e6f5312a3a56f6c4809"} Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.548059 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"797e86ef-8e0d-499e-859f-6f6732826d7e","Type":"ContainerStarted","Data":"f4a12b6ee7ad336091804289bc894c6a13745a83999dc17d80c439b15b9d8cac"} Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.548379 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.584442 4766 scope.go:117] "RemoveContainer" containerID="3e34c59bf5bf7ca25e183e71e40495c20d81fb1424f79727c366c75530b5398e" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.586361 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=42.586350064 podStartE2EDuration="42.586350064s" podCreationTimestamp="2025-11-26 00:51:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:51:51.569183112 +0000 UTC m=+1692.417953542" watchObservedRunningTime="2025-11-26 00:51:51.586350064 +0000 UTC m=+1692.435120494" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.601932 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=42.601914137 podStartE2EDuration="42.601914137s" podCreationTimestamp="2025-11-26 00:51:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 00:51:51.592191748 +0000 UTC m=+1692.440962178" watchObservedRunningTime="2025-11-26 00:51:51.601914137 +0000 UTC m=+1692.450684567" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.608790 4766 scope.go:117] "RemoveContainer" containerID="e70c99dab2f3e5e3dbfa4be4d4bf349e0fd90a658a331bd6182eda41ad03b53a" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.622700 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.634667 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.636186 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt" podStartSLOduration=2.255302795 podStartE2EDuration="11.636172959s" podCreationTimestamp="2025-11-26 00:51:40 +0000 UTC" firstStartedPulling="2025-11-26 00:51:41.11548441 +0000 UTC m=+1681.964254840" lastFinishedPulling="2025-11-26 00:51:50.496354574 +0000 UTC m=+1691.345125004" observedRunningTime="2025-11-26 00:51:51.62481323 +0000 UTC m=+1692.473583660" watchObservedRunningTime="2025-11-26 00:51:51.636172959 +0000 UTC m=+1692.484943389" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.643234 4766 scope.go:117] "RemoveContainer" containerID="dbb697fcd0acc97e4a2057b818fc9792cc3251bef001e93c129eed46cab3d934" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.658697 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 26 00:51:51 crc kubenswrapper[4766]: E1126 00:51:51.659127 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5383cd51-7225-43df-8637-9d8216505757" containerName="aodh-notifier" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.659145 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5383cd51-7225-43df-8637-9d8216505757" containerName="aodh-notifier" Nov 26 00:51:51 crc kubenswrapper[4766]: E1126 00:51:51.659159 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d37231fe-97e9-4f01-b914-e0bd6ebfcf21" containerName="heat-api" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.659166 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d37231fe-97e9-4f01-b914-e0bd6ebfcf21" containerName="heat-api" Nov 26 00:51:51 crc kubenswrapper[4766]: E1126 00:51:51.659181 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5383cd51-7225-43df-8637-9d8216505757" containerName="aodh-api" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.659187 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5383cd51-7225-43df-8637-9d8216505757" containerName="aodh-api" Nov 26 00:51:51 crc kubenswrapper[4766]: E1126 00:51:51.659198 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5383cd51-7225-43df-8637-9d8216505757" containerName="aodh-evaluator" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.659204 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5383cd51-7225-43df-8637-9d8216505757" containerName="aodh-evaluator" Nov 26 00:51:51 crc kubenswrapper[4766]: E1126 00:51:51.659221 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5383cd51-7225-43df-8637-9d8216505757" containerName="aodh-listener" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.659228 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5383cd51-7225-43df-8637-9d8216505757" containerName="aodh-listener" Nov 26 00:51:51 crc kubenswrapper[4766]: E1126 00:51:51.659244 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69584201-dc28-478a-a45c-ede768957725" containerName="heat-cfnapi" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.659251 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="69584201-dc28-478a-a45c-ede768957725" containerName="heat-cfnapi" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.659458 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="d37231fe-97e9-4f01-b914-e0bd6ebfcf21" containerName="heat-api" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.659477 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="5383cd51-7225-43df-8637-9d8216505757" containerName="aodh-evaluator" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.659492 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="5383cd51-7225-43df-8637-9d8216505757" containerName="aodh-listener" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.659504 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="5383cd51-7225-43df-8637-9d8216505757" containerName="aodh-notifier" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.659517 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="5383cd51-7225-43df-8637-9d8216505757" containerName="aodh-api" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.659537 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="69584201-dc28-478a-a45c-ede768957725" containerName="heat-cfnapi" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.661353 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.664786 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.667421 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.667728 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-9jvhv" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.667989 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.668206 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.668261 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.711505 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvlqb\" (UniqueName: \"kubernetes.io/projected/54262828-5166-42d6-934e-5eab8683ecbd-kube-api-access-pvlqb\") pod \"aodh-0\" (UID: \"54262828-5166-42d6-934e-5eab8683ecbd\") " pod="openstack/aodh-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.711561 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54262828-5166-42d6-934e-5eab8683ecbd-scripts\") pod \"aodh-0\" (UID: \"54262828-5166-42d6-934e-5eab8683ecbd\") " pod="openstack/aodh-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.711611 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/54262828-5166-42d6-934e-5eab8683ecbd-internal-tls-certs\") pod \"aodh-0\" (UID: \"54262828-5166-42d6-934e-5eab8683ecbd\") " pod="openstack/aodh-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.711637 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54262828-5166-42d6-934e-5eab8683ecbd-config-data\") pod \"aodh-0\" (UID: \"54262828-5166-42d6-934e-5eab8683ecbd\") " pod="openstack/aodh-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.711730 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54262828-5166-42d6-934e-5eab8683ecbd-combined-ca-bundle\") pod \"aodh-0\" (UID: \"54262828-5166-42d6-934e-5eab8683ecbd\") " pod="openstack/aodh-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.711847 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/54262828-5166-42d6-934e-5eab8683ecbd-public-tls-certs\") pod \"aodh-0\" (UID: \"54262828-5166-42d6-934e-5eab8683ecbd\") " pod="openstack/aodh-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.813453 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/54262828-5166-42d6-934e-5eab8683ecbd-internal-tls-certs\") pod \"aodh-0\" (UID: \"54262828-5166-42d6-934e-5eab8683ecbd\") " pod="openstack/aodh-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.813534 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54262828-5166-42d6-934e-5eab8683ecbd-config-data\") pod \"aodh-0\" (UID: \"54262828-5166-42d6-934e-5eab8683ecbd\") " pod="openstack/aodh-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.813568 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54262828-5166-42d6-934e-5eab8683ecbd-combined-ca-bundle\") pod \"aodh-0\" (UID: \"54262828-5166-42d6-934e-5eab8683ecbd\") " pod="openstack/aodh-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.813717 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/54262828-5166-42d6-934e-5eab8683ecbd-public-tls-certs\") pod \"aodh-0\" (UID: \"54262828-5166-42d6-934e-5eab8683ecbd\") " pod="openstack/aodh-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.813810 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvlqb\" (UniqueName: \"kubernetes.io/projected/54262828-5166-42d6-934e-5eab8683ecbd-kube-api-access-pvlqb\") pod \"aodh-0\" (UID: \"54262828-5166-42d6-934e-5eab8683ecbd\") " pod="openstack/aodh-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.814068 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54262828-5166-42d6-934e-5eab8683ecbd-scripts\") pod \"aodh-0\" (UID: \"54262828-5166-42d6-934e-5eab8683ecbd\") " pod="openstack/aodh-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.818161 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/54262828-5166-42d6-934e-5eab8683ecbd-public-tls-certs\") pod \"aodh-0\" (UID: \"54262828-5166-42d6-934e-5eab8683ecbd\") " pod="openstack/aodh-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.818256 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/54262828-5166-42d6-934e-5eab8683ecbd-internal-tls-certs\") pod \"aodh-0\" (UID: \"54262828-5166-42d6-934e-5eab8683ecbd\") " pod="openstack/aodh-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.820396 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54262828-5166-42d6-934e-5eab8683ecbd-scripts\") pod \"aodh-0\" (UID: \"54262828-5166-42d6-934e-5eab8683ecbd\") " pod="openstack/aodh-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.827456 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54262828-5166-42d6-934e-5eab8683ecbd-config-data\") pod \"aodh-0\" (UID: \"54262828-5166-42d6-934e-5eab8683ecbd\") " pod="openstack/aodh-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.832429 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54262828-5166-42d6-934e-5eab8683ecbd-combined-ca-bundle\") pod \"aodh-0\" (UID: \"54262828-5166-42d6-934e-5eab8683ecbd\") " pod="openstack/aodh-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.844295 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvlqb\" (UniqueName: \"kubernetes.io/projected/54262828-5166-42d6-934e-5eab8683ecbd-kube-api-access-pvlqb\") pod \"aodh-0\" (UID: \"54262828-5166-42d6-934e-5eab8683ecbd\") " pod="openstack/aodh-0" Nov 26 00:51:51 crc kubenswrapper[4766]: I1126 00:51:51.877253 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5383cd51-7225-43df-8637-9d8216505757" path="/var/lib/kubelet/pods/5383cd51-7225-43df-8637-9d8216505757/volumes" Nov 26 00:51:52 crc kubenswrapper[4766]: I1126 00:51:52.003344 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 26 00:51:52 crc kubenswrapper[4766]: W1126 00:51:52.542985 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54262828_5166_42d6_934e_5eab8683ecbd.slice/crio-0b1face941e7e0e9cb6feec7ac0b6346801d77f0938d13df60ef5cde4fbfecda WatchSource:0}: Error finding container 0b1face941e7e0e9cb6feec7ac0b6346801d77f0938d13df60ef5cde4fbfecda: Status 404 returned error can't find the container with id 0b1face941e7e0e9cb6feec7ac0b6346801d77f0938d13df60ef5cde4fbfecda Nov 26 00:51:52 crc kubenswrapper[4766]: I1126 00:51:52.544191 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 26 00:51:52 crc kubenswrapper[4766]: I1126 00:51:52.562764 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"54262828-5166-42d6-934e-5eab8683ecbd","Type":"ContainerStarted","Data":"0b1face941e7e0e9cb6feec7ac0b6346801d77f0938d13df60ef5cde4fbfecda"} Nov 26 00:51:53 crc kubenswrapper[4766]: I1126 00:51:53.575110 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"54262828-5166-42d6-934e-5eab8683ecbd","Type":"ContainerStarted","Data":"1dbc330678a302dbed6057f4425bd549acc9379dd9bee57c791d45de47767b5c"} Nov 26 00:51:54 crc kubenswrapper[4766]: I1126 00:51:54.591326 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"54262828-5166-42d6-934e-5eab8683ecbd","Type":"ContainerStarted","Data":"d38db06850616577b1715efe57d4eda33aabad9eda9cab77e9b43ec5d076d34a"} Nov 26 00:51:54 crc kubenswrapper[4766]: I1126 00:51:54.827375 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:51:54 crc kubenswrapper[4766]: E1126 00:51:54.827578 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:51:55 crc kubenswrapper[4766]: I1126 00:51:55.187611 4766 scope.go:117] "RemoveContainer" containerID="dd0a910ba1ed6f7f52274d8aba3d44c06ccdc43b8bfc50f731aad0b2322b2405" Nov 26 00:51:55 crc kubenswrapper[4766]: I1126 00:51:55.300760 4766 scope.go:117] "RemoveContainer" containerID="7f8c888a919d7a6d73f8981f0fc18c2ff35464b5c11779cedf8e8fa696efdfd9" Nov 26 00:51:55 crc kubenswrapper[4766]: I1126 00:51:55.609551 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"54262828-5166-42d6-934e-5eab8683ecbd","Type":"ContainerStarted","Data":"df8dccf890e4f6550d534ddf2393121cb642d998968c58b7ff3483a26466f919"} Nov 26 00:51:57 crc kubenswrapper[4766]: I1126 00:51:57.636506 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"54262828-5166-42d6-934e-5eab8683ecbd","Type":"ContainerStarted","Data":"7789182519dd3e715ee8bc5821cb3a72b8f20d82afc0c32c91c66c073268b976"} Nov 26 00:51:57 crc kubenswrapper[4766]: I1126 00:51:57.679835 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.8654750399999998 podStartE2EDuration="6.679807242s" podCreationTimestamp="2025-11-26 00:51:51 +0000 UTC" firstStartedPulling="2025-11-26 00:51:52.546247455 +0000 UTC m=+1693.395017885" lastFinishedPulling="2025-11-26 00:51:56.360579657 +0000 UTC m=+1697.209350087" observedRunningTime="2025-11-26 00:51:57.669709953 +0000 UTC m=+1698.518480443" watchObservedRunningTime="2025-11-26 00:51:57.679807242 +0000 UTC m=+1698.528577712" Nov 26 00:51:58 crc kubenswrapper[4766]: E1126 00:51:58.279801 4766 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c5c61611bbcbcf3f57f04c4c93929a9c73ad6b0ae1d455e0599b25ed4f3f2da3" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 26 00:51:58 crc kubenswrapper[4766]: E1126 00:51:58.281219 4766 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c5c61611bbcbcf3f57f04c4c93929a9c73ad6b0ae1d455e0599b25ed4f3f2da3" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 26 00:51:58 crc kubenswrapper[4766]: E1126 00:51:58.282575 4766 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c5c61611bbcbcf3f57f04c4c93929a9c73ad6b0ae1d455e0599b25ed4f3f2da3" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 26 00:51:58 crc kubenswrapper[4766]: E1126 00:51:58.282609 4766 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-668d6b6d67-r2rrd" podUID="c0d62d37-c9d8-46f3-8fcd-802578c573fb" containerName="heat-engine" Nov 26 00:52:00 crc kubenswrapper[4766]: I1126 00:52:00.354576 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 26 00:52:00 crc kubenswrapper[4766]: I1126 00:52:00.457959 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 26 00:52:00 crc kubenswrapper[4766]: I1126 00:52:00.671035 4766 generic.go:334] "Generic (PLEG): container finished" podID="c0d62d37-c9d8-46f3-8fcd-802578c573fb" containerID="c5c61611bbcbcf3f57f04c4c93929a9c73ad6b0ae1d455e0599b25ed4f3f2da3" exitCode=0 Nov 26 00:52:00 crc kubenswrapper[4766]: I1126 00:52:00.671437 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-668d6b6d67-r2rrd" event={"ID":"c0d62d37-c9d8-46f3-8fcd-802578c573fb","Type":"ContainerDied","Data":"c5c61611bbcbcf3f57f04c4c93929a9c73ad6b0ae1d455e0599b25ed4f3f2da3"} Nov 26 00:52:01 crc kubenswrapper[4766]: I1126 00:52:01.045370 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-668d6b6d67-r2rrd" Nov 26 00:52:01 crc kubenswrapper[4766]: I1126 00:52:01.212105 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0d62d37-c9d8-46f3-8fcd-802578c573fb-config-data-custom\") pod \"c0d62d37-c9d8-46f3-8fcd-802578c573fb\" (UID: \"c0d62d37-c9d8-46f3-8fcd-802578c573fb\") " Nov 26 00:52:01 crc kubenswrapper[4766]: I1126 00:52:01.212291 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njkxg\" (UniqueName: \"kubernetes.io/projected/c0d62d37-c9d8-46f3-8fcd-802578c573fb-kube-api-access-njkxg\") pod \"c0d62d37-c9d8-46f3-8fcd-802578c573fb\" (UID: \"c0d62d37-c9d8-46f3-8fcd-802578c573fb\") " Nov 26 00:52:01 crc kubenswrapper[4766]: I1126 00:52:01.212328 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0d62d37-c9d8-46f3-8fcd-802578c573fb-combined-ca-bundle\") pod \"c0d62d37-c9d8-46f3-8fcd-802578c573fb\" (UID: \"c0d62d37-c9d8-46f3-8fcd-802578c573fb\") " Nov 26 00:52:01 crc kubenswrapper[4766]: I1126 00:52:01.212416 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0d62d37-c9d8-46f3-8fcd-802578c573fb-config-data\") pod \"c0d62d37-c9d8-46f3-8fcd-802578c573fb\" (UID: \"c0d62d37-c9d8-46f3-8fcd-802578c573fb\") " Nov 26 00:52:01 crc kubenswrapper[4766]: I1126 00:52:01.229870 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0d62d37-c9d8-46f3-8fcd-802578c573fb-kube-api-access-njkxg" (OuterVolumeSpecName: "kube-api-access-njkxg") pod "c0d62d37-c9d8-46f3-8fcd-802578c573fb" (UID: "c0d62d37-c9d8-46f3-8fcd-802578c573fb"). InnerVolumeSpecName "kube-api-access-njkxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:52:01 crc kubenswrapper[4766]: I1126 00:52:01.237225 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0d62d37-c9d8-46f3-8fcd-802578c573fb-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c0d62d37-c9d8-46f3-8fcd-802578c573fb" (UID: "c0d62d37-c9d8-46f3-8fcd-802578c573fb"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:52:01 crc kubenswrapper[4766]: I1126 00:52:01.255496 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0d62d37-c9d8-46f3-8fcd-802578c573fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c0d62d37-c9d8-46f3-8fcd-802578c573fb" (UID: "c0d62d37-c9d8-46f3-8fcd-802578c573fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:52:01 crc kubenswrapper[4766]: I1126 00:52:01.272585 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0d62d37-c9d8-46f3-8fcd-802578c573fb-config-data" (OuterVolumeSpecName: "config-data") pod "c0d62d37-c9d8-46f3-8fcd-802578c573fb" (UID: "c0d62d37-c9d8-46f3-8fcd-802578c573fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:52:01 crc kubenswrapper[4766]: I1126 00:52:01.315746 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njkxg\" (UniqueName: \"kubernetes.io/projected/c0d62d37-c9d8-46f3-8fcd-802578c573fb-kube-api-access-njkxg\") on node \"crc\" DevicePath \"\"" Nov 26 00:52:01 crc kubenswrapper[4766]: I1126 00:52:01.315786 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0d62d37-c9d8-46f3-8fcd-802578c573fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:52:01 crc kubenswrapper[4766]: I1126 00:52:01.315800 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0d62d37-c9d8-46f3-8fcd-802578c573fb-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 00:52:01 crc kubenswrapper[4766]: I1126 00:52:01.315811 4766 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0d62d37-c9d8-46f3-8fcd-802578c573fb-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 26 00:52:01 crc kubenswrapper[4766]: I1126 00:52:01.685252 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-668d6b6d67-r2rrd" event={"ID":"c0d62d37-c9d8-46f3-8fcd-802578c573fb","Type":"ContainerDied","Data":"5d26fcc7a8368e29a1bb2c4edb3e42a99f9d7886bff196cd60e5372b1ead65b0"} Nov 26 00:52:01 crc kubenswrapper[4766]: I1126 00:52:01.685304 4766 scope.go:117] "RemoveContainer" containerID="c5c61611bbcbcf3f57f04c4c93929a9c73ad6b0ae1d455e0599b25ed4f3f2da3" Nov 26 00:52:01 crc kubenswrapper[4766]: I1126 00:52:01.685353 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-668d6b6d67-r2rrd" Nov 26 00:52:01 crc kubenswrapper[4766]: I1126 00:52:01.725220 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-668d6b6d67-r2rrd"] Nov 26 00:52:01 crc kubenswrapper[4766]: I1126 00:52:01.738461 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-668d6b6d67-r2rrd"] Nov 26 00:52:01 crc kubenswrapper[4766]: I1126 00:52:01.840562 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0d62d37-c9d8-46f3-8fcd-802578c573fb" path="/var/lib/kubelet/pods/c0d62d37-c9d8-46f3-8fcd-802578c573fb/volumes" Nov 26 00:52:03 crc kubenswrapper[4766]: I1126 00:52:03.743480 4766 generic.go:334] "Generic (PLEG): container finished" podID="7e4f73e8-58b4-47c3-83c6-75cc10ae1866" containerID="1b4c4650f175eb35d1916e6b05810a6a9f1a03d94cfb4e6f5312a3a56f6c4809" exitCode=0 Nov 26 00:52:03 crc kubenswrapper[4766]: I1126 00:52:03.743595 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt" event={"ID":"7e4f73e8-58b4-47c3-83c6-75cc10ae1866","Type":"ContainerDied","Data":"1b4c4650f175eb35d1916e6b05810a6a9f1a03d94cfb4e6f5312a3a56f6c4809"} Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.302110 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt" Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.417071 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-ssh-key\") pod \"7e4f73e8-58b4-47c3-83c6-75cc10ae1866\" (UID: \"7e4f73e8-58b4-47c3-83c6-75cc10ae1866\") " Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.417157 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-inventory\") pod \"7e4f73e8-58b4-47c3-83c6-75cc10ae1866\" (UID: \"7e4f73e8-58b4-47c3-83c6-75cc10ae1866\") " Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.417187 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-repo-setup-combined-ca-bundle\") pod \"7e4f73e8-58b4-47c3-83c6-75cc10ae1866\" (UID: \"7e4f73e8-58b4-47c3-83c6-75cc10ae1866\") " Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.417345 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bzvz\" (UniqueName: \"kubernetes.io/projected/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-kube-api-access-7bzvz\") pod \"7e4f73e8-58b4-47c3-83c6-75cc10ae1866\" (UID: \"7e4f73e8-58b4-47c3-83c6-75cc10ae1866\") " Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.426068 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-kube-api-access-7bzvz" (OuterVolumeSpecName: "kube-api-access-7bzvz") pod "7e4f73e8-58b4-47c3-83c6-75cc10ae1866" (UID: "7e4f73e8-58b4-47c3-83c6-75cc10ae1866"). InnerVolumeSpecName "kube-api-access-7bzvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.429493 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "7e4f73e8-58b4-47c3-83c6-75cc10ae1866" (UID: "7e4f73e8-58b4-47c3-83c6-75cc10ae1866"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.462843 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-inventory" (OuterVolumeSpecName: "inventory") pod "7e4f73e8-58b4-47c3-83c6-75cc10ae1866" (UID: "7e4f73e8-58b4-47c3-83c6-75cc10ae1866"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.465425 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7e4f73e8-58b4-47c3-83c6-75cc10ae1866" (UID: "7e4f73e8-58b4-47c3-83c6-75cc10ae1866"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.520356 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.520705 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.520721 4766 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.520737 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bzvz\" (UniqueName: \"kubernetes.io/projected/7e4f73e8-58b4-47c3-83c6-75cc10ae1866-kube-api-access-7bzvz\") on node \"crc\" DevicePath \"\"" Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.769719 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt" event={"ID":"7e4f73e8-58b4-47c3-83c6-75cc10ae1866","Type":"ContainerDied","Data":"e0bc6b65bbd9e1cba01718db332ebfe509a5f1873c7c6c1a8d40b322c2395722"} Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.769777 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0bc6b65bbd9e1cba01718db332ebfe509a5f1873c7c6c1a8d40b322c2395722" Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.769795 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt" Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.869720 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8"] Nov 26 00:52:05 crc kubenswrapper[4766]: E1126 00:52:05.870303 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0d62d37-c9d8-46f3-8fcd-802578c573fb" containerName="heat-engine" Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.870319 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0d62d37-c9d8-46f3-8fcd-802578c573fb" containerName="heat-engine" Nov 26 00:52:05 crc kubenswrapper[4766]: E1126 00:52:05.870356 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e4f73e8-58b4-47c3-83c6-75cc10ae1866" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.870365 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e4f73e8-58b4-47c3-83c6-75cc10ae1866" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.870684 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e4f73e8-58b4-47c3-83c6-75cc10ae1866" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.870738 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0d62d37-c9d8-46f3-8fcd-802578c573fb" containerName="heat-engine" Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.871667 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8" Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.875058 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.875353 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.875645 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.875936 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 00:52:05 crc kubenswrapper[4766]: I1126 00:52:05.885325 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8"] Nov 26 00:52:06 crc kubenswrapper[4766]: I1126 00:52:06.031298 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fee6935-5d49-4559-93de-8b244b5a6261-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8\" (UID: \"7fee6935-5d49-4559-93de-8b244b5a6261\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8" Nov 26 00:52:06 crc kubenswrapper[4766]: I1126 00:52:06.031392 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fee6935-5d49-4559-93de-8b244b5a6261-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8\" (UID: \"7fee6935-5d49-4559-93de-8b244b5a6261\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8" Nov 26 00:52:06 crc kubenswrapper[4766]: I1126 00:52:06.031461 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqzth\" (UniqueName: \"kubernetes.io/projected/7fee6935-5d49-4559-93de-8b244b5a6261-kube-api-access-sqzth\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8\" (UID: \"7fee6935-5d49-4559-93de-8b244b5a6261\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8" Nov 26 00:52:06 crc kubenswrapper[4766]: I1126 00:52:06.031560 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fee6935-5d49-4559-93de-8b244b5a6261-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8\" (UID: \"7fee6935-5d49-4559-93de-8b244b5a6261\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8" Nov 26 00:52:06 crc kubenswrapper[4766]: I1126 00:52:06.133775 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fee6935-5d49-4559-93de-8b244b5a6261-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8\" (UID: \"7fee6935-5d49-4559-93de-8b244b5a6261\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8" Nov 26 00:52:06 crc kubenswrapper[4766]: I1126 00:52:06.134060 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fee6935-5d49-4559-93de-8b244b5a6261-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8\" (UID: \"7fee6935-5d49-4559-93de-8b244b5a6261\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8" Nov 26 00:52:06 crc kubenswrapper[4766]: I1126 00:52:06.134149 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fee6935-5d49-4559-93de-8b244b5a6261-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8\" (UID: \"7fee6935-5d49-4559-93de-8b244b5a6261\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8" Nov 26 00:52:06 crc kubenswrapper[4766]: I1126 00:52:06.134202 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqzth\" (UniqueName: \"kubernetes.io/projected/7fee6935-5d49-4559-93de-8b244b5a6261-kube-api-access-sqzth\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8\" (UID: \"7fee6935-5d49-4559-93de-8b244b5a6261\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8" Nov 26 00:52:06 crc kubenswrapper[4766]: I1126 00:52:06.139354 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fee6935-5d49-4559-93de-8b244b5a6261-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8\" (UID: \"7fee6935-5d49-4559-93de-8b244b5a6261\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8" Nov 26 00:52:06 crc kubenswrapper[4766]: I1126 00:52:06.139363 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fee6935-5d49-4559-93de-8b244b5a6261-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8\" (UID: \"7fee6935-5d49-4559-93de-8b244b5a6261\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8" Nov 26 00:52:06 crc kubenswrapper[4766]: I1126 00:52:06.157403 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fee6935-5d49-4559-93de-8b244b5a6261-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8\" (UID: \"7fee6935-5d49-4559-93de-8b244b5a6261\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8" Nov 26 00:52:06 crc kubenswrapper[4766]: I1126 00:52:06.163143 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqzth\" (UniqueName: \"kubernetes.io/projected/7fee6935-5d49-4559-93de-8b244b5a6261-kube-api-access-sqzth\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8\" (UID: \"7fee6935-5d49-4559-93de-8b244b5a6261\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8" Nov 26 00:52:06 crc kubenswrapper[4766]: I1126 00:52:06.191409 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8" Nov 26 00:52:06 crc kubenswrapper[4766]: W1126 00:52:06.798023 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fee6935_5d49_4559_93de_8b244b5a6261.slice/crio-5987a53ae26c7c4be98c1c4f9a4e59c077745c424d596f036cd3f0bfe922f660 WatchSource:0}: Error finding container 5987a53ae26c7c4be98c1c4f9a4e59c077745c424d596f036cd3f0bfe922f660: Status 404 returned error can't find the container with id 5987a53ae26c7c4be98c1c4f9a4e59c077745c424d596f036cd3f0bfe922f660 Nov 26 00:52:06 crc kubenswrapper[4766]: I1126 00:52:06.806538 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8"] Nov 26 00:52:07 crc kubenswrapper[4766]: I1126 00:52:07.798710 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8" event={"ID":"7fee6935-5d49-4559-93de-8b244b5a6261","Type":"ContainerStarted","Data":"aba7c1527460d45400164d3ba4df993725b996e9df22aebb71e574413d815763"} Nov 26 00:52:07 crc kubenswrapper[4766]: I1126 00:52:07.799028 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8" event={"ID":"7fee6935-5d49-4559-93de-8b244b5a6261","Type":"ContainerStarted","Data":"5987a53ae26c7c4be98c1c4f9a4e59c077745c424d596f036cd3f0bfe922f660"} Nov 26 00:52:07 crc kubenswrapper[4766]: I1126 00:52:07.833315 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8" podStartSLOduration=2.174154388 podStartE2EDuration="2.833295795s" podCreationTimestamp="2025-11-26 00:52:05 +0000 UTC" firstStartedPulling="2025-11-26 00:52:06.803125706 +0000 UTC m=+1707.651896146" lastFinishedPulling="2025-11-26 00:52:07.462267123 +0000 UTC m=+1708.311037553" observedRunningTime="2025-11-26 00:52:07.819720981 +0000 UTC m=+1708.668491411" watchObservedRunningTime="2025-11-26 00:52:07.833295795 +0000 UTC m=+1708.682066235" Nov 26 00:52:08 crc kubenswrapper[4766]: I1126 00:52:08.829345 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:52:08 crc kubenswrapper[4766]: E1126 00:52:08.832044 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:52:21 crc kubenswrapper[4766]: I1126 00:52:21.826446 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:52:21 crc kubenswrapper[4766]: E1126 00:52:21.827114 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:52:33 crc kubenswrapper[4766]: I1126 00:52:33.828760 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:52:33 crc kubenswrapper[4766]: E1126 00:52:33.830017 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:52:44 crc kubenswrapper[4766]: I1126 00:52:44.828176 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:52:44 crc kubenswrapper[4766]: E1126 00:52:44.829186 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:52:55 crc kubenswrapper[4766]: I1126 00:52:55.603089 4766 scope.go:117] "RemoveContainer" containerID="9242e3ede9b06520f5b153cd3544ca4062ce8e5818ecaf3d453337c2fddf0419" Nov 26 00:52:59 crc kubenswrapper[4766]: I1126 00:52:59.844504 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:52:59 crc kubenswrapper[4766]: E1126 00:52:59.846067 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:53:12 crc kubenswrapper[4766]: I1126 00:53:12.829040 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:53:12 crc kubenswrapper[4766]: E1126 00:53:12.830291 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:53:24 crc kubenswrapper[4766]: I1126 00:53:24.827455 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:53:24 crc kubenswrapper[4766]: E1126 00:53:24.828284 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:53:35 crc kubenswrapper[4766]: I1126 00:53:35.826486 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:53:35 crc kubenswrapper[4766]: E1126 00:53:35.827156 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:53:46 crc kubenswrapper[4766]: I1126 00:53:46.826459 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:53:46 crc kubenswrapper[4766]: E1126 00:53:46.827120 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:53:55 crc kubenswrapper[4766]: I1126 00:53:55.716858 4766 scope.go:117] "RemoveContainer" containerID="bceafd245f3339232ecc96719a07509eae7eaa5d91b5564c4c361ab21fbea67c" Nov 26 00:53:55 crc kubenswrapper[4766]: I1126 00:53:55.752393 4766 scope.go:117] "RemoveContainer" containerID="1d21e1aedc1c9ddc220987318fdec667060660941c075ff804b51171730506de" Nov 26 00:53:55 crc kubenswrapper[4766]: I1126 00:53:55.782703 4766 scope.go:117] "RemoveContainer" containerID="27937a6f6f497f5f7c3f3a89a54f4857db004c27ee4f74df829476bd1a716a89" Nov 26 00:53:55 crc kubenswrapper[4766]: I1126 00:53:55.842437 4766 scope.go:117] "RemoveContainer" containerID="9237bc93c46f5b46707c562126f17465bf9c732a639be92b586142a8d543f870" Nov 26 00:53:57 crc kubenswrapper[4766]: I1126 00:53:57.827348 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:53:57 crc kubenswrapper[4766]: E1126 00:53:57.828205 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:54:09 crc kubenswrapper[4766]: I1126 00:54:09.837908 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:54:09 crc kubenswrapper[4766]: E1126 00:54:09.839052 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 00:54:16 crc kubenswrapper[4766]: I1126 00:54:16.081509 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-zn96g"] Nov 26 00:54:16 crc kubenswrapper[4766]: I1126 00:54:16.092987 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-b1a9-account-create-update-tzrh4"] Nov 26 00:54:16 crc kubenswrapper[4766]: I1126 00:54:16.104176 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-zn96g"] Nov 26 00:54:16 crc kubenswrapper[4766]: I1126 00:54:16.114081 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-b1a9-account-create-update-tzrh4"] Nov 26 00:54:16 crc kubenswrapper[4766]: I1126 00:54:16.123887 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-xx494"] Nov 26 00:54:16 crc kubenswrapper[4766]: I1126 00:54:16.132643 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-fbd2-account-create-update-pcmbx"] Nov 26 00:54:16 crc kubenswrapper[4766]: I1126 00:54:16.141086 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-fbd2-account-create-update-pcmbx"] Nov 26 00:54:16 crc kubenswrapper[4766]: I1126 00:54:16.177377 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-xx494"] Nov 26 00:54:17 crc kubenswrapper[4766]: I1126 00:54:17.852193 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="095a8e06-fc50-4b1c-b27d-9efb621c523a" path="/var/lib/kubelet/pods/095a8e06-fc50-4b1c-b27d-9efb621c523a/volumes" Nov 26 00:54:17 crc kubenswrapper[4766]: I1126 00:54:17.854035 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34ebe8c7-f89d-478a-ba71-4160910955c6" path="/var/lib/kubelet/pods/34ebe8c7-f89d-478a-ba71-4160910955c6/volumes" Nov 26 00:54:17 crc kubenswrapper[4766]: I1126 00:54:17.857036 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcf23dba-a1e2-41be-b115-777bda2d8dc4" path="/var/lib/kubelet/pods/bcf23dba-a1e2-41be-b115-777bda2d8dc4/volumes" Nov 26 00:54:17 crc kubenswrapper[4766]: I1126 00:54:17.858354 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2ec3192-d2ac-4f92-abd2-4c20548dceb3" path="/var/lib/kubelet/pods/c2ec3192-d2ac-4f92-abd2-4c20548dceb3/volumes" Nov 26 00:54:19 crc kubenswrapper[4766]: I1126 00:54:19.036044 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-92677"] Nov 26 00:54:19 crc kubenswrapper[4766]: I1126 00:54:19.048355 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-29559"] Nov 26 00:54:19 crc kubenswrapper[4766]: I1126 00:54:19.058351 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-92677"] Nov 26 00:54:19 crc kubenswrapper[4766]: I1126 00:54:19.070245 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-29559"] Nov 26 00:54:19 crc kubenswrapper[4766]: I1126 00:54:19.856098 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68b0745b-3ae1-420b-ba0f-3cdbec6c0878" path="/var/lib/kubelet/pods/68b0745b-3ae1-420b-ba0f-3cdbec6c0878/volumes" Nov 26 00:54:19 crc kubenswrapper[4766]: I1126 00:54:19.856907 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78944618-3a02-482a-9965-ef46c0ece134" path="/var/lib/kubelet/pods/78944618-3a02-482a-9965-ef46c0ece134/volumes" Nov 26 00:54:20 crc kubenswrapper[4766]: I1126 00:54:20.028420 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-efe0-account-create-update-2kxdm"] Nov 26 00:54:20 crc kubenswrapper[4766]: I1126 00:54:20.042216 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-3f15-account-create-update-5l657"] Nov 26 00:54:20 crc kubenswrapper[4766]: I1126 00:54:20.117703 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-efe0-account-create-update-2kxdm"] Nov 26 00:54:20 crc kubenswrapper[4766]: I1126 00:54:20.128300 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-3f15-account-create-update-5l657"] Nov 26 00:54:21 crc kubenswrapper[4766]: I1126 00:54:21.852285 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7a0a3f7-ac14-4270-a51e-35738403f130" path="/var/lib/kubelet/pods/f7a0a3f7-ac14-4270-a51e-35738403f130/volumes" Nov 26 00:54:21 crc kubenswrapper[4766]: I1126 00:54:21.854289 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e" path="/var/lib/kubelet/pods/feb7ecda-6e5c-4a6e-baf8-fe4c87f50d4e/volumes" Nov 26 00:54:24 crc kubenswrapper[4766]: I1126 00:54:24.827147 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:54:25 crc kubenswrapper[4766]: I1126 00:54:25.766182 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"31c5fe6cf4a7b0b0322cd91369e8b92e06d2251c1e99313b96d5eb55cc36b4e6"} Nov 26 00:54:30 crc kubenswrapper[4766]: I1126 00:54:30.053300 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-v2bnt"] Nov 26 00:54:30 crc kubenswrapper[4766]: I1126 00:54:30.073212 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-def0-account-create-update-rmc7d"] Nov 26 00:54:30 crc kubenswrapper[4766]: I1126 00:54:30.089195 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-def0-account-create-update-rmc7d"] Nov 26 00:54:30 crc kubenswrapper[4766]: I1126 00:54:30.103369 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-v2bnt"] Nov 26 00:54:31 crc kubenswrapper[4766]: I1126 00:54:31.851047 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="618ca9d6-4ffb-4c75-9b16-b471471f91c3" path="/var/lib/kubelet/pods/618ca9d6-4ffb-4c75-9b16-b471471f91c3/volumes" Nov 26 00:54:31 crc kubenswrapper[4766]: I1126 00:54:31.855454 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0" path="/var/lib/kubelet/pods/c9bc49f8-8ab0-4f1a-b32c-677ab40cc3f0/volumes" Nov 26 00:54:50 crc kubenswrapper[4766]: I1126 00:54:50.050768 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-7hkqm"] Nov 26 00:54:50 crc kubenswrapper[4766]: I1126 00:54:50.065847 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-7hkqm"] Nov 26 00:54:50 crc kubenswrapper[4766]: I1126 00:54:50.079826 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-6008-account-create-update-dbw86"] Nov 26 00:54:50 crc kubenswrapper[4766]: I1126 00:54:50.093297 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-6008-account-create-update-dbw86"] Nov 26 00:54:51 crc kubenswrapper[4766]: I1126 00:54:51.840111 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b80a95b-04b4-4b49-a799-cb6469c7b198" path="/var/lib/kubelet/pods/0b80a95b-04b4-4b49-a799-cb6469c7b198/volumes" Nov 26 00:54:51 crc kubenswrapper[4766]: I1126 00:54:51.841365 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec" path="/var/lib/kubelet/pods/ef5c1bcc-7d69-4ec1-848b-244c7bc4b3ec/volumes" Nov 26 00:54:52 crc kubenswrapper[4766]: I1126 00:54:52.043857 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-79smp"] Nov 26 00:54:52 crc kubenswrapper[4766]: I1126 00:54:52.057517 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-79smp"] Nov 26 00:54:52 crc kubenswrapper[4766]: I1126 00:54:52.074240 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-0831-account-create-update-l6rfv"] Nov 26 00:54:52 crc kubenswrapper[4766]: I1126 00:54:52.086739 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-9e39-account-create-update-9s4pv"] Nov 26 00:54:52 crc kubenswrapper[4766]: I1126 00:54:52.096561 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-zfpqn"] Nov 26 00:54:52 crc kubenswrapper[4766]: I1126 00:54:52.104974 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-zfpqn"] Nov 26 00:54:52 crc kubenswrapper[4766]: I1126 00:54:52.117203 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-9e39-account-create-update-9s4pv"] Nov 26 00:54:52 crc kubenswrapper[4766]: I1126 00:54:52.141980 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-0831-account-create-update-l6rfv"] Nov 26 00:54:52 crc kubenswrapper[4766]: I1126 00:54:52.154932 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-6wtrt"] Nov 26 00:54:52 crc kubenswrapper[4766]: I1126 00:54:52.165839 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-6wtrt"] Nov 26 00:54:53 crc kubenswrapper[4766]: I1126 00:54:53.856139 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4873ca17-8a36-4c66-b86b-5a6784c7e4f2" path="/var/lib/kubelet/pods/4873ca17-8a36-4c66-b86b-5a6784c7e4f2/volumes" Nov 26 00:54:53 crc kubenswrapper[4766]: I1126 00:54:53.858126 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5abdded9-8645-49e8-910e-3ce04650f19e" path="/var/lib/kubelet/pods/5abdded9-8645-49e8-910e-3ce04650f19e/volumes" Nov 26 00:54:53 crc kubenswrapper[4766]: I1126 00:54:53.859575 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92fd9e2d-9152-46a4-8686-ef9fe6d0b790" path="/var/lib/kubelet/pods/92fd9e2d-9152-46a4-8686-ef9fe6d0b790/volumes" Nov 26 00:54:53 crc kubenswrapper[4766]: I1126 00:54:53.861341 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ced743c6-6d0e-439e-9d16-46ed28450a25" path="/var/lib/kubelet/pods/ced743c6-6d0e-439e-9d16-46ed28450a25/volumes" Nov 26 00:54:53 crc kubenswrapper[4766]: I1126 00:54:53.864164 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e25f6c19-7b36-4144-8416-7df74580e906" path="/var/lib/kubelet/pods/e25f6c19-7b36-4144-8416-7df74580e906/volumes" Nov 26 00:54:55 crc kubenswrapper[4766]: I1126 00:54:55.938710 4766 scope.go:117] "RemoveContainer" containerID="2582ba9afa41abc0ae5d3e47291ea56d3d29f9cc2862d627fa98cbf0f68f93b9" Nov 26 00:54:55 crc kubenswrapper[4766]: I1126 00:54:55.971613 4766 scope.go:117] "RemoveContainer" containerID="d379cb5e72f4f66844ab6f351a9c5cdec0bbee9e536777d4b418b62a5c1a2b1b" Nov 26 00:54:56 crc kubenswrapper[4766]: I1126 00:54:56.053386 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-5dd7-account-create-update-q2fxk"] Nov 26 00:54:56 crc kubenswrapper[4766]: I1126 00:54:56.067900 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-5dd7-account-create-update-q2fxk"] Nov 26 00:54:56 crc kubenswrapper[4766]: I1126 00:54:56.069502 4766 scope.go:117] "RemoveContainer" containerID="0fc5bc8ceed46933f386b560ac88b6d46950acf7e3fca8e34371c897c1589ba6" Nov 26 00:54:56 crc kubenswrapper[4766]: I1126 00:54:56.080403 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-vh2cq"] Nov 26 00:54:56 crc kubenswrapper[4766]: I1126 00:54:56.091797 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-vh2cq"] Nov 26 00:54:56 crc kubenswrapper[4766]: I1126 00:54:56.108310 4766 scope.go:117] "RemoveContainer" containerID="355ab80164cf5594f074f0ee22d5d68a441801af3e96264aca5d72a76f929c68" Nov 26 00:54:56 crc kubenswrapper[4766]: I1126 00:54:56.148502 4766 scope.go:117] "RemoveContainer" containerID="e2cd4853efb6adc3594e8c0b2ed32609691e4a0a453aeb7a20751172fce520dc" Nov 26 00:54:56 crc kubenswrapper[4766]: I1126 00:54:56.196676 4766 scope.go:117] "RemoveContainer" containerID="4328dcf41f178bd6b54b42796098e40895df9e94b9bcebf8994c433d89780586" Nov 26 00:54:56 crc kubenswrapper[4766]: I1126 00:54:56.251177 4766 scope.go:117] "RemoveContainer" containerID="f57bf5ff37c0b77f28e0e5beab34da08d33c0e96573d481165a5a40dbb3ffadd" Nov 26 00:54:56 crc kubenswrapper[4766]: I1126 00:54:56.281036 4766 scope.go:117] "RemoveContainer" containerID="192c1c9406b9ae618d7e353c7cf5e6b63badfda270f1bdfb72db674c960f3664" Nov 26 00:54:56 crc kubenswrapper[4766]: I1126 00:54:56.305818 4766 scope.go:117] "RemoveContainer" containerID="8cdd22a5a27b876b5463eb0cadb66912b4a961a13e341b9c2f852800c9abf62c" Nov 26 00:54:56 crc kubenswrapper[4766]: I1126 00:54:56.353085 4766 scope.go:117] "RemoveContainer" containerID="3dec080f0849364d716930c5377aabf98641dbe93cb0acf2b18ebfbac25f7daf" Nov 26 00:54:56 crc kubenswrapper[4766]: I1126 00:54:56.374006 4766 scope.go:117] "RemoveContainer" containerID="88bf9d7a8f06ab171a2e337b223f19130ce74feffb2baaa08750f237903784fe" Nov 26 00:54:56 crc kubenswrapper[4766]: I1126 00:54:56.392784 4766 scope.go:117] "RemoveContainer" containerID="4e715311f76e353b420abed0419f3ac40cd086666dfd92cac3e058990bcc330b" Nov 26 00:54:56 crc kubenswrapper[4766]: I1126 00:54:56.430886 4766 scope.go:117] "RemoveContainer" containerID="6a94d4eb134b3e7e31ef1a063827a330051743ba95d2b1a225613f8053ac406d" Nov 26 00:54:56 crc kubenswrapper[4766]: I1126 00:54:56.458328 4766 scope.go:117] "RemoveContainer" containerID="dcc1454763f21e13b343f565f06d9546ef5926d64aa121466f653116ecada477" Nov 26 00:54:56 crc kubenswrapper[4766]: I1126 00:54:56.492094 4766 scope.go:117] "RemoveContainer" containerID="ffe5fab3d7cf671c80721bf7746656bd38ec42271a90b576b96afcc749301564" Nov 26 00:54:56 crc kubenswrapper[4766]: I1126 00:54:56.533816 4766 scope.go:117] "RemoveContainer" containerID="d4c6a754f1e46d07682e51b007f6b063944a5b52ef1a6fec9d341fff511e1639" Nov 26 00:54:56 crc kubenswrapper[4766]: I1126 00:54:56.561645 4766 scope.go:117] "RemoveContainer" containerID="617dc37b7d60acafb755ea09418c5c542e2891fb518eacc33195b5d71d7d306c" Nov 26 00:54:57 crc kubenswrapper[4766]: I1126 00:54:57.856236 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04d8e581-a471-4aea-8c66-016ae86cd9e8" path="/var/lib/kubelet/pods/04d8e581-a471-4aea-8c66-016ae86cd9e8/volumes" Nov 26 00:54:57 crc kubenswrapper[4766]: I1126 00:54:57.858431 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adde2fcf-e073-4936-a856-16bd660658a5" path="/var/lib/kubelet/pods/adde2fcf-e073-4936-a856-16bd660658a5/volumes" Nov 26 00:55:02 crc kubenswrapper[4766]: I1126 00:55:02.048348 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-g5j89"] Nov 26 00:55:02 crc kubenswrapper[4766]: I1126 00:55:02.067074 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-g5j89"] Nov 26 00:55:03 crc kubenswrapper[4766]: I1126 00:55:03.841299 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e0f8d57-5be8-4851-b8cf-e4ebdf753048" path="/var/lib/kubelet/pods/7e0f8d57-5be8-4851-b8cf-e4ebdf753048/volumes" Nov 26 00:55:06 crc kubenswrapper[4766]: I1126 00:55:06.389454 4766 generic.go:334] "Generic (PLEG): container finished" podID="7fee6935-5d49-4559-93de-8b244b5a6261" containerID="aba7c1527460d45400164d3ba4df993725b996e9df22aebb71e574413d815763" exitCode=0 Nov 26 00:55:06 crc kubenswrapper[4766]: I1126 00:55:06.389966 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8" event={"ID":"7fee6935-5d49-4559-93de-8b244b5a6261","Type":"ContainerDied","Data":"aba7c1527460d45400164d3ba4df993725b996e9df22aebb71e574413d815763"} Nov 26 00:55:07 crc kubenswrapper[4766]: I1126 00:55:07.973618 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.097855 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqzth\" (UniqueName: \"kubernetes.io/projected/7fee6935-5d49-4559-93de-8b244b5a6261-kube-api-access-sqzth\") pod \"7fee6935-5d49-4559-93de-8b244b5a6261\" (UID: \"7fee6935-5d49-4559-93de-8b244b5a6261\") " Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.098012 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fee6935-5d49-4559-93de-8b244b5a6261-ssh-key\") pod \"7fee6935-5d49-4559-93de-8b244b5a6261\" (UID: \"7fee6935-5d49-4559-93de-8b244b5a6261\") " Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.098186 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fee6935-5d49-4559-93de-8b244b5a6261-bootstrap-combined-ca-bundle\") pod \"7fee6935-5d49-4559-93de-8b244b5a6261\" (UID: \"7fee6935-5d49-4559-93de-8b244b5a6261\") " Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.098275 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fee6935-5d49-4559-93de-8b244b5a6261-inventory\") pod \"7fee6935-5d49-4559-93de-8b244b5a6261\" (UID: \"7fee6935-5d49-4559-93de-8b244b5a6261\") " Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.104145 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fee6935-5d49-4559-93de-8b244b5a6261-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "7fee6935-5d49-4559-93de-8b244b5a6261" (UID: "7fee6935-5d49-4559-93de-8b244b5a6261"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.108390 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fee6935-5d49-4559-93de-8b244b5a6261-kube-api-access-sqzth" (OuterVolumeSpecName: "kube-api-access-sqzth") pod "7fee6935-5d49-4559-93de-8b244b5a6261" (UID: "7fee6935-5d49-4559-93de-8b244b5a6261"). InnerVolumeSpecName "kube-api-access-sqzth". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.133254 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fee6935-5d49-4559-93de-8b244b5a6261-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7fee6935-5d49-4559-93de-8b244b5a6261" (UID: "7fee6935-5d49-4559-93de-8b244b5a6261"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.145987 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fee6935-5d49-4559-93de-8b244b5a6261-inventory" (OuterVolumeSpecName: "inventory") pod "7fee6935-5d49-4559-93de-8b244b5a6261" (UID: "7fee6935-5d49-4559-93de-8b244b5a6261"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.201154 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fee6935-5d49-4559-93de-8b244b5a6261-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.201185 4766 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fee6935-5d49-4559-93de-8b244b5a6261-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.201196 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fee6935-5d49-4559-93de-8b244b5a6261-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.201206 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqzth\" (UniqueName: \"kubernetes.io/projected/7fee6935-5d49-4559-93de-8b244b5a6261-kube-api-access-sqzth\") on node \"crc\" DevicePath \"\"" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.415206 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8" event={"ID":"7fee6935-5d49-4559-93de-8b244b5a6261","Type":"ContainerDied","Data":"5987a53ae26c7c4be98c1c4f9a4e59c077745c424d596f036cd3f0bfe922f660"} Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.415247 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5987a53ae26c7c4be98c1c4f9a4e59c077745c424d596f036cd3f0bfe922f660" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.415305 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.537834 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x"] Nov 26 00:55:08 crc kubenswrapper[4766]: E1126 00:55:08.538578 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fee6935-5d49-4559-93de-8b244b5a6261" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.538606 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fee6935-5d49-4559-93de-8b244b5a6261" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.539080 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fee6935-5d49-4559-93de-8b244b5a6261" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.540482 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.543872 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.544098 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.544107 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.544156 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.556305 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x"] Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.718588 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5356fee3-3206-46eb-b466-4eb864ecb237-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x\" (UID: \"5356fee3-3206-46eb-b466-4eb864ecb237\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.718810 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5356fee3-3206-46eb-b466-4eb864ecb237-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x\" (UID: \"5356fee3-3206-46eb-b466-4eb864ecb237\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.718926 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bvtv\" (UniqueName: \"kubernetes.io/projected/5356fee3-3206-46eb-b466-4eb864ecb237-kube-api-access-7bvtv\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x\" (UID: \"5356fee3-3206-46eb-b466-4eb864ecb237\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.821052 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5356fee3-3206-46eb-b466-4eb864ecb237-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x\" (UID: \"5356fee3-3206-46eb-b466-4eb864ecb237\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.821306 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5356fee3-3206-46eb-b466-4eb864ecb237-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x\" (UID: \"5356fee3-3206-46eb-b466-4eb864ecb237\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.821343 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bvtv\" (UniqueName: \"kubernetes.io/projected/5356fee3-3206-46eb-b466-4eb864ecb237-kube-api-access-7bvtv\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x\" (UID: \"5356fee3-3206-46eb-b466-4eb864ecb237\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.831454 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5356fee3-3206-46eb-b466-4eb864ecb237-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x\" (UID: \"5356fee3-3206-46eb-b466-4eb864ecb237\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.832093 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5356fee3-3206-46eb-b466-4eb864ecb237-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x\" (UID: \"5356fee3-3206-46eb-b466-4eb864ecb237\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.850574 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bvtv\" (UniqueName: \"kubernetes.io/projected/5356fee3-3206-46eb-b466-4eb864ecb237-kube-api-access-7bvtv\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x\" (UID: \"5356fee3-3206-46eb-b466-4eb864ecb237\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x" Nov 26 00:55:08 crc kubenswrapper[4766]: I1126 00:55:08.868692 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x" Nov 26 00:55:09 crc kubenswrapper[4766]: I1126 00:55:09.752589 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 00:55:09 crc kubenswrapper[4766]: I1126 00:55:09.754006 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x"] Nov 26 00:55:10 crc kubenswrapper[4766]: I1126 00:55:10.661910 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x" event={"ID":"5356fee3-3206-46eb-b466-4eb864ecb237","Type":"ContainerStarted","Data":"5e89130c9d6c70aa05b6f3167f77889c6d7bf84025092501efda329aea717017"} Nov 26 00:55:11 crc kubenswrapper[4766]: I1126 00:55:11.688520 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x" event={"ID":"5356fee3-3206-46eb-b466-4eb864ecb237","Type":"ContainerStarted","Data":"c20e1ac5e10091502b723502fb56034373de52996074b49ba01137cad6b8a1df"} Nov 26 00:55:11 crc kubenswrapper[4766]: I1126 00:55:11.711235 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x" podStartSLOduration=3.013908957 podStartE2EDuration="3.711215784s" podCreationTimestamp="2025-11-26 00:55:08 +0000 UTC" firstStartedPulling="2025-11-26 00:55:09.752407084 +0000 UTC m=+1890.601177514" lastFinishedPulling="2025-11-26 00:55:10.449713901 +0000 UTC m=+1891.298484341" observedRunningTime="2025-11-26 00:55:11.70816046 +0000 UTC m=+1892.556930910" watchObservedRunningTime="2025-11-26 00:55:11.711215784 +0000 UTC m=+1892.559986224" Nov 26 00:55:31 crc kubenswrapper[4766]: I1126 00:55:31.059729 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-8bfjz"] Nov 26 00:55:31 crc kubenswrapper[4766]: I1126 00:55:31.075736 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-8bfjz"] Nov 26 00:55:31 crc kubenswrapper[4766]: I1126 00:55:31.874087 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b156a1fc-37a3-45d3-999d-55852f10bcc3" path="/var/lib/kubelet/pods/b156a1fc-37a3-45d3-999d-55852f10bcc3/volumes" Nov 26 00:55:37 crc kubenswrapper[4766]: I1126 00:55:37.073778 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-d26z8"] Nov 26 00:55:37 crc kubenswrapper[4766]: I1126 00:55:37.089498 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-7rfnj"] Nov 26 00:55:37 crc kubenswrapper[4766]: I1126 00:55:37.101713 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-d26z8"] Nov 26 00:55:37 crc kubenswrapper[4766]: I1126 00:55:37.111881 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-pd2t2"] Nov 26 00:55:37 crc kubenswrapper[4766]: I1126 00:55:37.123215 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-7rfnj"] Nov 26 00:55:37 crc kubenswrapper[4766]: I1126 00:55:37.132570 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-pd2t2"] Nov 26 00:55:37 crc kubenswrapper[4766]: I1126 00:55:37.847086 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66feb239-5790-4cfe-96db-e9dd6413bb09" path="/var/lib/kubelet/pods/66feb239-5790-4cfe-96db-e9dd6413bb09/volumes" Nov 26 00:55:37 crc kubenswrapper[4766]: I1126 00:55:37.847826 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6969780f-beaf-4216-8516-1eea9445f0c4" path="/var/lib/kubelet/pods/6969780f-beaf-4216-8516-1eea9445f0c4/volumes" Nov 26 00:55:37 crc kubenswrapper[4766]: I1126 00:55:37.848565 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d2fbed9-754d-45c7-a8c1-7dc4828b26a8" path="/var/lib/kubelet/pods/7d2fbed9-754d-45c7-a8c1-7dc4828b26a8/volumes" Nov 26 00:55:57 crc kubenswrapper[4766]: I1126 00:55:57.035725 4766 scope.go:117] "RemoveContainer" containerID="a0222e88d05b9e6f27da2a12b178be333fccf25b344163e8ae9b92d4c7116e94" Nov 26 00:55:57 crc kubenswrapper[4766]: I1126 00:55:57.103981 4766 scope.go:117] "RemoveContainer" containerID="c41bd398fb0aeb4ea7712daecc1565be2bc9a263b141abc17405ebb24c8de9ce" Nov 26 00:55:57 crc kubenswrapper[4766]: I1126 00:55:57.153085 4766 scope.go:117] "RemoveContainer" containerID="89528903de6bdb75b16ed3af567fb4d84f8c699933dfe36d9f49023cb06484b2" Nov 26 00:55:57 crc kubenswrapper[4766]: I1126 00:55:57.212153 4766 scope.go:117] "RemoveContainer" containerID="28f98fa01a48d0f4bded085fc78451fad87337d0a0eecc968095e0c60f6edffe" Nov 26 00:55:57 crc kubenswrapper[4766]: I1126 00:55:57.270749 4766 scope.go:117] "RemoveContainer" containerID="bb7285908ffbbf45d040e880724046e64e81ab2eea095e40609884bee5494e08" Nov 26 00:55:57 crc kubenswrapper[4766]: I1126 00:55:57.301806 4766 scope.go:117] "RemoveContainer" containerID="6673e046e2197a4b6fac11035a178781f4e83f5c3084eaaeefdd9e783da3f0ff" Nov 26 00:55:57 crc kubenswrapper[4766]: I1126 00:55:57.356009 4766 scope.go:117] "RemoveContainer" containerID="e0731b84189b5cf1d81b460c924c87afd8f2e54e09bd83d8ffc0da67f42e28fc" Nov 26 00:55:57 crc kubenswrapper[4766]: I1126 00:55:57.404906 4766 scope.go:117] "RemoveContainer" containerID="e0fc6733b38804513dacda343aa208bb447365560aaa4baf35e5bbe9390d787a" Nov 26 00:55:57 crc kubenswrapper[4766]: I1126 00:55:57.446062 4766 scope.go:117] "RemoveContainer" containerID="063c2b5642c0debff8bdd6d1fb44f28d827eb58ca69c0244076b2830713167a0" Nov 26 00:55:57 crc kubenswrapper[4766]: I1126 00:55:57.476237 4766 scope.go:117] "RemoveContainer" containerID="3718f9376f2bea3d70776252e88a7eb5fcc27e2ec33341a258d4bd09f79aac4c" Nov 26 00:55:57 crc kubenswrapper[4766]: I1126 00:55:57.522391 4766 scope.go:117] "RemoveContainer" containerID="b2138ff204dd583ad9f427fd6407d82c57b26b12b9bb1433ac804293133a3552" Nov 26 00:55:57 crc kubenswrapper[4766]: I1126 00:55:57.550031 4766 scope.go:117] "RemoveContainer" containerID="2a9815005b07139bc4c1db61fdce11fd291bd00028e65ab55088bfc31dbe5001" Nov 26 00:55:57 crc kubenswrapper[4766]: I1126 00:55:57.580309 4766 scope.go:117] "RemoveContainer" containerID="64ee6e6eb224e1d80393c27b907b09bb3ed85df968b78774ec928cde0defffcb" Nov 26 00:55:57 crc kubenswrapper[4766]: I1126 00:55:57.602750 4766 scope.go:117] "RemoveContainer" containerID="252626f0723e2a175f72d821ec6908dff31e1409f3ca28fd691e29d2869b2be6" Nov 26 00:55:57 crc kubenswrapper[4766]: I1126 00:55:57.627293 4766 scope.go:117] "RemoveContainer" containerID="81a18a1a1a33fce2ef59e20de6029e3f5ecb95513bd41c361e5d3a25170b3a8c" Nov 26 00:55:57 crc kubenswrapper[4766]: I1126 00:55:57.649063 4766 scope.go:117] "RemoveContainer" containerID="530f9b909755d33d9f92a02281b30b40fd91c52515c3b64e26f9b0b5e00801c5" Nov 26 00:56:01 crc kubenswrapper[4766]: I1126 00:56:01.046881 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-8sbgx"] Nov 26 00:56:01 crc kubenswrapper[4766]: I1126 00:56:01.060629 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-8sbgx"] Nov 26 00:56:01 crc kubenswrapper[4766]: I1126 00:56:01.848089 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffaf6d04-ed5d-4438-a285-bc4981448be5" path="/var/lib/kubelet/pods/ffaf6d04-ed5d-4438-a285-bc4981448be5/volumes" Nov 26 00:56:27 crc kubenswrapper[4766]: I1126 00:56:27.733907 4766 generic.go:334] "Generic (PLEG): container finished" podID="5356fee3-3206-46eb-b466-4eb864ecb237" containerID="c20e1ac5e10091502b723502fb56034373de52996074b49ba01137cad6b8a1df" exitCode=0 Nov 26 00:56:27 crc kubenswrapper[4766]: I1126 00:56:27.733978 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x" event={"ID":"5356fee3-3206-46eb-b466-4eb864ecb237","Type":"ContainerDied","Data":"c20e1ac5e10091502b723502fb56034373de52996074b49ba01137cad6b8a1df"} Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.278831 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x" Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.341563 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5356fee3-3206-46eb-b466-4eb864ecb237-ssh-key\") pod \"5356fee3-3206-46eb-b466-4eb864ecb237\" (UID: \"5356fee3-3206-46eb-b466-4eb864ecb237\") " Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.341733 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bvtv\" (UniqueName: \"kubernetes.io/projected/5356fee3-3206-46eb-b466-4eb864ecb237-kube-api-access-7bvtv\") pod \"5356fee3-3206-46eb-b466-4eb864ecb237\" (UID: \"5356fee3-3206-46eb-b466-4eb864ecb237\") " Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.341791 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5356fee3-3206-46eb-b466-4eb864ecb237-inventory\") pod \"5356fee3-3206-46eb-b466-4eb864ecb237\" (UID: \"5356fee3-3206-46eb-b466-4eb864ecb237\") " Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.350866 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5356fee3-3206-46eb-b466-4eb864ecb237-kube-api-access-7bvtv" (OuterVolumeSpecName: "kube-api-access-7bvtv") pod "5356fee3-3206-46eb-b466-4eb864ecb237" (UID: "5356fee3-3206-46eb-b466-4eb864ecb237"). InnerVolumeSpecName "kube-api-access-7bvtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.373626 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5356fee3-3206-46eb-b466-4eb864ecb237-inventory" (OuterVolumeSpecName: "inventory") pod "5356fee3-3206-46eb-b466-4eb864ecb237" (UID: "5356fee3-3206-46eb-b466-4eb864ecb237"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.396003 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5356fee3-3206-46eb-b466-4eb864ecb237-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5356fee3-3206-46eb-b466-4eb864ecb237" (UID: "5356fee3-3206-46eb-b466-4eb864ecb237"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.444341 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bvtv\" (UniqueName: \"kubernetes.io/projected/5356fee3-3206-46eb-b466-4eb864ecb237-kube-api-access-7bvtv\") on node \"crc\" DevicePath \"\"" Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.444390 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5356fee3-3206-46eb-b466-4eb864ecb237-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.444402 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5356fee3-3206-46eb-b466-4eb864ecb237-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.764034 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x" event={"ID":"5356fee3-3206-46eb-b466-4eb864ecb237","Type":"ContainerDied","Data":"5e89130c9d6c70aa05b6f3167f77889c6d7bf84025092501efda329aea717017"} Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.764413 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e89130c9d6c70aa05b6f3167f77889c6d7bf84025092501efda329aea717017" Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.764120 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x" Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.894985 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x"] Nov 26 00:56:29 crc kubenswrapper[4766]: E1126 00:56:29.895483 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5356fee3-3206-46eb-b466-4eb864ecb237" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.895496 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5356fee3-3206-46eb-b466-4eb864ecb237" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.895723 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="5356fee3-3206-46eb-b466-4eb864ecb237" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.896465 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x" Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.898670 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.899228 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.899445 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.899834 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 00:56:29 crc kubenswrapper[4766]: I1126 00:56:29.909126 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x"] Nov 26 00:56:30 crc kubenswrapper[4766]: I1126 00:56:30.059355 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x\" (UID: \"ac83f8e3-4d14-4f01-992d-ab5c7fcacaed\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x" Nov 26 00:56:30 crc kubenswrapper[4766]: I1126 00:56:30.059601 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hcb2\" (UniqueName: \"kubernetes.io/projected/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed-kube-api-access-7hcb2\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x\" (UID: \"ac83f8e3-4d14-4f01-992d-ab5c7fcacaed\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x" Nov 26 00:56:30 crc kubenswrapper[4766]: I1126 00:56:30.060066 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x\" (UID: \"ac83f8e3-4d14-4f01-992d-ab5c7fcacaed\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x" Nov 26 00:56:30 crc kubenswrapper[4766]: I1126 00:56:30.162832 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x\" (UID: \"ac83f8e3-4d14-4f01-992d-ab5c7fcacaed\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x" Nov 26 00:56:30 crc kubenswrapper[4766]: I1126 00:56:30.163050 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x\" (UID: \"ac83f8e3-4d14-4f01-992d-ab5c7fcacaed\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x" Nov 26 00:56:30 crc kubenswrapper[4766]: I1126 00:56:30.163160 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hcb2\" (UniqueName: \"kubernetes.io/projected/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed-kube-api-access-7hcb2\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x\" (UID: \"ac83f8e3-4d14-4f01-992d-ab5c7fcacaed\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x" Nov 26 00:56:30 crc kubenswrapper[4766]: I1126 00:56:30.170484 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x\" (UID: \"ac83f8e3-4d14-4f01-992d-ab5c7fcacaed\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x" Nov 26 00:56:30 crc kubenswrapper[4766]: I1126 00:56:30.176916 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x\" (UID: \"ac83f8e3-4d14-4f01-992d-ab5c7fcacaed\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x" Nov 26 00:56:30 crc kubenswrapper[4766]: I1126 00:56:30.193854 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hcb2\" (UniqueName: \"kubernetes.io/projected/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed-kube-api-access-7hcb2\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x\" (UID: \"ac83f8e3-4d14-4f01-992d-ab5c7fcacaed\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x" Nov 26 00:56:30 crc kubenswrapper[4766]: I1126 00:56:30.219684 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x" Nov 26 00:56:30 crc kubenswrapper[4766]: W1126 00:56:30.870084 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac83f8e3_4d14_4f01_992d_ab5c7fcacaed.slice/crio-dfac22bd2561e8a0185a61b47599195884730f81981dad562260016967c566c2 WatchSource:0}: Error finding container dfac22bd2561e8a0185a61b47599195884730f81981dad562260016967c566c2: Status 404 returned error can't find the container with id dfac22bd2561e8a0185a61b47599195884730f81981dad562260016967c566c2 Nov 26 00:56:30 crc kubenswrapper[4766]: I1126 00:56:30.871734 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x"] Nov 26 00:56:31 crc kubenswrapper[4766]: I1126 00:56:31.790461 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x" event={"ID":"ac83f8e3-4d14-4f01-992d-ab5c7fcacaed","Type":"ContainerStarted","Data":"68c7016a858c4a7a21c6443a85bc5d35e9693d290303733a56a2c665e77fb55f"} Nov 26 00:56:31 crc kubenswrapper[4766]: I1126 00:56:31.791510 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x" event={"ID":"ac83f8e3-4d14-4f01-992d-ab5c7fcacaed","Type":"ContainerStarted","Data":"dfac22bd2561e8a0185a61b47599195884730f81981dad562260016967c566c2"} Nov 26 00:56:31 crc kubenswrapper[4766]: I1126 00:56:31.818944 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x" podStartSLOduration=2.306257678 podStartE2EDuration="2.818925437s" podCreationTimestamp="2025-11-26 00:56:29 +0000 UTC" firstStartedPulling="2025-11-26 00:56:30.875765846 +0000 UTC m=+1971.724536306" lastFinishedPulling="2025-11-26 00:56:31.388433625 +0000 UTC m=+1972.237204065" observedRunningTime="2025-11-26 00:56:31.812574672 +0000 UTC m=+1972.661345142" watchObservedRunningTime="2025-11-26 00:56:31.818925437 +0000 UTC m=+1972.667695867" Nov 26 00:56:38 crc kubenswrapper[4766]: I1126 00:56:38.902391 4766 generic.go:334] "Generic (PLEG): container finished" podID="ac83f8e3-4d14-4f01-992d-ab5c7fcacaed" containerID="68c7016a858c4a7a21c6443a85bc5d35e9693d290303733a56a2c665e77fb55f" exitCode=0 Nov 26 00:56:38 crc kubenswrapper[4766]: I1126 00:56:38.902552 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x" event={"ID":"ac83f8e3-4d14-4f01-992d-ab5c7fcacaed","Type":"ContainerDied","Data":"68c7016a858c4a7a21c6443a85bc5d35e9693d290303733a56a2c665e77fb55f"} Nov 26 00:56:40 crc kubenswrapper[4766]: I1126 00:56:40.426952 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x" Nov 26 00:56:40 crc kubenswrapper[4766]: I1126 00:56:40.541247 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed-inventory\") pod \"ac83f8e3-4d14-4f01-992d-ab5c7fcacaed\" (UID: \"ac83f8e3-4d14-4f01-992d-ab5c7fcacaed\") " Nov 26 00:56:40 crc kubenswrapper[4766]: I1126 00:56:40.541424 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hcb2\" (UniqueName: \"kubernetes.io/projected/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed-kube-api-access-7hcb2\") pod \"ac83f8e3-4d14-4f01-992d-ab5c7fcacaed\" (UID: \"ac83f8e3-4d14-4f01-992d-ab5c7fcacaed\") " Nov 26 00:56:40 crc kubenswrapper[4766]: I1126 00:56:40.541561 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed-ssh-key\") pod \"ac83f8e3-4d14-4f01-992d-ab5c7fcacaed\" (UID: \"ac83f8e3-4d14-4f01-992d-ab5c7fcacaed\") " Nov 26 00:56:40 crc kubenswrapper[4766]: I1126 00:56:40.551111 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed-kube-api-access-7hcb2" (OuterVolumeSpecName: "kube-api-access-7hcb2") pod "ac83f8e3-4d14-4f01-992d-ab5c7fcacaed" (UID: "ac83f8e3-4d14-4f01-992d-ab5c7fcacaed"). InnerVolumeSpecName "kube-api-access-7hcb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:56:40 crc kubenswrapper[4766]: E1126 00:56:40.571010 4766 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed-inventory podName:ac83f8e3-4d14-4f01-992d-ab5c7fcacaed nodeName:}" failed. No retries permitted until 2025-11-26 00:56:41.070956926 +0000 UTC m=+1981.919727356 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed-inventory") pod "ac83f8e3-4d14-4f01-992d-ab5c7fcacaed" (UID: "ac83f8e3-4d14-4f01-992d-ab5c7fcacaed") : error deleting /var/lib/kubelet/pods/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed/volume-subpaths: remove /var/lib/kubelet/pods/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed/volume-subpaths: no such file or directory Nov 26 00:56:40 crc kubenswrapper[4766]: I1126 00:56:40.577629 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ac83f8e3-4d14-4f01-992d-ab5c7fcacaed" (UID: "ac83f8e3-4d14-4f01-992d-ab5c7fcacaed"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:56:40 crc kubenswrapper[4766]: I1126 00:56:40.643833 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hcb2\" (UniqueName: \"kubernetes.io/projected/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed-kube-api-access-7hcb2\") on node \"crc\" DevicePath \"\"" Nov 26 00:56:40 crc kubenswrapper[4766]: I1126 00:56:40.643875 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 00:56:40 crc kubenswrapper[4766]: I1126 00:56:40.945387 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x" event={"ID":"ac83f8e3-4d14-4f01-992d-ab5c7fcacaed","Type":"ContainerDied","Data":"dfac22bd2561e8a0185a61b47599195884730f81981dad562260016967c566c2"} Nov 26 00:56:40 crc kubenswrapper[4766]: I1126 00:56:40.945730 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfac22bd2561e8a0185a61b47599195884730f81981dad562260016967c566c2" Nov 26 00:56:40 crc kubenswrapper[4766]: I1126 00:56:40.945603 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x" Nov 26 00:56:41 crc kubenswrapper[4766]: I1126 00:56:41.085197 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb"] Nov 26 00:56:41 crc kubenswrapper[4766]: E1126 00:56:41.085620 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac83f8e3-4d14-4f01-992d-ab5c7fcacaed" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 26 00:56:41 crc kubenswrapper[4766]: I1126 00:56:41.085637 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac83f8e3-4d14-4f01-992d-ab5c7fcacaed" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 26 00:56:41 crc kubenswrapper[4766]: I1126 00:56:41.085866 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac83f8e3-4d14-4f01-992d-ab5c7fcacaed" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 26 00:56:41 crc kubenswrapper[4766]: I1126 00:56:41.086638 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb" Nov 26 00:56:41 crc kubenswrapper[4766]: I1126 00:56:41.102938 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb"] Nov 26 00:56:41 crc kubenswrapper[4766]: I1126 00:56:41.153996 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed-inventory\") pod \"ac83f8e3-4d14-4f01-992d-ab5c7fcacaed\" (UID: \"ac83f8e3-4d14-4f01-992d-ab5c7fcacaed\") " Nov 26 00:56:41 crc kubenswrapper[4766]: I1126 00:56:41.162503 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed-inventory" (OuterVolumeSpecName: "inventory") pod "ac83f8e3-4d14-4f01-992d-ab5c7fcacaed" (UID: "ac83f8e3-4d14-4f01-992d-ab5c7fcacaed"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:56:41 crc kubenswrapper[4766]: I1126 00:56:41.257546 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecefedc9-a319-4ada-99e6-8e7bf2ad5da5-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6t2cb\" (UID: \"ecefedc9-a319-4ada-99e6-8e7bf2ad5da5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb" Nov 26 00:56:41 crc kubenswrapper[4766]: I1126 00:56:41.257591 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf4rz\" (UniqueName: \"kubernetes.io/projected/ecefedc9-a319-4ada-99e6-8e7bf2ad5da5-kube-api-access-tf4rz\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6t2cb\" (UID: \"ecefedc9-a319-4ada-99e6-8e7bf2ad5da5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb" Nov 26 00:56:41 crc kubenswrapper[4766]: I1126 00:56:41.258057 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecefedc9-a319-4ada-99e6-8e7bf2ad5da5-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6t2cb\" (UID: \"ecefedc9-a319-4ada-99e6-8e7bf2ad5da5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb" Nov 26 00:56:41 crc kubenswrapper[4766]: I1126 00:56:41.258217 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 00:56:41 crc kubenswrapper[4766]: I1126 00:56:41.360120 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecefedc9-a319-4ada-99e6-8e7bf2ad5da5-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6t2cb\" (UID: \"ecefedc9-a319-4ada-99e6-8e7bf2ad5da5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb" Nov 26 00:56:41 crc kubenswrapper[4766]: I1126 00:56:41.360217 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecefedc9-a319-4ada-99e6-8e7bf2ad5da5-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6t2cb\" (UID: \"ecefedc9-a319-4ada-99e6-8e7bf2ad5da5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb" Nov 26 00:56:41 crc kubenswrapper[4766]: I1126 00:56:41.360235 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf4rz\" (UniqueName: \"kubernetes.io/projected/ecefedc9-a319-4ada-99e6-8e7bf2ad5da5-kube-api-access-tf4rz\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6t2cb\" (UID: \"ecefedc9-a319-4ada-99e6-8e7bf2ad5da5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb" Nov 26 00:56:41 crc kubenswrapper[4766]: I1126 00:56:41.367223 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecefedc9-a319-4ada-99e6-8e7bf2ad5da5-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6t2cb\" (UID: \"ecefedc9-a319-4ada-99e6-8e7bf2ad5da5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb" Nov 26 00:56:41 crc kubenswrapper[4766]: I1126 00:56:41.369599 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecefedc9-a319-4ada-99e6-8e7bf2ad5da5-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6t2cb\" (UID: \"ecefedc9-a319-4ada-99e6-8e7bf2ad5da5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb" Nov 26 00:56:41 crc kubenswrapper[4766]: I1126 00:56:41.395309 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf4rz\" (UniqueName: \"kubernetes.io/projected/ecefedc9-a319-4ada-99e6-8e7bf2ad5da5-kube-api-access-tf4rz\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6t2cb\" (UID: \"ecefedc9-a319-4ada-99e6-8e7bf2ad5da5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb" Nov 26 00:56:41 crc kubenswrapper[4766]: I1126 00:56:41.415906 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb" Nov 26 00:56:41 crc kubenswrapper[4766]: I1126 00:56:41.480240 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:56:41 crc kubenswrapper[4766]: I1126 00:56:41.480305 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:56:42 crc kubenswrapper[4766]: I1126 00:56:42.020976 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb"] Nov 26 00:56:42 crc kubenswrapper[4766]: W1126 00:56:42.021418 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecefedc9_a319_4ada_99e6_8e7bf2ad5da5.slice/crio-8e5f258993c14bc1ec7b0d4b364497925d8e95dbfd4205081688aa0c269fa5e9 WatchSource:0}: Error finding container 8e5f258993c14bc1ec7b0d4b364497925d8e95dbfd4205081688aa0c269fa5e9: Status 404 returned error can't find the container with id 8e5f258993c14bc1ec7b0d4b364497925d8e95dbfd4205081688aa0c269fa5e9 Nov 26 00:56:42 crc kubenswrapper[4766]: I1126 00:56:42.971832 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb" event={"ID":"ecefedc9-a319-4ada-99e6-8e7bf2ad5da5","Type":"ContainerStarted","Data":"7262c3217748cb4f2c388179da8f8d6b47d5240c1fe26815cc76eb98a2f6dec3"} Nov 26 00:56:42 crc kubenswrapper[4766]: I1126 00:56:42.972185 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb" event={"ID":"ecefedc9-a319-4ada-99e6-8e7bf2ad5da5","Type":"ContainerStarted","Data":"8e5f258993c14bc1ec7b0d4b364497925d8e95dbfd4205081688aa0c269fa5e9"} Nov 26 00:56:43 crc kubenswrapper[4766]: I1126 00:56:43.001510 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb" podStartSLOduration=1.550743839 podStartE2EDuration="2.001491896s" podCreationTimestamp="2025-11-26 00:56:41 +0000 UTC" firstStartedPulling="2025-11-26 00:56:42.024406067 +0000 UTC m=+1982.873176497" lastFinishedPulling="2025-11-26 00:56:42.475154094 +0000 UTC m=+1983.323924554" observedRunningTime="2025-11-26 00:56:42.986948381 +0000 UTC m=+1983.835718911" watchObservedRunningTime="2025-11-26 00:56:43.001491896 +0000 UTC m=+1983.850262336" Nov 26 00:56:52 crc kubenswrapper[4766]: I1126 00:56:52.074488 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-k2tlc"] Nov 26 00:56:52 crc kubenswrapper[4766]: I1126 00:56:52.093619 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-t4vp6"] Nov 26 00:56:52 crc kubenswrapper[4766]: I1126 00:56:52.107644 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-88b3-account-create-update-zvx6c"] Nov 26 00:56:52 crc kubenswrapper[4766]: I1126 00:56:52.120756 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-t4vp6"] Nov 26 00:56:52 crc kubenswrapper[4766]: I1126 00:56:52.134281 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-88b3-account-create-update-zvx6c"] Nov 26 00:56:52 crc kubenswrapper[4766]: I1126 00:56:52.142928 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-k2tlc"] Nov 26 00:56:53 crc kubenswrapper[4766]: I1126 00:56:53.039869 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-d77a-account-create-update-8gxtj"] Nov 26 00:56:53 crc kubenswrapper[4766]: I1126 00:56:53.054097 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-jgd8f"] Nov 26 00:56:53 crc kubenswrapper[4766]: I1126 00:56:53.068446 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-f00c-account-create-update-7jpjm"] Nov 26 00:56:53 crc kubenswrapper[4766]: I1126 00:56:53.078492 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-d77a-account-create-update-8gxtj"] Nov 26 00:56:53 crc kubenswrapper[4766]: I1126 00:56:53.088780 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-jgd8f"] Nov 26 00:56:53 crc kubenswrapper[4766]: I1126 00:56:53.109096 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-f00c-account-create-update-7jpjm"] Nov 26 00:56:53 crc kubenswrapper[4766]: I1126 00:56:53.847199 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2046b547-611f-4e98-ba95-a34cd0d45f50" path="/var/lib/kubelet/pods/2046b547-611f-4e98-ba95-a34cd0d45f50/volumes" Nov 26 00:56:53 crc kubenswrapper[4766]: I1126 00:56:53.847751 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d4b645a-ddab-4a29-8ffa-77e7873591c8" path="/var/lib/kubelet/pods/2d4b645a-ddab-4a29-8ffa-77e7873591c8/volumes" Nov 26 00:56:53 crc kubenswrapper[4766]: I1126 00:56:53.848273 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4794066a-0e54-4387-ae01-078abc184684" path="/var/lib/kubelet/pods/4794066a-0e54-4387-ae01-078abc184684/volumes" Nov 26 00:56:53 crc kubenswrapper[4766]: I1126 00:56:53.848783 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="619650be-59d2-47c8-815d-3337c156c83f" path="/var/lib/kubelet/pods/619650be-59d2-47c8-815d-3337c156c83f/volumes" Nov 26 00:56:53 crc kubenswrapper[4766]: I1126 00:56:53.849746 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="751e5ae1-be9d-48e8-8cc1-e08012854203" path="/var/lib/kubelet/pods/751e5ae1-be9d-48e8-8cc1-e08012854203/volumes" Nov 26 00:56:53 crc kubenswrapper[4766]: I1126 00:56:53.850249 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b20f822d-71d5-4cb8-823b-8c261af60109" path="/var/lib/kubelet/pods/b20f822d-71d5-4cb8-823b-8c261af60109/volumes" Nov 26 00:56:57 crc kubenswrapper[4766]: I1126 00:56:57.938504 4766 scope.go:117] "RemoveContainer" containerID="97606350c9b9dcf26aa878a45371aa3ff32e85934db2d23fad1153f8d0243913" Nov 26 00:56:57 crc kubenswrapper[4766]: I1126 00:56:57.969390 4766 scope.go:117] "RemoveContainer" containerID="4cd953a4be032f4df75da118bdc15379473d4b9254e10720f6d0045cb08ac8a2" Nov 26 00:56:58 crc kubenswrapper[4766]: I1126 00:56:58.063529 4766 scope.go:117] "RemoveContainer" containerID="ad711cfd11b280b1818b457932f68682de01db5dad05a749c800cdb8c87f64f3" Nov 26 00:56:58 crc kubenswrapper[4766]: I1126 00:56:58.130043 4766 scope.go:117] "RemoveContainer" containerID="7ef0ad34acd79fd20333d11bb7ff0b1dcfccd0cb0517a49063c065d8eaf6eeb0" Nov 26 00:56:58 crc kubenswrapper[4766]: I1126 00:56:58.197705 4766 scope.go:117] "RemoveContainer" containerID="5c730b2e7bbdd15b5a9c9c7517ec31cbe511dffb42429118fd3355ce409a52af" Nov 26 00:56:58 crc kubenswrapper[4766]: I1126 00:56:58.282506 4766 scope.go:117] "RemoveContainer" containerID="fd8749b745d14b29e866e787691471be81cb43466bb8932ef270ac001dc988eb" Nov 26 00:56:58 crc kubenswrapper[4766]: I1126 00:56:58.324849 4766 scope.go:117] "RemoveContainer" containerID="e18a280668d8cba4f4a4e353024890c172f809b1751ab6ef5b6847fd11f4e2e3" Nov 26 00:57:11 crc kubenswrapper[4766]: I1126 00:57:11.479835 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:57:11 crc kubenswrapper[4766]: I1126 00:57:11.481285 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:57:20 crc kubenswrapper[4766]: I1126 00:57:20.059207 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2t966"] Nov 26 00:57:20 crc kubenswrapper[4766]: I1126 00:57:20.075429 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-2t966"] Nov 26 00:57:21 crc kubenswrapper[4766]: I1126 00:57:21.838670 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6612de23-8b59-474c-9089-2559ecda8811" path="/var/lib/kubelet/pods/6612de23-8b59-474c-9089-2559ecda8811/volumes" Nov 26 00:57:26 crc kubenswrapper[4766]: I1126 00:57:26.536528 4766 generic.go:334] "Generic (PLEG): container finished" podID="ecefedc9-a319-4ada-99e6-8e7bf2ad5da5" containerID="7262c3217748cb4f2c388179da8f8d6b47d5240c1fe26815cc76eb98a2f6dec3" exitCode=0 Nov 26 00:57:26 crc kubenswrapper[4766]: I1126 00:57:26.536617 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb" event={"ID":"ecefedc9-a319-4ada-99e6-8e7bf2ad5da5","Type":"ContainerDied","Data":"7262c3217748cb4f2c388179da8f8d6b47d5240c1fe26815cc76eb98a2f6dec3"} Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.057360 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.096801 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecefedc9-a319-4ada-99e6-8e7bf2ad5da5-ssh-key\") pod \"ecefedc9-a319-4ada-99e6-8e7bf2ad5da5\" (UID: \"ecefedc9-a319-4ada-99e6-8e7bf2ad5da5\") " Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.097198 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tf4rz\" (UniqueName: \"kubernetes.io/projected/ecefedc9-a319-4ada-99e6-8e7bf2ad5da5-kube-api-access-tf4rz\") pod \"ecefedc9-a319-4ada-99e6-8e7bf2ad5da5\" (UID: \"ecefedc9-a319-4ada-99e6-8e7bf2ad5da5\") " Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.097279 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecefedc9-a319-4ada-99e6-8e7bf2ad5da5-inventory\") pod \"ecefedc9-a319-4ada-99e6-8e7bf2ad5da5\" (UID: \"ecefedc9-a319-4ada-99e6-8e7bf2ad5da5\") " Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.107853 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecefedc9-a319-4ada-99e6-8e7bf2ad5da5-kube-api-access-tf4rz" (OuterVolumeSpecName: "kube-api-access-tf4rz") pod "ecefedc9-a319-4ada-99e6-8e7bf2ad5da5" (UID: "ecefedc9-a319-4ada-99e6-8e7bf2ad5da5"). InnerVolumeSpecName "kube-api-access-tf4rz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.136587 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecefedc9-a319-4ada-99e6-8e7bf2ad5da5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ecefedc9-a319-4ada-99e6-8e7bf2ad5da5" (UID: "ecefedc9-a319-4ada-99e6-8e7bf2ad5da5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.154368 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecefedc9-a319-4ada-99e6-8e7bf2ad5da5-inventory" (OuterVolumeSpecName: "inventory") pod "ecefedc9-a319-4ada-99e6-8e7bf2ad5da5" (UID: "ecefedc9-a319-4ada-99e6-8e7bf2ad5da5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.200563 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tf4rz\" (UniqueName: \"kubernetes.io/projected/ecefedc9-a319-4ada-99e6-8e7bf2ad5da5-kube-api-access-tf4rz\") on node \"crc\" DevicePath \"\"" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.200599 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecefedc9-a319-4ada-99e6-8e7bf2ad5da5-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.200610 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecefedc9-a319-4ada-99e6-8e7bf2ad5da5-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.559491 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb" event={"ID":"ecefedc9-a319-4ada-99e6-8e7bf2ad5da5","Type":"ContainerDied","Data":"8e5f258993c14bc1ec7b0d4b364497925d8e95dbfd4205081688aa0c269fa5e9"} Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.559534 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e5f258993c14bc1ec7b0d4b364497925d8e95dbfd4205081688aa0c269fa5e9" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.559568 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.671603 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8"] Nov 26 00:57:28 crc kubenswrapper[4766]: E1126 00:57:28.672191 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecefedc9-a319-4ada-99e6-8e7bf2ad5da5" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.672222 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecefedc9-a319-4ada-99e6-8e7bf2ad5da5" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.672561 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecefedc9-a319-4ada-99e6-8e7bf2ad5da5" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.673398 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.676378 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.678161 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.679279 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.694698 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.714009 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3e66227e-c68b-47cd-97e6-91a7e447ea81-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8\" (UID: \"3e66227e-c68b-47cd-97e6-91a7e447ea81\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.714779 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh7k4\" (UniqueName: \"kubernetes.io/projected/3e66227e-c68b-47cd-97e6-91a7e447ea81-kube-api-access-lh7k4\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8\" (UID: \"3e66227e-c68b-47cd-97e6-91a7e447ea81\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.714982 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3e66227e-c68b-47cd-97e6-91a7e447ea81-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8\" (UID: \"3e66227e-c68b-47cd-97e6-91a7e447ea81\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.719066 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8"] Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.818522 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3e66227e-c68b-47cd-97e6-91a7e447ea81-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8\" (UID: \"3e66227e-c68b-47cd-97e6-91a7e447ea81\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.818617 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3e66227e-c68b-47cd-97e6-91a7e447ea81-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8\" (UID: \"3e66227e-c68b-47cd-97e6-91a7e447ea81\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.818775 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh7k4\" (UniqueName: \"kubernetes.io/projected/3e66227e-c68b-47cd-97e6-91a7e447ea81-kube-api-access-lh7k4\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8\" (UID: \"3e66227e-c68b-47cd-97e6-91a7e447ea81\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.824013 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3e66227e-c68b-47cd-97e6-91a7e447ea81-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8\" (UID: \"3e66227e-c68b-47cd-97e6-91a7e447ea81\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.826178 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3e66227e-c68b-47cd-97e6-91a7e447ea81-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8\" (UID: \"3e66227e-c68b-47cd-97e6-91a7e447ea81\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8" Nov 26 00:57:28 crc kubenswrapper[4766]: I1126 00:57:28.835554 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh7k4\" (UniqueName: \"kubernetes.io/projected/3e66227e-c68b-47cd-97e6-91a7e447ea81-kube-api-access-lh7k4\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8\" (UID: \"3e66227e-c68b-47cd-97e6-91a7e447ea81\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8" Nov 26 00:57:29 crc kubenswrapper[4766]: I1126 00:57:29.002534 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8" Nov 26 00:57:29 crc kubenswrapper[4766]: I1126 00:57:29.748578 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8"] Nov 26 00:57:29 crc kubenswrapper[4766]: W1126 00:57:29.754992 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e66227e_c68b_47cd_97e6_91a7e447ea81.slice/crio-6e0622ac66c13bf713e5e473d94b20426b605f99c9153cec24e659fdb07ef4be WatchSource:0}: Error finding container 6e0622ac66c13bf713e5e473d94b20426b605f99c9153cec24e659fdb07ef4be: Status 404 returned error can't find the container with id 6e0622ac66c13bf713e5e473d94b20426b605f99c9153cec24e659fdb07ef4be Nov 26 00:57:30 crc kubenswrapper[4766]: I1126 00:57:30.586630 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8" event={"ID":"3e66227e-c68b-47cd-97e6-91a7e447ea81","Type":"ContainerStarted","Data":"c95a0eaf6920e52eacf4489f1067b99e68b0b7651c58fd717451749e094d814e"} Nov 26 00:57:30 crc kubenswrapper[4766]: I1126 00:57:30.587235 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8" event={"ID":"3e66227e-c68b-47cd-97e6-91a7e447ea81","Type":"ContainerStarted","Data":"6e0622ac66c13bf713e5e473d94b20426b605f99c9153cec24e659fdb07ef4be"} Nov 26 00:57:30 crc kubenswrapper[4766]: I1126 00:57:30.610555 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8" podStartSLOduration=2.18632883 podStartE2EDuration="2.610529557s" podCreationTimestamp="2025-11-26 00:57:28 +0000 UTC" firstStartedPulling="2025-11-26 00:57:29.761929147 +0000 UTC m=+2030.610699607" lastFinishedPulling="2025-11-26 00:57:30.186129894 +0000 UTC m=+2031.034900334" observedRunningTime="2025-11-26 00:57:30.608667892 +0000 UTC m=+2031.457438352" watchObservedRunningTime="2025-11-26 00:57:30.610529557 +0000 UTC m=+2031.459300017" Nov 26 00:57:35 crc kubenswrapper[4766]: I1126 00:57:35.656712 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8" event={"ID":"3e66227e-c68b-47cd-97e6-91a7e447ea81","Type":"ContainerDied","Data":"c95a0eaf6920e52eacf4489f1067b99e68b0b7651c58fd717451749e094d814e"} Nov 26 00:57:35 crc kubenswrapper[4766]: I1126 00:57:35.656713 4766 generic.go:334] "Generic (PLEG): container finished" podID="3e66227e-c68b-47cd-97e6-91a7e447ea81" containerID="c95a0eaf6920e52eacf4489f1067b99e68b0b7651c58fd717451749e094d814e" exitCode=0 Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.206077 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.324656 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh7k4\" (UniqueName: \"kubernetes.io/projected/3e66227e-c68b-47cd-97e6-91a7e447ea81-kube-api-access-lh7k4\") pod \"3e66227e-c68b-47cd-97e6-91a7e447ea81\" (UID: \"3e66227e-c68b-47cd-97e6-91a7e447ea81\") " Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.325152 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3e66227e-c68b-47cd-97e6-91a7e447ea81-ssh-key\") pod \"3e66227e-c68b-47cd-97e6-91a7e447ea81\" (UID: \"3e66227e-c68b-47cd-97e6-91a7e447ea81\") " Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.325783 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3e66227e-c68b-47cd-97e6-91a7e447ea81-inventory\") pod \"3e66227e-c68b-47cd-97e6-91a7e447ea81\" (UID: \"3e66227e-c68b-47cd-97e6-91a7e447ea81\") " Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.333427 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e66227e-c68b-47cd-97e6-91a7e447ea81-kube-api-access-lh7k4" (OuterVolumeSpecName: "kube-api-access-lh7k4") pod "3e66227e-c68b-47cd-97e6-91a7e447ea81" (UID: "3e66227e-c68b-47cd-97e6-91a7e447ea81"). InnerVolumeSpecName "kube-api-access-lh7k4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.367423 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e66227e-c68b-47cd-97e6-91a7e447ea81-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3e66227e-c68b-47cd-97e6-91a7e447ea81" (UID: "3e66227e-c68b-47cd-97e6-91a7e447ea81"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.375419 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e66227e-c68b-47cd-97e6-91a7e447ea81-inventory" (OuterVolumeSpecName: "inventory") pod "3e66227e-c68b-47cd-97e6-91a7e447ea81" (UID: "3e66227e-c68b-47cd-97e6-91a7e447ea81"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.428298 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3e66227e-c68b-47cd-97e6-91a7e447ea81-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.428339 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3e66227e-c68b-47cd-97e6-91a7e447ea81-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.428353 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh7k4\" (UniqueName: \"kubernetes.io/projected/3e66227e-c68b-47cd-97e6-91a7e447ea81-kube-api-access-lh7k4\") on node \"crc\" DevicePath \"\"" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.684535 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8" event={"ID":"3e66227e-c68b-47cd-97e6-91a7e447ea81","Type":"ContainerDied","Data":"6e0622ac66c13bf713e5e473d94b20426b605f99c9153cec24e659fdb07ef4be"} Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.684595 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e0622ac66c13bf713e5e473d94b20426b605f99c9153cec24e659fdb07ef4be" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.684630 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.778821 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p"] Nov 26 00:57:37 crc kubenswrapper[4766]: E1126 00:57:37.779361 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e66227e-c68b-47cd-97e6-91a7e447ea81" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.779382 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e66227e-c68b-47cd-97e6-91a7e447ea81" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.779707 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e66227e-c68b-47cd-97e6-91a7e447ea81" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.780643 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.782986 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.783600 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.783995 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.785104 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.788972 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p"] Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.836972 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhzzx\" (UniqueName: \"kubernetes.io/projected/0ba3a09a-d108-4c92-b3db-4f531ba53427-kube-api-access-fhzzx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-57p4p\" (UID: \"0ba3a09a-d108-4c92-b3db-4f531ba53427\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.837052 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ba3a09a-d108-4c92-b3db-4f531ba53427-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-57p4p\" (UID: \"0ba3a09a-d108-4c92-b3db-4f531ba53427\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.837607 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ba3a09a-d108-4c92-b3db-4f531ba53427-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-57p4p\" (UID: \"0ba3a09a-d108-4c92-b3db-4f531ba53427\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.939567 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhzzx\" (UniqueName: \"kubernetes.io/projected/0ba3a09a-d108-4c92-b3db-4f531ba53427-kube-api-access-fhzzx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-57p4p\" (UID: \"0ba3a09a-d108-4c92-b3db-4f531ba53427\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.939986 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ba3a09a-d108-4c92-b3db-4f531ba53427-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-57p4p\" (UID: \"0ba3a09a-d108-4c92-b3db-4f531ba53427\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.940136 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ba3a09a-d108-4c92-b3db-4f531ba53427-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-57p4p\" (UID: \"0ba3a09a-d108-4c92-b3db-4f531ba53427\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.945910 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ba3a09a-d108-4c92-b3db-4f531ba53427-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-57p4p\" (UID: \"0ba3a09a-d108-4c92-b3db-4f531ba53427\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.952298 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ba3a09a-d108-4c92-b3db-4f531ba53427-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-57p4p\" (UID: \"0ba3a09a-d108-4c92-b3db-4f531ba53427\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p" Nov 26 00:57:37 crc kubenswrapper[4766]: I1126 00:57:37.959162 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhzzx\" (UniqueName: \"kubernetes.io/projected/0ba3a09a-d108-4c92-b3db-4f531ba53427-kube-api-access-fhzzx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-57p4p\" (UID: \"0ba3a09a-d108-4c92-b3db-4f531ba53427\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p" Nov 26 00:57:38 crc kubenswrapper[4766]: I1126 00:57:38.116322 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p" Nov 26 00:57:38 crc kubenswrapper[4766]: I1126 00:57:38.712168 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p"] Nov 26 00:57:38 crc kubenswrapper[4766]: W1126 00:57:38.722177 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ba3a09a_d108_4c92_b3db_4f531ba53427.slice/crio-6de86c53a67e9919a10e9e9cd61a7d60db86c5563b26ef9919c049572f033f6a WatchSource:0}: Error finding container 6de86c53a67e9919a10e9e9cd61a7d60db86c5563b26ef9919c049572f033f6a: Status 404 returned error can't find the container with id 6de86c53a67e9919a10e9e9cd61a7d60db86c5563b26ef9919c049572f033f6a Nov 26 00:57:39 crc kubenswrapper[4766]: I1126 00:57:39.703257 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p" event={"ID":"0ba3a09a-d108-4c92-b3db-4f531ba53427","Type":"ContainerStarted","Data":"22db7d2b98931ff64bbf2313a66c843a87160a766dbd40e16b52adf6f125f17c"} Nov 26 00:57:39 crc kubenswrapper[4766]: I1126 00:57:39.703871 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p" event={"ID":"0ba3a09a-d108-4c92-b3db-4f531ba53427","Type":"ContainerStarted","Data":"6de86c53a67e9919a10e9e9cd61a7d60db86c5563b26ef9919c049572f033f6a"} Nov 26 00:57:41 crc kubenswrapper[4766]: I1126 00:57:41.479229 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 00:57:41 crc kubenswrapper[4766]: I1126 00:57:41.479617 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 00:57:41 crc kubenswrapper[4766]: I1126 00:57:41.479709 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 00:57:41 crc kubenswrapper[4766]: I1126 00:57:41.720384 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"31c5fe6cf4a7b0b0322cd91369e8b92e06d2251c1e99313b96d5eb55cc36b4e6"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 00:57:41 crc kubenswrapper[4766]: I1126 00:57:41.720444 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://31c5fe6cf4a7b0b0322cd91369e8b92e06d2251c1e99313b96d5eb55cc36b4e6" gracePeriod=600 Nov 26 00:57:42 crc kubenswrapper[4766]: I1126 00:57:42.736061 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="31c5fe6cf4a7b0b0322cd91369e8b92e06d2251c1e99313b96d5eb55cc36b4e6" exitCode=0 Nov 26 00:57:42 crc kubenswrapper[4766]: I1126 00:57:42.736107 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"31c5fe6cf4a7b0b0322cd91369e8b92e06d2251c1e99313b96d5eb55cc36b4e6"} Nov 26 00:57:42 crc kubenswrapper[4766]: I1126 00:57:42.736614 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c"} Nov 26 00:57:42 crc kubenswrapper[4766]: I1126 00:57:42.736635 4766 scope.go:117] "RemoveContainer" containerID="bd39ca00a17f3750da4f2e2b18a34c688a52888f3b4be45ab5fea86b98bc94cc" Nov 26 00:57:42 crc kubenswrapper[4766]: I1126 00:57:42.765629 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p" podStartSLOduration=5.268421674 podStartE2EDuration="5.765602534s" podCreationTimestamp="2025-11-26 00:57:37 +0000 UTC" firstStartedPulling="2025-11-26 00:57:38.725104602 +0000 UTC m=+2039.573875022" lastFinishedPulling="2025-11-26 00:57:39.222285442 +0000 UTC m=+2040.071055882" observedRunningTime="2025-11-26 00:57:39.718844618 +0000 UTC m=+2040.567615048" watchObservedRunningTime="2025-11-26 00:57:42.765602534 +0000 UTC m=+2043.614373004" Nov 26 00:57:58 crc kubenswrapper[4766]: I1126 00:57:58.552164 4766 scope.go:117] "RemoveContainer" containerID="9846fffec4a9f47094a0f5e4313eea57458477b1bb7d993864e3aab65ad4b893" Nov 26 00:58:41 crc kubenswrapper[4766]: I1126 00:58:41.440921 4766 generic.go:334] "Generic (PLEG): container finished" podID="0ba3a09a-d108-4c92-b3db-4f531ba53427" containerID="22db7d2b98931ff64bbf2313a66c843a87160a766dbd40e16b52adf6f125f17c" exitCode=0 Nov 26 00:58:41 crc kubenswrapper[4766]: I1126 00:58:41.441064 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p" event={"ID":"0ba3a09a-d108-4c92-b3db-4f531ba53427","Type":"ContainerDied","Data":"22db7d2b98931ff64bbf2313a66c843a87160a766dbd40e16b52adf6f125f17c"} Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.042764 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.133810 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhzzx\" (UniqueName: \"kubernetes.io/projected/0ba3a09a-d108-4c92-b3db-4f531ba53427-kube-api-access-fhzzx\") pod \"0ba3a09a-d108-4c92-b3db-4f531ba53427\" (UID: \"0ba3a09a-d108-4c92-b3db-4f531ba53427\") " Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.133978 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ba3a09a-d108-4c92-b3db-4f531ba53427-ssh-key\") pod \"0ba3a09a-d108-4c92-b3db-4f531ba53427\" (UID: \"0ba3a09a-d108-4c92-b3db-4f531ba53427\") " Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.134016 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ba3a09a-d108-4c92-b3db-4f531ba53427-inventory\") pod \"0ba3a09a-d108-4c92-b3db-4f531ba53427\" (UID: \"0ba3a09a-d108-4c92-b3db-4f531ba53427\") " Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.141282 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ba3a09a-d108-4c92-b3db-4f531ba53427-kube-api-access-fhzzx" (OuterVolumeSpecName: "kube-api-access-fhzzx") pod "0ba3a09a-d108-4c92-b3db-4f531ba53427" (UID: "0ba3a09a-d108-4c92-b3db-4f531ba53427"). InnerVolumeSpecName "kube-api-access-fhzzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.173587 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ba3a09a-d108-4c92-b3db-4f531ba53427-inventory" (OuterVolumeSpecName: "inventory") pod "0ba3a09a-d108-4c92-b3db-4f531ba53427" (UID: "0ba3a09a-d108-4c92-b3db-4f531ba53427"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.177979 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ba3a09a-d108-4c92-b3db-4f531ba53427-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0ba3a09a-d108-4c92-b3db-4f531ba53427" (UID: "0ba3a09a-d108-4c92-b3db-4f531ba53427"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.236670 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhzzx\" (UniqueName: \"kubernetes.io/projected/0ba3a09a-d108-4c92-b3db-4f531ba53427-kube-api-access-fhzzx\") on node \"crc\" DevicePath \"\"" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.236785 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ba3a09a-d108-4c92-b3db-4f531ba53427-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.236858 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ba3a09a-d108-4c92-b3db-4f531ba53427-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.474187 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p" event={"ID":"0ba3a09a-d108-4c92-b3db-4f531ba53427","Type":"ContainerDied","Data":"6de86c53a67e9919a10e9e9cd61a7d60db86c5563b26ef9919c049572f033f6a"} Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.474629 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6de86c53a67e9919a10e9e9cd61a7d60db86c5563b26ef9919c049572f033f6a" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.474324 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.575504 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cmb77"] Nov 26 00:58:43 crc kubenswrapper[4766]: E1126 00:58:43.576140 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ba3a09a-d108-4c92-b3db-4f531ba53427" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.576170 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ba3a09a-d108-4c92-b3db-4f531ba53427" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.576522 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ba3a09a-d108-4c92-b3db-4f531ba53427" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.577722 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cmb77" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.581355 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.582840 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.583023 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.593603 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.603369 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cmb77"] Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.747296 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2e890aa1-53db-4e49-a3d2-f2ba570bfd7e-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-cmb77\" (UID: \"2e890aa1-53db-4e49-a3d2-f2ba570bfd7e\") " pod="openstack/ssh-known-hosts-edpm-deployment-cmb77" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.747370 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvjbs\" (UniqueName: \"kubernetes.io/projected/2e890aa1-53db-4e49-a3d2-f2ba570bfd7e-kube-api-access-qvjbs\") pod \"ssh-known-hosts-edpm-deployment-cmb77\" (UID: \"2e890aa1-53db-4e49-a3d2-f2ba570bfd7e\") " pod="openstack/ssh-known-hosts-edpm-deployment-cmb77" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.747476 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e890aa1-53db-4e49-a3d2-f2ba570bfd7e-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-cmb77\" (UID: \"2e890aa1-53db-4e49-a3d2-f2ba570bfd7e\") " pod="openstack/ssh-known-hosts-edpm-deployment-cmb77" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.851704 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2e890aa1-53db-4e49-a3d2-f2ba570bfd7e-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-cmb77\" (UID: \"2e890aa1-53db-4e49-a3d2-f2ba570bfd7e\") " pod="openstack/ssh-known-hosts-edpm-deployment-cmb77" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.851774 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvjbs\" (UniqueName: \"kubernetes.io/projected/2e890aa1-53db-4e49-a3d2-f2ba570bfd7e-kube-api-access-qvjbs\") pod \"ssh-known-hosts-edpm-deployment-cmb77\" (UID: \"2e890aa1-53db-4e49-a3d2-f2ba570bfd7e\") " pod="openstack/ssh-known-hosts-edpm-deployment-cmb77" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.851886 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e890aa1-53db-4e49-a3d2-f2ba570bfd7e-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-cmb77\" (UID: \"2e890aa1-53db-4e49-a3d2-f2ba570bfd7e\") " pod="openstack/ssh-known-hosts-edpm-deployment-cmb77" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.862853 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2e890aa1-53db-4e49-a3d2-f2ba570bfd7e-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-cmb77\" (UID: \"2e890aa1-53db-4e49-a3d2-f2ba570bfd7e\") " pod="openstack/ssh-known-hosts-edpm-deployment-cmb77" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.866788 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e890aa1-53db-4e49-a3d2-f2ba570bfd7e-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-cmb77\" (UID: \"2e890aa1-53db-4e49-a3d2-f2ba570bfd7e\") " pod="openstack/ssh-known-hosts-edpm-deployment-cmb77" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.881106 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvjbs\" (UniqueName: \"kubernetes.io/projected/2e890aa1-53db-4e49-a3d2-f2ba570bfd7e-kube-api-access-qvjbs\") pod \"ssh-known-hosts-edpm-deployment-cmb77\" (UID: \"2e890aa1-53db-4e49-a3d2-f2ba570bfd7e\") " pod="openstack/ssh-known-hosts-edpm-deployment-cmb77" Nov 26 00:58:43 crc kubenswrapper[4766]: I1126 00:58:43.917174 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cmb77" Nov 26 00:58:44 crc kubenswrapper[4766]: I1126 00:58:44.508613 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cmb77"] Nov 26 00:58:45 crc kubenswrapper[4766]: I1126 00:58:45.511518 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cmb77" event={"ID":"2e890aa1-53db-4e49-a3d2-f2ba570bfd7e","Type":"ContainerStarted","Data":"c898f90f982326804573682ce3aa916a45b1ec550c2211d44c67b313df2c79e0"} Nov 26 00:58:45 crc kubenswrapper[4766]: I1126 00:58:45.511864 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cmb77" event={"ID":"2e890aa1-53db-4e49-a3d2-f2ba570bfd7e","Type":"ContainerStarted","Data":"93da16ae605f5f4e99a4bed8a6092c20834e0718816128dcaf28cac0327ca1c5"} Nov 26 00:58:45 crc kubenswrapper[4766]: I1126 00:58:45.551202 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-cmb77" podStartSLOduration=2.069588698 podStartE2EDuration="2.551182488s" podCreationTimestamp="2025-11-26 00:58:43 +0000 UTC" firstStartedPulling="2025-11-26 00:58:44.515772686 +0000 UTC m=+2105.364543146" lastFinishedPulling="2025-11-26 00:58:44.997366466 +0000 UTC m=+2105.846136936" observedRunningTime="2025-11-26 00:58:45.532798724 +0000 UTC m=+2106.381569184" watchObservedRunningTime="2025-11-26 00:58:45.551182488 +0000 UTC m=+2106.399952918" Nov 26 00:58:53 crc kubenswrapper[4766]: I1126 00:58:53.613516 4766 generic.go:334] "Generic (PLEG): container finished" podID="2e890aa1-53db-4e49-a3d2-f2ba570bfd7e" containerID="c898f90f982326804573682ce3aa916a45b1ec550c2211d44c67b313df2c79e0" exitCode=0 Nov 26 00:58:53 crc kubenswrapper[4766]: I1126 00:58:53.613603 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cmb77" event={"ID":"2e890aa1-53db-4e49-a3d2-f2ba570bfd7e","Type":"ContainerDied","Data":"c898f90f982326804573682ce3aa916a45b1ec550c2211d44c67b313df2c79e0"} Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.140484 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cmb77" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.227823 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvjbs\" (UniqueName: \"kubernetes.io/projected/2e890aa1-53db-4e49-a3d2-f2ba570bfd7e-kube-api-access-qvjbs\") pod \"2e890aa1-53db-4e49-a3d2-f2ba570bfd7e\" (UID: \"2e890aa1-53db-4e49-a3d2-f2ba570bfd7e\") " Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.228084 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2e890aa1-53db-4e49-a3d2-f2ba570bfd7e-inventory-0\") pod \"2e890aa1-53db-4e49-a3d2-f2ba570bfd7e\" (UID: \"2e890aa1-53db-4e49-a3d2-f2ba570bfd7e\") " Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.228122 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e890aa1-53db-4e49-a3d2-f2ba570bfd7e-ssh-key-openstack-edpm-ipam\") pod \"2e890aa1-53db-4e49-a3d2-f2ba570bfd7e\" (UID: \"2e890aa1-53db-4e49-a3d2-f2ba570bfd7e\") " Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.262729 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e890aa1-53db-4e49-a3d2-f2ba570bfd7e-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2e890aa1-53db-4e49-a3d2-f2ba570bfd7e" (UID: "2e890aa1-53db-4e49-a3d2-f2ba570bfd7e"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.275839 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e890aa1-53db-4e49-a3d2-f2ba570bfd7e-kube-api-access-qvjbs" (OuterVolumeSpecName: "kube-api-access-qvjbs") pod "2e890aa1-53db-4e49-a3d2-f2ba570bfd7e" (UID: "2e890aa1-53db-4e49-a3d2-f2ba570bfd7e"). InnerVolumeSpecName "kube-api-access-qvjbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.327872 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e890aa1-53db-4e49-a3d2-f2ba570bfd7e-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "2e890aa1-53db-4e49-a3d2-f2ba570bfd7e" (UID: "2e890aa1-53db-4e49-a3d2-f2ba570bfd7e"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.331948 4766 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2e890aa1-53db-4e49-a3d2-f2ba570bfd7e-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.331979 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e890aa1-53db-4e49-a3d2-f2ba570bfd7e-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.331992 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvjbs\" (UniqueName: \"kubernetes.io/projected/2e890aa1-53db-4e49-a3d2-f2ba570bfd7e-kube-api-access-qvjbs\") on node \"crc\" DevicePath \"\"" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.637397 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cmb77" event={"ID":"2e890aa1-53db-4e49-a3d2-f2ba570bfd7e","Type":"ContainerDied","Data":"93da16ae605f5f4e99a4bed8a6092c20834e0718816128dcaf28cac0327ca1c5"} Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.637428 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93da16ae605f5f4e99a4bed8a6092c20834e0718816128dcaf28cac0327ca1c5" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.637483 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cmb77" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.722431 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp"] Nov 26 00:58:55 crc kubenswrapper[4766]: E1126 00:58:55.722844 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e890aa1-53db-4e49-a3d2-f2ba570bfd7e" containerName="ssh-known-hosts-edpm-deployment" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.722863 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e890aa1-53db-4e49-a3d2-f2ba570bfd7e" containerName="ssh-known-hosts-edpm-deployment" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.723130 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e890aa1-53db-4e49-a3d2-f2ba570bfd7e" containerName="ssh-known-hosts-edpm-deployment" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.723818 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.727343 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.728010 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.728255 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.729195 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.737572 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp"] Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.841472 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-sp6fp\" (UID: \"9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.841677 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-sp6fp\" (UID: \"9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.842159 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gt9nn\" (UniqueName: \"kubernetes.io/projected/9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2-kube-api-access-gt9nn\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-sp6fp\" (UID: \"9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.944810 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-sp6fp\" (UID: \"9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.945191 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gt9nn\" (UniqueName: \"kubernetes.io/projected/9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2-kube-api-access-gt9nn\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-sp6fp\" (UID: \"9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.945383 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-sp6fp\" (UID: \"9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.950756 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-sp6fp\" (UID: \"9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.950779 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-sp6fp\" (UID: \"9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp" Nov 26 00:58:55 crc kubenswrapper[4766]: I1126 00:58:55.964884 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gt9nn\" (UniqueName: \"kubernetes.io/projected/9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2-kube-api-access-gt9nn\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-sp6fp\" (UID: \"9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp" Nov 26 00:58:56 crc kubenswrapper[4766]: I1126 00:58:56.040973 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp" Nov 26 00:58:56 crc kubenswrapper[4766]: I1126 00:58:56.607360 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp"] Nov 26 00:58:56 crc kubenswrapper[4766]: I1126 00:58:56.650475 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp" event={"ID":"9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2","Type":"ContainerStarted","Data":"3438ccb52d47ed5d57ccbb5fcfb17f6ddb147368bd8181694821208ea5471dcb"} Nov 26 00:58:57 crc kubenswrapper[4766]: I1126 00:58:57.666320 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp" event={"ID":"9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2","Type":"ContainerStarted","Data":"bcf2656245fd5f4a1b2616954f9180c62d508dead67f1c3a7ab62a38bfbc9aa0"} Nov 26 00:58:57 crc kubenswrapper[4766]: I1126 00:58:57.688872 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp" podStartSLOduration=2.256851497 podStartE2EDuration="2.688847504s" podCreationTimestamp="2025-11-26 00:58:55 +0000 UTC" firstStartedPulling="2025-11-26 00:58:56.619313969 +0000 UTC m=+2117.468084439" lastFinishedPulling="2025-11-26 00:58:57.051309936 +0000 UTC m=+2117.900080446" observedRunningTime="2025-11-26 00:58:57.682037116 +0000 UTC m=+2118.530807536" watchObservedRunningTime="2025-11-26 00:58:57.688847504 +0000 UTC m=+2118.537617974" Nov 26 00:59:06 crc kubenswrapper[4766]: I1126 00:59:06.412216 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h5w6j"] Nov 26 00:59:06 crc kubenswrapper[4766]: I1126 00:59:06.415942 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h5w6j" Nov 26 00:59:06 crc kubenswrapper[4766]: I1126 00:59:06.440937 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h5w6j"] Nov 26 00:59:06 crc kubenswrapper[4766]: I1126 00:59:06.488702 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-576nx\" (UniqueName: \"kubernetes.io/projected/5ce0ff67-1fdd-486a-a0a3-b26b054a4414-kube-api-access-576nx\") pod \"certified-operators-h5w6j\" (UID: \"5ce0ff67-1fdd-486a-a0a3-b26b054a4414\") " pod="openshift-marketplace/certified-operators-h5w6j" Nov 26 00:59:06 crc kubenswrapper[4766]: I1126 00:59:06.488847 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ce0ff67-1fdd-486a-a0a3-b26b054a4414-utilities\") pod \"certified-operators-h5w6j\" (UID: \"5ce0ff67-1fdd-486a-a0a3-b26b054a4414\") " pod="openshift-marketplace/certified-operators-h5w6j" Nov 26 00:59:06 crc kubenswrapper[4766]: I1126 00:59:06.489134 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ce0ff67-1fdd-486a-a0a3-b26b054a4414-catalog-content\") pod \"certified-operators-h5w6j\" (UID: \"5ce0ff67-1fdd-486a-a0a3-b26b054a4414\") " pod="openshift-marketplace/certified-operators-h5w6j" Nov 26 00:59:06 crc kubenswrapper[4766]: I1126 00:59:06.591273 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-576nx\" (UniqueName: \"kubernetes.io/projected/5ce0ff67-1fdd-486a-a0a3-b26b054a4414-kube-api-access-576nx\") pod \"certified-operators-h5w6j\" (UID: \"5ce0ff67-1fdd-486a-a0a3-b26b054a4414\") " pod="openshift-marketplace/certified-operators-h5w6j" Nov 26 00:59:06 crc kubenswrapper[4766]: I1126 00:59:06.591364 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ce0ff67-1fdd-486a-a0a3-b26b054a4414-utilities\") pod \"certified-operators-h5w6j\" (UID: \"5ce0ff67-1fdd-486a-a0a3-b26b054a4414\") " pod="openshift-marketplace/certified-operators-h5w6j" Nov 26 00:59:06 crc kubenswrapper[4766]: I1126 00:59:06.591419 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ce0ff67-1fdd-486a-a0a3-b26b054a4414-catalog-content\") pod \"certified-operators-h5w6j\" (UID: \"5ce0ff67-1fdd-486a-a0a3-b26b054a4414\") " pod="openshift-marketplace/certified-operators-h5w6j" Nov 26 00:59:06 crc kubenswrapper[4766]: I1126 00:59:06.591922 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ce0ff67-1fdd-486a-a0a3-b26b054a4414-catalog-content\") pod \"certified-operators-h5w6j\" (UID: \"5ce0ff67-1fdd-486a-a0a3-b26b054a4414\") " pod="openshift-marketplace/certified-operators-h5w6j" Nov 26 00:59:06 crc kubenswrapper[4766]: I1126 00:59:06.591924 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ce0ff67-1fdd-486a-a0a3-b26b054a4414-utilities\") pod \"certified-operators-h5w6j\" (UID: \"5ce0ff67-1fdd-486a-a0a3-b26b054a4414\") " pod="openshift-marketplace/certified-operators-h5w6j" Nov 26 00:59:06 crc kubenswrapper[4766]: I1126 00:59:06.617941 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-576nx\" (UniqueName: \"kubernetes.io/projected/5ce0ff67-1fdd-486a-a0a3-b26b054a4414-kube-api-access-576nx\") pod \"certified-operators-h5w6j\" (UID: \"5ce0ff67-1fdd-486a-a0a3-b26b054a4414\") " pod="openshift-marketplace/certified-operators-h5w6j" Nov 26 00:59:06 crc kubenswrapper[4766]: I1126 00:59:06.781477 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h5w6j" Nov 26 00:59:06 crc kubenswrapper[4766]: I1126 00:59:06.815606 4766 generic.go:334] "Generic (PLEG): container finished" podID="9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2" containerID="bcf2656245fd5f4a1b2616954f9180c62d508dead67f1c3a7ab62a38bfbc9aa0" exitCode=0 Nov 26 00:59:06 crc kubenswrapper[4766]: I1126 00:59:06.815663 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp" event={"ID":"9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2","Type":"ContainerDied","Data":"bcf2656245fd5f4a1b2616954f9180c62d508dead67f1c3a7ab62a38bfbc9aa0"} Nov 26 00:59:07 crc kubenswrapper[4766]: I1126 00:59:07.388672 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h5w6j"] Nov 26 00:59:07 crc kubenswrapper[4766]: I1126 00:59:07.836025 4766 generic.go:334] "Generic (PLEG): container finished" podID="5ce0ff67-1fdd-486a-a0a3-b26b054a4414" containerID="8d5145b686926b5c1abb8fc6a4fca50b7b0499a9255bcbc4c7c6a9d68cf86584" exitCode=0 Nov 26 00:59:07 crc kubenswrapper[4766]: I1126 00:59:07.848336 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5w6j" event={"ID":"5ce0ff67-1fdd-486a-a0a3-b26b054a4414","Type":"ContainerDied","Data":"8d5145b686926b5c1abb8fc6a4fca50b7b0499a9255bcbc4c7c6a9d68cf86584"} Nov 26 00:59:07 crc kubenswrapper[4766]: I1126 00:59:07.848405 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5w6j" event={"ID":"5ce0ff67-1fdd-486a-a0a3-b26b054a4414","Type":"ContainerStarted","Data":"e397c2a3dfb7821c8d80e70da8254dc51419ca36f90e38fa085242d3fa792b5f"} Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.333964 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp" Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.432266 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gt9nn\" (UniqueName: \"kubernetes.io/projected/9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2-kube-api-access-gt9nn\") pod \"9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2\" (UID: \"9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2\") " Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.432568 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2-ssh-key\") pod \"9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2\" (UID: \"9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2\") " Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.432603 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2-inventory\") pod \"9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2\" (UID: \"9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2\") " Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.438634 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2-kube-api-access-gt9nn" (OuterVolumeSpecName: "kube-api-access-gt9nn") pod "9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2" (UID: "9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2"). InnerVolumeSpecName "kube-api-access-gt9nn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.470386 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2" (UID: "9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.474419 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2-inventory" (OuterVolumeSpecName: "inventory") pod "9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2" (UID: "9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.535626 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.535668 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gt9nn\" (UniqueName: \"kubernetes.io/projected/9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2-kube-api-access-gt9nn\") on node \"crc\" DevicePath \"\"" Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.535681 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.848377 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5w6j" event={"ID":"5ce0ff67-1fdd-486a-a0a3-b26b054a4414","Type":"ContainerStarted","Data":"ee0416d171c38b97643fd814c70ce65b936145a7228a3f2cf84b39bc6dbb111a"} Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.850959 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp" event={"ID":"9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2","Type":"ContainerDied","Data":"3438ccb52d47ed5d57ccbb5fcfb17f6ddb147368bd8181694821208ea5471dcb"} Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.851020 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp" Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.850999 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3438ccb52d47ed5d57ccbb5fcfb17f6ddb147368bd8181694821208ea5471dcb" Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.932575 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr"] Nov 26 00:59:08 crc kubenswrapper[4766]: E1126 00:59:08.933183 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.933217 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.933455 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.934324 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr" Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.937520 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.937632 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.937775 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.938243 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 00:59:08 crc kubenswrapper[4766]: I1126 00:59:08.967747 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr"] Nov 26 00:59:09 crc kubenswrapper[4766]: I1126 00:59:09.046999 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1fb35a35-5b73-4790-bd2f-b90260847990-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr\" (UID: \"1fb35a35-5b73-4790-bd2f-b90260847990\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr" Nov 26 00:59:09 crc kubenswrapper[4766]: I1126 00:59:09.047098 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd2m4\" (UniqueName: \"kubernetes.io/projected/1fb35a35-5b73-4790-bd2f-b90260847990-kube-api-access-wd2m4\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr\" (UID: \"1fb35a35-5b73-4790-bd2f-b90260847990\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr" Nov 26 00:59:09 crc kubenswrapper[4766]: I1126 00:59:09.047165 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fb35a35-5b73-4790-bd2f-b90260847990-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr\" (UID: \"1fb35a35-5b73-4790-bd2f-b90260847990\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr" Nov 26 00:59:09 crc kubenswrapper[4766]: I1126 00:59:09.149569 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1fb35a35-5b73-4790-bd2f-b90260847990-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr\" (UID: \"1fb35a35-5b73-4790-bd2f-b90260847990\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr" Nov 26 00:59:09 crc kubenswrapper[4766]: I1126 00:59:09.149685 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd2m4\" (UniqueName: \"kubernetes.io/projected/1fb35a35-5b73-4790-bd2f-b90260847990-kube-api-access-wd2m4\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr\" (UID: \"1fb35a35-5b73-4790-bd2f-b90260847990\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr" Nov 26 00:59:09 crc kubenswrapper[4766]: I1126 00:59:09.149763 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fb35a35-5b73-4790-bd2f-b90260847990-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr\" (UID: \"1fb35a35-5b73-4790-bd2f-b90260847990\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr" Nov 26 00:59:09 crc kubenswrapper[4766]: I1126 00:59:09.156122 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1fb35a35-5b73-4790-bd2f-b90260847990-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr\" (UID: \"1fb35a35-5b73-4790-bd2f-b90260847990\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr" Nov 26 00:59:09 crc kubenswrapper[4766]: I1126 00:59:09.158729 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fb35a35-5b73-4790-bd2f-b90260847990-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr\" (UID: \"1fb35a35-5b73-4790-bd2f-b90260847990\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr" Nov 26 00:59:09 crc kubenswrapper[4766]: I1126 00:59:09.178404 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd2m4\" (UniqueName: \"kubernetes.io/projected/1fb35a35-5b73-4790-bd2f-b90260847990-kube-api-access-wd2m4\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr\" (UID: \"1fb35a35-5b73-4790-bd2f-b90260847990\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr" Nov 26 00:59:09 crc kubenswrapper[4766]: I1126 00:59:09.258683 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr" Nov 26 00:59:09 crc kubenswrapper[4766]: I1126 00:59:09.859193 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr"] Nov 26 00:59:10 crc kubenswrapper[4766]: I1126 00:59:10.881417 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr" event={"ID":"1fb35a35-5b73-4790-bd2f-b90260847990","Type":"ContainerStarted","Data":"6d9d72adc4a3064d64bf5e7372ee64711950dee9b1160e08bf2dca091eb825d5"} Nov 26 00:59:10 crc kubenswrapper[4766]: I1126 00:59:10.881780 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr" event={"ID":"1fb35a35-5b73-4790-bd2f-b90260847990","Type":"ContainerStarted","Data":"fa398808fdc08e35a180c0ea66ef0ff113a3929df59dcfb93a4fc715fc6be5e7"} Nov 26 00:59:10 crc kubenswrapper[4766]: I1126 00:59:10.884628 4766 generic.go:334] "Generic (PLEG): container finished" podID="5ce0ff67-1fdd-486a-a0a3-b26b054a4414" containerID="ee0416d171c38b97643fd814c70ce65b936145a7228a3f2cf84b39bc6dbb111a" exitCode=0 Nov 26 00:59:10 crc kubenswrapper[4766]: I1126 00:59:10.884694 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5w6j" event={"ID":"5ce0ff67-1fdd-486a-a0a3-b26b054a4414","Type":"ContainerDied","Data":"ee0416d171c38b97643fd814c70ce65b936145a7228a3f2cf84b39bc6dbb111a"} Nov 26 00:59:10 crc kubenswrapper[4766]: I1126 00:59:10.914176 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr" podStartSLOduration=2.4854985259999998 podStartE2EDuration="2.914154501s" podCreationTimestamp="2025-11-26 00:59:08 +0000 UTC" firstStartedPulling="2025-11-26 00:59:09.876073973 +0000 UTC m=+2130.724844403" lastFinishedPulling="2025-11-26 00:59:10.304729938 +0000 UTC m=+2131.153500378" observedRunningTime="2025-11-26 00:59:10.906140693 +0000 UTC m=+2131.754911163" watchObservedRunningTime="2025-11-26 00:59:10.914154501 +0000 UTC m=+2131.762924941" Nov 26 00:59:11 crc kubenswrapper[4766]: I1126 00:59:11.903930 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5w6j" event={"ID":"5ce0ff67-1fdd-486a-a0a3-b26b054a4414","Type":"ContainerStarted","Data":"43d19e2b8d7c6097d366980fb67035d3df21e8bf61f15e1773bab28f9a8c55be"} Nov 26 00:59:11 crc kubenswrapper[4766]: I1126 00:59:11.938265 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h5w6j" podStartSLOduration=2.486747099 podStartE2EDuration="5.938237235s" podCreationTimestamp="2025-11-26 00:59:06 +0000 UTC" firstStartedPulling="2025-11-26 00:59:07.838135378 +0000 UTC m=+2128.686905828" lastFinishedPulling="2025-11-26 00:59:11.289625534 +0000 UTC m=+2132.138395964" observedRunningTime="2025-11-26 00:59:11.930693728 +0000 UTC m=+2132.779464158" watchObservedRunningTime="2025-11-26 00:59:11.938237235 +0000 UTC m=+2132.787007705" Nov 26 00:59:13 crc kubenswrapper[4766]: I1126 00:59:13.041129 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-gk5ww"] Nov 26 00:59:13 crc kubenswrapper[4766]: I1126 00:59:13.052567 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-gk5ww"] Nov 26 00:59:13 crc kubenswrapper[4766]: I1126 00:59:13.850875 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f161def-0ff3-4f5d-8f9f-083800b5e1f2" path="/var/lib/kubelet/pods/8f161def-0ff3-4f5d-8f9f-083800b5e1f2/volumes" Nov 26 00:59:14 crc kubenswrapper[4766]: I1126 00:59:14.032207 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-c562-account-create-update-gpcbx"] Nov 26 00:59:14 crc kubenswrapper[4766]: I1126 00:59:14.042468 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-c562-account-create-update-gpcbx"] Nov 26 00:59:15 crc kubenswrapper[4766]: I1126 00:59:15.313231 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tklm4"] Nov 26 00:59:15 crc kubenswrapper[4766]: I1126 00:59:15.317638 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tklm4" Nov 26 00:59:15 crc kubenswrapper[4766]: I1126 00:59:15.342194 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tklm4"] Nov 26 00:59:15 crc kubenswrapper[4766]: I1126 00:59:15.392080 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szgq5\" (UniqueName: \"kubernetes.io/projected/0cfa7585-d116-4353-b227-7fcaed2f7101-kube-api-access-szgq5\") pod \"community-operators-tklm4\" (UID: \"0cfa7585-d116-4353-b227-7fcaed2f7101\") " pod="openshift-marketplace/community-operators-tklm4" Nov 26 00:59:15 crc kubenswrapper[4766]: I1126 00:59:15.392169 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cfa7585-d116-4353-b227-7fcaed2f7101-catalog-content\") pod \"community-operators-tklm4\" (UID: \"0cfa7585-d116-4353-b227-7fcaed2f7101\") " pod="openshift-marketplace/community-operators-tklm4" Nov 26 00:59:15 crc kubenswrapper[4766]: I1126 00:59:15.392291 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cfa7585-d116-4353-b227-7fcaed2f7101-utilities\") pod \"community-operators-tklm4\" (UID: \"0cfa7585-d116-4353-b227-7fcaed2f7101\") " pod="openshift-marketplace/community-operators-tklm4" Nov 26 00:59:15 crc kubenswrapper[4766]: I1126 00:59:15.494888 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szgq5\" (UniqueName: \"kubernetes.io/projected/0cfa7585-d116-4353-b227-7fcaed2f7101-kube-api-access-szgq5\") pod \"community-operators-tklm4\" (UID: \"0cfa7585-d116-4353-b227-7fcaed2f7101\") " pod="openshift-marketplace/community-operators-tklm4" Nov 26 00:59:15 crc kubenswrapper[4766]: I1126 00:59:15.494970 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cfa7585-d116-4353-b227-7fcaed2f7101-catalog-content\") pod \"community-operators-tklm4\" (UID: \"0cfa7585-d116-4353-b227-7fcaed2f7101\") " pod="openshift-marketplace/community-operators-tklm4" Nov 26 00:59:15 crc kubenswrapper[4766]: I1126 00:59:15.495034 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cfa7585-d116-4353-b227-7fcaed2f7101-utilities\") pod \"community-operators-tklm4\" (UID: \"0cfa7585-d116-4353-b227-7fcaed2f7101\") " pod="openshift-marketplace/community-operators-tklm4" Nov 26 00:59:15 crc kubenswrapper[4766]: I1126 00:59:15.495499 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cfa7585-d116-4353-b227-7fcaed2f7101-utilities\") pod \"community-operators-tklm4\" (UID: \"0cfa7585-d116-4353-b227-7fcaed2f7101\") " pod="openshift-marketplace/community-operators-tklm4" Nov 26 00:59:15 crc kubenswrapper[4766]: I1126 00:59:15.495558 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cfa7585-d116-4353-b227-7fcaed2f7101-catalog-content\") pod \"community-operators-tklm4\" (UID: \"0cfa7585-d116-4353-b227-7fcaed2f7101\") " pod="openshift-marketplace/community-operators-tklm4" Nov 26 00:59:15 crc kubenswrapper[4766]: I1126 00:59:15.527212 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szgq5\" (UniqueName: \"kubernetes.io/projected/0cfa7585-d116-4353-b227-7fcaed2f7101-kube-api-access-szgq5\") pod \"community-operators-tklm4\" (UID: \"0cfa7585-d116-4353-b227-7fcaed2f7101\") " pod="openshift-marketplace/community-operators-tklm4" Nov 26 00:59:15 crc kubenswrapper[4766]: I1126 00:59:15.652692 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tklm4" Nov 26 00:59:15 crc kubenswrapper[4766]: I1126 00:59:15.869150 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53093344-eff4-438d-8782-9d131efa8427" path="/var/lib/kubelet/pods/53093344-eff4-438d-8782-9d131efa8427/volumes" Nov 26 00:59:16 crc kubenswrapper[4766]: I1126 00:59:16.206571 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tklm4"] Nov 26 00:59:16 crc kubenswrapper[4766]: I1126 00:59:16.782537 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h5w6j" Nov 26 00:59:16 crc kubenswrapper[4766]: I1126 00:59:16.783006 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h5w6j" Nov 26 00:59:16 crc kubenswrapper[4766]: I1126 00:59:16.862386 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h5w6j" Nov 26 00:59:16 crc kubenswrapper[4766]: I1126 00:59:16.971667 4766 generic.go:334] "Generic (PLEG): container finished" podID="0cfa7585-d116-4353-b227-7fcaed2f7101" containerID="495f7f2c82fc742b51920d593dfa8d2d2510fe1ce565f67fff394508f44efb76" exitCode=0 Nov 26 00:59:16 crc kubenswrapper[4766]: I1126 00:59:16.972856 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tklm4" event={"ID":"0cfa7585-d116-4353-b227-7fcaed2f7101","Type":"ContainerDied","Data":"495f7f2c82fc742b51920d593dfa8d2d2510fe1ce565f67fff394508f44efb76"} Nov 26 00:59:16 crc kubenswrapper[4766]: I1126 00:59:16.972892 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tklm4" event={"ID":"0cfa7585-d116-4353-b227-7fcaed2f7101","Type":"ContainerStarted","Data":"d00cf5d7f1ce64744e85561362ab89c87b96892459a3f185afd8a9898ba5bcf2"} Nov 26 00:59:17 crc kubenswrapper[4766]: I1126 00:59:17.028605 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h5w6j" Nov 26 00:59:17 crc kubenswrapper[4766]: I1126 00:59:17.984422 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tklm4" event={"ID":"0cfa7585-d116-4353-b227-7fcaed2f7101","Type":"ContainerStarted","Data":"96d38e86af53adfa4e4b26b0b8fbe183559ed4d349f20e9d3402c866fdd9f4b9"} Nov 26 00:59:19 crc kubenswrapper[4766]: I1126 00:59:19.288408 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h5w6j"] Nov 26 00:59:19 crc kubenswrapper[4766]: I1126 00:59:19.289030 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-h5w6j" podUID="5ce0ff67-1fdd-486a-a0a3-b26b054a4414" containerName="registry-server" containerID="cri-o://43d19e2b8d7c6097d366980fb67035d3df21e8bf61f15e1773bab28f9a8c55be" gracePeriod=2 Nov 26 00:59:19 crc kubenswrapper[4766]: I1126 00:59:19.879149 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h5w6j" Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.000881 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ce0ff67-1fdd-486a-a0a3-b26b054a4414-catalog-content\") pod \"5ce0ff67-1fdd-486a-a0a3-b26b054a4414\" (UID: \"5ce0ff67-1fdd-486a-a0a3-b26b054a4414\") " Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.001106 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-576nx\" (UniqueName: \"kubernetes.io/projected/5ce0ff67-1fdd-486a-a0a3-b26b054a4414-kube-api-access-576nx\") pod \"5ce0ff67-1fdd-486a-a0a3-b26b054a4414\" (UID: \"5ce0ff67-1fdd-486a-a0a3-b26b054a4414\") " Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.001156 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ce0ff67-1fdd-486a-a0a3-b26b054a4414-utilities\") pod \"5ce0ff67-1fdd-486a-a0a3-b26b054a4414\" (UID: \"5ce0ff67-1fdd-486a-a0a3-b26b054a4414\") " Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.001909 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ce0ff67-1fdd-486a-a0a3-b26b054a4414-utilities" (OuterVolumeSpecName: "utilities") pod "5ce0ff67-1fdd-486a-a0a3-b26b054a4414" (UID: "5ce0ff67-1fdd-486a-a0a3-b26b054a4414"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.010632 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ce0ff67-1fdd-486a-a0a3-b26b054a4414-kube-api-access-576nx" (OuterVolumeSpecName: "kube-api-access-576nx") pod "5ce0ff67-1fdd-486a-a0a3-b26b054a4414" (UID: "5ce0ff67-1fdd-486a-a0a3-b26b054a4414"). InnerVolumeSpecName "kube-api-access-576nx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.016511 4766 generic.go:334] "Generic (PLEG): container finished" podID="5ce0ff67-1fdd-486a-a0a3-b26b054a4414" containerID="43d19e2b8d7c6097d366980fb67035d3df21e8bf61f15e1773bab28f9a8c55be" exitCode=0 Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.016575 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h5w6j" Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.016576 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5w6j" event={"ID":"5ce0ff67-1fdd-486a-a0a3-b26b054a4414","Type":"ContainerDied","Data":"43d19e2b8d7c6097d366980fb67035d3df21e8bf61f15e1773bab28f9a8c55be"} Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.016670 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5w6j" event={"ID":"5ce0ff67-1fdd-486a-a0a3-b26b054a4414","Type":"ContainerDied","Data":"e397c2a3dfb7821c8d80e70da8254dc51419ca36f90e38fa085242d3fa792b5f"} Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.016696 4766 scope.go:117] "RemoveContainer" containerID="43d19e2b8d7c6097d366980fb67035d3df21e8bf61f15e1773bab28f9a8c55be" Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.018277 4766 generic.go:334] "Generic (PLEG): container finished" podID="0cfa7585-d116-4353-b227-7fcaed2f7101" containerID="96d38e86af53adfa4e4b26b0b8fbe183559ed4d349f20e9d3402c866fdd9f4b9" exitCode=0 Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.018315 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tklm4" event={"ID":"0cfa7585-d116-4353-b227-7fcaed2f7101","Type":"ContainerDied","Data":"96d38e86af53adfa4e4b26b0b8fbe183559ed4d349f20e9d3402c866fdd9f4b9"} Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.052611 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ce0ff67-1fdd-486a-a0a3-b26b054a4414-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ce0ff67-1fdd-486a-a0a3-b26b054a4414" (UID: "5ce0ff67-1fdd-486a-a0a3-b26b054a4414"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.093731 4766 scope.go:117] "RemoveContainer" containerID="ee0416d171c38b97643fd814c70ce65b936145a7228a3f2cf84b39bc6dbb111a" Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.104117 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-576nx\" (UniqueName: \"kubernetes.io/projected/5ce0ff67-1fdd-486a-a0a3-b26b054a4414-kube-api-access-576nx\") on node \"crc\" DevicePath \"\"" Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.104180 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ce0ff67-1fdd-486a-a0a3-b26b054a4414-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.104194 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ce0ff67-1fdd-486a-a0a3-b26b054a4414-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.142359 4766 scope.go:117] "RemoveContainer" containerID="8d5145b686926b5c1abb8fc6a4fca50b7b0499a9255bcbc4c7c6a9d68cf86584" Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.181914 4766 scope.go:117] "RemoveContainer" containerID="43d19e2b8d7c6097d366980fb67035d3df21e8bf61f15e1773bab28f9a8c55be" Nov 26 00:59:20 crc kubenswrapper[4766]: E1126 00:59:20.182337 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43d19e2b8d7c6097d366980fb67035d3df21e8bf61f15e1773bab28f9a8c55be\": container with ID starting with 43d19e2b8d7c6097d366980fb67035d3df21e8bf61f15e1773bab28f9a8c55be not found: ID does not exist" containerID="43d19e2b8d7c6097d366980fb67035d3df21e8bf61f15e1773bab28f9a8c55be" Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.182370 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43d19e2b8d7c6097d366980fb67035d3df21e8bf61f15e1773bab28f9a8c55be"} err="failed to get container status \"43d19e2b8d7c6097d366980fb67035d3df21e8bf61f15e1773bab28f9a8c55be\": rpc error: code = NotFound desc = could not find container \"43d19e2b8d7c6097d366980fb67035d3df21e8bf61f15e1773bab28f9a8c55be\": container with ID starting with 43d19e2b8d7c6097d366980fb67035d3df21e8bf61f15e1773bab28f9a8c55be not found: ID does not exist" Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.182389 4766 scope.go:117] "RemoveContainer" containerID="ee0416d171c38b97643fd814c70ce65b936145a7228a3f2cf84b39bc6dbb111a" Nov 26 00:59:20 crc kubenswrapper[4766]: E1126 00:59:20.182796 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee0416d171c38b97643fd814c70ce65b936145a7228a3f2cf84b39bc6dbb111a\": container with ID starting with ee0416d171c38b97643fd814c70ce65b936145a7228a3f2cf84b39bc6dbb111a not found: ID does not exist" containerID="ee0416d171c38b97643fd814c70ce65b936145a7228a3f2cf84b39bc6dbb111a" Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.182834 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee0416d171c38b97643fd814c70ce65b936145a7228a3f2cf84b39bc6dbb111a"} err="failed to get container status \"ee0416d171c38b97643fd814c70ce65b936145a7228a3f2cf84b39bc6dbb111a\": rpc error: code = NotFound desc = could not find container \"ee0416d171c38b97643fd814c70ce65b936145a7228a3f2cf84b39bc6dbb111a\": container with ID starting with ee0416d171c38b97643fd814c70ce65b936145a7228a3f2cf84b39bc6dbb111a not found: ID does not exist" Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.182860 4766 scope.go:117] "RemoveContainer" containerID="8d5145b686926b5c1abb8fc6a4fca50b7b0499a9255bcbc4c7c6a9d68cf86584" Nov 26 00:59:20 crc kubenswrapper[4766]: E1126 00:59:20.183208 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d5145b686926b5c1abb8fc6a4fca50b7b0499a9255bcbc4c7c6a9d68cf86584\": container with ID starting with 8d5145b686926b5c1abb8fc6a4fca50b7b0499a9255bcbc4c7c6a9d68cf86584 not found: ID does not exist" containerID="8d5145b686926b5c1abb8fc6a4fca50b7b0499a9255bcbc4c7c6a9d68cf86584" Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.183270 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d5145b686926b5c1abb8fc6a4fca50b7b0499a9255bcbc4c7c6a9d68cf86584"} err="failed to get container status \"8d5145b686926b5c1abb8fc6a4fca50b7b0499a9255bcbc4c7c6a9d68cf86584\": rpc error: code = NotFound desc = could not find container \"8d5145b686926b5c1abb8fc6a4fca50b7b0499a9255bcbc4c7c6a9d68cf86584\": container with ID starting with 8d5145b686926b5c1abb8fc6a4fca50b7b0499a9255bcbc4c7c6a9d68cf86584 not found: ID does not exist" Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.361572 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h5w6j"] Nov 26 00:59:20 crc kubenswrapper[4766]: I1126 00:59:20.370284 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-h5w6j"] Nov 26 00:59:21 crc kubenswrapper[4766]: I1126 00:59:21.033757 4766 generic.go:334] "Generic (PLEG): container finished" podID="1fb35a35-5b73-4790-bd2f-b90260847990" containerID="6d9d72adc4a3064d64bf5e7372ee64711950dee9b1160e08bf2dca091eb825d5" exitCode=0 Nov 26 00:59:21 crc kubenswrapper[4766]: I1126 00:59:21.033838 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr" event={"ID":"1fb35a35-5b73-4790-bd2f-b90260847990","Type":"ContainerDied","Data":"6d9d72adc4a3064d64bf5e7372ee64711950dee9b1160e08bf2dca091eb825d5"} Nov 26 00:59:21 crc kubenswrapper[4766]: I1126 00:59:21.040376 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tklm4" event={"ID":"0cfa7585-d116-4353-b227-7fcaed2f7101","Type":"ContainerStarted","Data":"c31dd6edcd94ab93e554b2fa7c70d35cd74e803b3020cfa6f51168e1cf3510df"} Nov 26 00:59:21 crc kubenswrapper[4766]: I1126 00:59:21.110555 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tklm4" podStartSLOduration=2.623802823 podStartE2EDuration="6.110528717s" podCreationTimestamp="2025-11-26 00:59:15 +0000 UTC" firstStartedPulling="2025-11-26 00:59:16.974192648 +0000 UTC m=+2137.822963088" lastFinishedPulling="2025-11-26 00:59:20.460918512 +0000 UTC m=+2141.309688982" observedRunningTime="2025-11-26 00:59:21.078386214 +0000 UTC m=+2141.927156684" watchObservedRunningTime="2025-11-26 00:59:21.110528717 +0000 UTC m=+2141.959299177" Nov 26 00:59:21 crc kubenswrapper[4766]: I1126 00:59:21.849799 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ce0ff67-1fdd-486a-a0a3-b26b054a4414" path="/var/lib/kubelet/pods/5ce0ff67-1fdd-486a-a0a3-b26b054a4414/volumes" Nov 26 00:59:22 crc kubenswrapper[4766]: I1126 00:59:22.604395 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr" Nov 26 00:59:22 crc kubenswrapper[4766]: I1126 00:59:22.758699 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wd2m4\" (UniqueName: \"kubernetes.io/projected/1fb35a35-5b73-4790-bd2f-b90260847990-kube-api-access-wd2m4\") pod \"1fb35a35-5b73-4790-bd2f-b90260847990\" (UID: \"1fb35a35-5b73-4790-bd2f-b90260847990\") " Nov 26 00:59:22 crc kubenswrapper[4766]: I1126 00:59:22.759320 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1fb35a35-5b73-4790-bd2f-b90260847990-ssh-key\") pod \"1fb35a35-5b73-4790-bd2f-b90260847990\" (UID: \"1fb35a35-5b73-4790-bd2f-b90260847990\") " Nov 26 00:59:22 crc kubenswrapper[4766]: I1126 00:59:22.759361 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fb35a35-5b73-4790-bd2f-b90260847990-inventory\") pod \"1fb35a35-5b73-4790-bd2f-b90260847990\" (UID: \"1fb35a35-5b73-4790-bd2f-b90260847990\") " Nov 26 00:59:22 crc kubenswrapper[4766]: I1126 00:59:22.766366 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fb35a35-5b73-4790-bd2f-b90260847990-kube-api-access-wd2m4" (OuterVolumeSpecName: "kube-api-access-wd2m4") pod "1fb35a35-5b73-4790-bd2f-b90260847990" (UID: "1fb35a35-5b73-4790-bd2f-b90260847990"). InnerVolumeSpecName "kube-api-access-wd2m4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:59:22 crc kubenswrapper[4766]: I1126 00:59:22.791908 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fb35a35-5b73-4790-bd2f-b90260847990-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1fb35a35-5b73-4790-bd2f-b90260847990" (UID: "1fb35a35-5b73-4790-bd2f-b90260847990"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:59:22 crc kubenswrapper[4766]: I1126 00:59:22.817588 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fb35a35-5b73-4790-bd2f-b90260847990-inventory" (OuterVolumeSpecName: "inventory") pod "1fb35a35-5b73-4790-bd2f-b90260847990" (UID: "1fb35a35-5b73-4790-bd2f-b90260847990"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 00:59:22 crc kubenswrapper[4766]: I1126 00:59:22.861902 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wd2m4\" (UniqueName: \"kubernetes.io/projected/1fb35a35-5b73-4790-bd2f-b90260847990-kube-api-access-wd2m4\") on node \"crc\" DevicePath \"\"" Nov 26 00:59:22 crc kubenswrapper[4766]: I1126 00:59:22.861934 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1fb35a35-5b73-4790-bd2f-b90260847990-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 00:59:22 crc kubenswrapper[4766]: I1126 00:59:22.861944 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fb35a35-5b73-4790-bd2f-b90260847990-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.066957 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr" event={"ID":"1fb35a35-5b73-4790-bd2f-b90260847990","Type":"ContainerDied","Data":"fa398808fdc08e35a180c0ea66ef0ff113a3929df59dcfb93a4fc715fc6be5e7"} Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.066995 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa398808fdc08e35a180c0ea66ef0ff113a3929df59dcfb93a4fc715fc6be5e7" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.067032 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.165802 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq"] Nov 26 00:59:23 crc kubenswrapper[4766]: E1126 00:59:23.166207 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ce0ff67-1fdd-486a-a0a3-b26b054a4414" containerName="extract-content" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.166225 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ce0ff67-1fdd-486a-a0a3-b26b054a4414" containerName="extract-content" Nov 26 00:59:23 crc kubenswrapper[4766]: E1126 00:59:23.166242 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ce0ff67-1fdd-486a-a0a3-b26b054a4414" containerName="registry-server" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.166248 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ce0ff67-1fdd-486a-a0a3-b26b054a4414" containerName="registry-server" Nov 26 00:59:23 crc kubenswrapper[4766]: E1126 00:59:23.166269 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ce0ff67-1fdd-486a-a0a3-b26b054a4414" containerName="extract-utilities" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.166276 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ce0ff67-1fdd-486a-a0a3-b26b054a4414" containerName="extract-utilities" Nov 26 00:59:23 crc kubenswrapper[4766]: E1126 00:59:23.166295 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fb35a35-5b73-4790-bd2f-b90260847990" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.166302 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fb35a35-5b73-4790-bd2f-b90260847990" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.166488 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ce0ff67-1fdd-486a-a0a3-b26b054a4414" containerName="registry-server" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.166519 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fb35a35-5b73-4790-bd2f-b90260847990" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.167176 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.174354 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.174558 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.176584 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.176822 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.176879 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.176987 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.177216 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.177366 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.186827 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq"] Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.268734 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.268790 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.268820 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.269012 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.269159 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g75k2\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-kube-api-access-g75k2\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.269213 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.269362 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.269427 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.269556 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.269922 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.269992 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.270066 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.270121 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.371713 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.371757 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.371783 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.371804 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.371876 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.371903 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.371922 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.371942 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.371973 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g75k2\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-kube-api-access-g75k2\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.371989 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.372025 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.372046 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.372080 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.376665 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.376726 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.377104 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.377769 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.377808 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.378368 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.378408 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.379641 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.379791 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.379957 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.381499 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.381576 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.397202 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g75k2\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-kube-api-access-g75k2\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:23 crc kubenswrapper[4766]: I1126 00:59:23.483247 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 00:59:24 crc kubenswrapper[4766]: I1126 00:59:24.140231 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq"] Nov 26 00:59:24 crc kubenswrapper[4766]: W1126 00:59:24.151167 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf264bace_32cb_4fb6_acab_40d59db68d6e.slice/crio-eeeac981735f1493788c1301ce1e4ede3bc32615bb83626f316ea2ed8aab3ac0 WatchSource:0}: Error finding container eeeac981735f1493788c1301ce1e4ede3bc32615bb83626f316ea2ed8aab3ac0: Status 404 returned error can't find the container with id eeeac981735f1493788c1301ce1e4ede3bc32615bb83626f316ea2ed8aab3ac0 Nov 26 00:59:25 crc kubenswrapper[4766]: I1126 00:59:25.097531 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" event={"ID":"f264bace-32cb-4fb6-acab-40d59db68d6e","Type":"ContainerStarted","Data":"c1e94194b85eda72901f1729372aa218cc8883221639975d9e52f2a1c7979027"} Nov 26 00:59:25 crc kubenswrapper[4766]: I1126 00:59:25.097924 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" event={"ID":"f264bace-32cb-4fb6-acab-40d59db68d6e","Type":"ContainerStarted","Data":"eeeac981735f1493788c1301ce1e4ede3bc32615bb83626f316ea2ed8aab3ac0"} Nov 26 00:59:25 crc kubenswrapper[4766]: I1126 00:59:25.652977 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tklm4" Nov 26 00:59:25 crc kubenswrapper[4766]: I1126 00:59:25.654171 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tklm4" Nov 26 00:59:25 crc kubenswrapper[4766]: I1126 00:59:25.718630 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tklm4" Nov 26 00:59:25 crc kubenswrapper[4766]: I1126 00:59:25.772361 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" podStartSLOduration=2.2819141419999998 podStartE2EDuration="2.77233927s" podCreationTimestamp="2025-11-26 00:59:23 +0000 UTC" firstStartedPulling="2025-11-26 00:59:24.155426203 +0000 UTC m=+2145.004196633" lastFinishedPulling="2025-11-26 00:59:24.645851291 +0000 UTC m=+2145.494621761" observedRunningTime="2025-11-26 00:59:25.134483874 +0000 UTC m=+2145.983254334" watchObservedRunningTime="2025-11-26 00:59:25.77233927 +0000 UTC m=+2146.621109710" Nov 26 00:59:26 crc kubenswrapper[4766]: I1126 00:59:26.169989 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tklm4" Nov 26 00:59:26 crc kubenswrapper[4766]: I1126 00:59:26.224777 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tklm4"] Nov 26 00:59:28 crc kubenswrapper[4766]: I1126 00:59:28.136461 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tklm4" podUID="0cfa7585-d116-4353-b227-7fcaed2f7101" containerName="registry-server" containerID="cri-o://c31dd6edcd94ab93e554b2fa7c70d35cd74e803b3020cfa6f51168e1cf3510df" gracePeriod=2 Nov 26 00:59:28 crc kubenswrapper[4766]: I1126 00:59:28.686385 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tklm4" Nov 26 00:59:28 crc kubenswrapper[4766]: I1126 00:59:28.802079 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cfa7585-d116-4353-b227-7fcaed2f7101-utilities\") pod \"0cfa7585-d116-4353-b227-7fcaed2f7101\" (UID: \"0cfa7585-d116-4353-b227-7fcaed2f7101\") " Nov 26 00:59:28 crc kubenswrapper[4766]: I1126 00:59:28.802143 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cfa7585-d116-4353-b227-7fcaed2f7101-catalog-content\") pod \"0cfa7585-d116-4353-b227-7fcaed2f7101\" (UID: \"0cfa7585-d116-4353-b227-7fcaed2f7101\") " Nov 26 00:59:28 crc kubenswrapper[4766]: I1126 00:59:28.802311 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szgq5\" (UniqueName: \"kubernetes.io/projected/0cfa7585-d116-4353-b227-7fcaed2f7101-kube-api-access-szgq5\") pod \"0cfa7585-d116-4353-b227-7fcaed2f7101\" (UID: \"0cfa7585-d116-4353-b227-7fcaed2f7101\") " Nov 26 00:59:28 crc kubenswrapper[4766]: I1126 00:59:28.802891 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cfa7585-d116-4353-b227-7fcaed2f7101-utilities" (OuterVolumeSpecName: "utilities") pod "0cfa7585-d116-4353-b227-7fcaed2f7101" (UID: "0cfa7585-d116-4353-b227-7fcaed2f7101"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:59:28 crc kubenswrapper[4766]: I1126 00:59:28.813470 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cfa7585-d116-4353-b227-7fcaed2f7101-kube-api-access-szgq5" (OuterVolumeSpecName: "kube-api-access-szgq5") pod "0cfa7585-d116-4353-b227-7fcaed2f7101" (UID: "0cfa7585-d116-4353-b227-7fcaed2f7101"). InnerVolumeSpecName "kube-api-access-szgq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 00:59:28 crc kubenswrapper[4766]: I1126 00:59:28.865382 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cfa7585-d116-4353-b227-7fcaed2f7101-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0cfa7585-d116-4353-b227-7fcaed2f7101" (UID: "0cfa7585-d116-4353-b227-7fcaed2f7101"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 00:59:28 crc kubenswrapper[4766]: I1126 00:59:28.905851 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cfa7585-d116-4353-b227-7fcaed2f7101-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 00:59:28 crc kubenswrapper[4766]: I1126 00:59:28.905886 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cfa7585-d116-4353-b227-7fcaed2f7101-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 00:59:28 crc kubenswrapper[4766]: I1126 00:59:28.905925 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szgq5\" (UniqueName: \"kubernetes.io/projected/0cfa7585-d116-4353-b227-7fcaed2f7101-kube-api-access-szgq5\") on node \"crc\" DevicePath \"\"" Nov 26 00:59:29 crc kubenswrapper[4766]: I1126 00:59:29.155474 4766 generic.go:334] "Generic (PLEG): container finished" podID="0cfa7585-d116-4353-b227-7fcaed2f7101" containerID="c31dd6edcd94ab93e554b2fa7c70d35cd74e803b3020cfa6f51168e1cf3510df" exitCode=0 Nov 26 00:59:29 crc kubenswrapper[4766]: I1126 00:59:29.155515 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tklm4" Nov 26 00:59:29 crc kubenswrapper[4766]: I1126 00:59:29.155523 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tklm4" event={"ID":"0cfa7585-d116-4353-b227-7fcaed2f7101","Type":"ContainerDied","Data":"c31dd6edcd94ab93e554b2fa7c70d35cd74e803b3020cfa6f51168e1cf3510df"} Nov 26 00:59:29 crc kubenswrapper[4766]: I1126 00:59:29.155554 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tklm4" event={"ID":"0cfa7585-d116-4353-b227-7fcaed2f7101","Type":"ContainerDied","Data":"d00cf5d7f1ce64744e85561362ab89c87b96892459a3f185afd8a9898ba5bcf2"} Nov 26 00:59:29 crc kubenswrapper[4766]: I1126 00:59:29.155576 4766 scope.go:117] "RemoveContainer" containerID="c31dd6edcd94ab93e554b2fa7c70d35cd74e803b3020cfa6f51168e1cf3510df" Nov 26 00:59:29 crc kubenswrapper[4766]: I1126 00:59:29.186387 4766 scope.go:117] "RemoveContainer" containerID="96d38e86af53adfa4e4b26b0b8fbe183559ed4d349f20e9d3402c866fdd9f4b9" Nov 26 00:59:29 crc kubenswrapper[4766]: I1126 00:59:29.193728 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tklm4"] Nov 26 00:59:29 crc kubenswrapper[4766]: I1126 00:59:29.209971 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tklm4"] Nov 26 00:59:29 crc kubenswrapper[4766]: I1126 00:59:29.221112 4766 scope.go:117] "RemoveContainer" containerID="495f7f2c82fc742b51920d593dfa8d2d2510fe1ce565f67fff394508f44efb76" Nov 26 00:59:29 crc kubenswrapper[4766]: I1126 00:59:29.284044 4766 scope.go:117] "RemoveContainer" containerID="c31dd6edcd94ab93e554b2fa7c70d35cd74e803b3020cfa6f51168e1cf3510df" Nov 26 00:59:29 crc kubenswrapper[4766]: E1126 00:59:29.284818 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c31dd6edcd94ab93e554b2fa7c70d35cd74e803b3020cfa6f51168e1cf3510df\": container with ID starting with c31dd6edcd94ab93e554b2fa7c70d35cd74e803b3020cfa6f51168e1cf3510df not found: ID does not exist" containerID="c31dd6edcd94ab93e554b2fa7c70d35cd74e803b3020cfa6f51168e1cf3510df" Nov 26 00:59:29 crc kubenswrapper[4766]: I1126 00:59:29.284859 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c31dd6edcd94ab93e554b2fa7c70d35cd74e803b3020cfa6f51168e1cf3510df"} err="failed to get container status \"c31dd6edcd94ab93e554b2fa7c70d35cd74e803b3020cfa6f51168e1cf3510df\": rpc error: code = NotFound desc = could not find container \"c31dd6edcd94ab93e554b2fa7c70d35cd74e803b3020cfa6f51168e1cf3510df\": container with ID starting with c31dd6edcd94ab93e554b2fa7c70d35cd74e803b3020cfa6f51168e1cf3510df not found: ID does not exist" Nov 26 00:59:29 crc kubenswrapper[4766]: I1126 00:59:29.284881 4766 scope.go:117] "RemoveContainer" containerID="96d38e86af53adfa4e4b26b0b8fbe183559ed4d349f20e9d3402c866fdd9f4b9" Nov 26 00:59:29 crc kubenswrapper[4766]: E1126 00:59:29.285198 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96d38e86af53adfa4e4b26b0b8fbe183559ed4d349f20e9d3402c866fdd9f4b9\": container with ID starting with 96d38e86af53adfa4e4b26b0b8fbe183559ed4d349f20e9d3402c866fdd9f4b9 not found: ID does not exist" containerID="96d38e86af53adfa4e4b26b0b8fbe183559ed4d349f20e9d3402c866fdd9f4b9" Nov 26 00:59:29 crc kubenswrapper[4766]: I1126 00:59:29.285216 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96d38e86af53adfa4e4b26b0b8fbe183559ed4d349f20e9d3402c866fdd9f4b9"} err="failed to get container status \"96d38e86af53adfa4e4b26b0b8fbe183559ed4d349f20e9d3402c866fdd9f4b9\": rpc error: code = NotFound desc = could not find container \"96d38e86af53adfa4e4b26b0b8fbe183559ed4d349f20e9d3402c866fdd9f4b9\": container with ID starting with 96d38e86af53adfa4e4b26b0b8fbe183559ed4d349f20e9d3402c866fdd9f4b9 not found: ID does not exist" Nov 26 00:59:29 crc kubenswrapper[4766]: I1126 00:59:29.285228 4766 scope.go:117] "RemoveContainer" containerID="495f7f2c82fc742b51920d593dfa8d2d2510fe1ce565f67fff394508f44efb76" Nov 26 00:59:29 crc kubenswrapper[4766]: E1126 00:59:29.285510 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"495f7f2c82fc742b51920d593dfa8d2d2510fe1ce565f67fff394508f44efb76\": container with ID starting with 495f7f2c82fc742b51920d593dfa8d2d2510fe1ce565f67fff394508f44efb76 not found: ID does not exist" containerID="495f7f2c82fc742b51920d593dfa8d2d2510fe1ce565f67fff394508f44efb76" Nov 26 00:59:29 crc kubenswrapper[4766]: I1126 00:59:29.285534 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"495f7f2c82fc742b51920d593dfa8d2d2510fe1ce565f67fff394508f44efb76"} err="failed to get container status \"495f7f2c82fc742b51920d593dfa8d2d2510fe1ce565f67fff394508f44efb76\": rpc error: code = NotFound desc = could not find container \"495f7f2c82fc742b51920d593dfa8d2d2510fe1ce565f67fff394508f44efb76\": container with ID starting with 495f7f2c82fc742b51920d593dfa8d2d2510fe1ce565f67fff394508f44efb76 not found: ID does not exist" Nov 26 00:59:29 crc kubenswrapper[4766]: I1126 00:59:29.853028 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cfa7585-d116-4353-b227-7fcaed2f7101" path="/var/lib/kubelet/pods/0cfa7585-d116-4353-b227-7fcaed2f7101/volumes" Nov 26 00:59:30 crc kubenswrapper[4766]: I1126 00:59:30.057122 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-vc4ps"] Nov 26 00:59:30 crc kubenswrapper[4766]: I1126 00:59:30.072858 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-vc4ps"] Nov 26 00:59:30 crc kubenswrapper[4766]: I1126 00:59:30.085156 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-wxhtm"] Nov 26 00:59:30 crc kubenswrapper[4766]: I1126 00:59:30.099127 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-wxhtm"] Nov 26 00:59:31 crc kubenswrapper[4766]: I1126 00:59:31.848423 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c" path="/var/lib/kubelet/pods/03e87e7f-6aa5-45f9-8e0e-acb7c0087f1c/volumes" Nov 26 00:59:31 crc kubenswrapper[4766]: I1126 00:59:31.849923 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12f40773-d3bf-46ac-96a0-8f114083229f" path="/var/lib/kubelet/pods/12f40773-d3bf-46ac-96a0-8f114083229f/volumes" Nov 26 00:59:54 crc kubenswrapper[4766]: I1126 00:59:54.045750 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-dn4zp"] Nov 26 00:59:54 crc kubenswrapper[4766]: I1126 00:59:54.055663 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-dn4zp"] Nov 26 00:59:55 crc kubenswrapper[4766]: I1126 00:59:55.848150 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4252a4a9-9a21-4006-8d77-1128fdedd601" path="/var/lib/kubelet/pods/4252a4a9-9a21-4006-8d77-1128fdedd601/volumes" Nov 26 00:59:58 crc kubenswrapper[4766]: I1126 00:59:58.715019 4766 scope.go:117] "RemoveContainer" containerID="451e952ceb6097241b3aa93cc5e036cd73babfd8d613c4fc3109e5138908acde" Nov 26 00:59:58 crc kubenswrapper[4766]: I1126 00:59:58.777995 4766 scope.go:117] "RemoveContainer" containerID="91430a52bc08672d7a76f2968ea638d9102ce29b7f28f477f8635795a2f46779" Nov 26 00:59:58 crc kubenswrapper[4766]: I1126 00:59:58.851993 4766 scope.go:117] "RemoveContainer" containerID="92aef5fe319c6e4705cccd3f94d4735c9ef2561999c3cbc1bcb70eb61d7d4ffd" Nov 26 00:59:58 crc kubenswrapper[4766]: I1126 00:59:58.958636 4766 scope.go:117] "RemoveContainer" containerID="01b645a0911c62a1c44fec9ad6c73b934f98107042782e07bb3bc616789e8a50" Nov 26 00:59:58 crc kubenswrapper[4766]: I1126 00:59:58.996328 4766 scope.go:117] "RemoveContainer" containerID="b1c01ee5e01c4926e7120c25299f00cb50055c6b3458163d33ae4cdbad953a9e" Nov 26 01:00:00 crc kubenswrapper[4766]: I1126 01:00:00.213223 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w"] Nov 26 01:00:00 crc kubenswrapper[4766]: E1126 01:00:00.213907 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfa7585-d116-4353-b227-7fcaed2f7101" containerName="extract-content" Nov 26 01:00:00 crc kubenswrapper[4766]: I1126 01:00:00.213926 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfa7585-d116-4353-b227-7fcaed2f7101" containerName="extract-content" Nov 26 01:00:00 crc kubenswrapper[4766]: E1126 01:00:00.213947 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfa7585-d116-4353-b227-7fcaed2f7101" containerName="extract-utilities" Nov 26 01:00:00 crc kubenswrapper[4766]: I1126 01:00:00.213956 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfa7585-d116-4353-b227-7fcaed2f7101" containerName="extract-utilities" Nov 26 01:00:00 crc kubenswrapper[4766]: E1126 01:00:00.213976 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfa7585-d116-4353-b227-7fcaed2f7101" containerName="registry-server" Nov 26 01:00:00 crc kubenswrapper[4766]: I1126 01:00:00.213991 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfa7585-d116-4353-b227-7fcaed2f7101" containerName="registry-server" Nov 26 01:00:00 crc kubenswrapper[4766]: I1126 01:00:00.214374 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cfa7585-d116-4353-b227-7fcaed2f7101" containerName="registry-server" Nov 26 01:00:00 crc kubenswrapper[4766]: I1126 01:00:00.215390 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w" Nov 26 01:00:00 crc kubenswrapper[4766]: I1126 01:00:00.220715 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 26 01:00:00 crc kubenswrapper[4766]: I1126 01:00:00.221307 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 26 01:00:00 crc kubenswrapper[4766]: I1126 01:00:00.225169 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w"] Nov 26 01:00:00 crc kubenswrapper[4766]: I1126 01:00:00.309379 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a13a895b-28a8-46d1-895d-f9f6cce17490-config-volume\") pod \"collect-profiles-29401980-n2x8w\" (UID: \"a13a895b-28a8-46d1-895d-f9f6cce17490\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w" Nov 26 01:00:00 crc kubenswrapper[4766]: I1126 01:00:00.309431 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a13a895b-28a8-46d1-895d-f9f6cce17490-secret-volume\") pod \"collect-profiles-29401980-n2x8w\" (UID: \"a13a895b-28a8-46d1-895d-f9f6cce17490\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w" Nov 26 01:00:00 crc kubenswrapper[4766]: I1126 01:00:00.309469 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pt54k\" (UniqueName: \"kubernetes.io/projected/a13a895b-28a8-46d1-895d-f9f6cce17490-kube-api-access-pt54k\") pod \"collect-profiles-29401980-n2x8w\" (UID: \"a13a895b-28a8-46d1-895d-f9f6cce17490\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w" Nov 26 01:00:00 crc kubenswrapper[4766]: I1126 01:00:00.412416 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a13a895b-28a8-46d1-895d-f9f6cce17490-config-volume\") pod \"collect-profiles-29401980-n2x8w\" (UID: \"a13a895b-28a8-46d1-895d-f9f6cce17490\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w" Nov 26 01:00:00 crc kubenswrapper[4766]: I1126 01:00:00.412490 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a13a895b-28a8-46d1-895d-f9f6cce17490-secret-volume\") pod \"collect-profiles-29401980-n2x8w\" (UID: \"a13a895b-28a8-46d1-895d-f9f6cce17490\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w" Nov 26 01:00:00 crc kubenswrapper[4766]: I1126 01:00:00.412552 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pt54k\" (UniqueName: \"kubernetes.io/projected/a13a895b-28a8-46d1-895d-f9f6cce17490-kube-api-access-pt54k\") pod \"collect-profiles-29401980-n2x8w\" (UID: \"a13a895b-28a8-46d1-895d-f9f6cce17490\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w" Nov 26 01:00:00 crc kubenswrapper[4766]: I1126 01:00:00.413755 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a13a895b-28a8-46d1-895d-f9f6cce17490-config-volume\") pod \"collect-profiles-29401980-n2x8w\" (UID: \"a13a895b-28a8-46d1-895d-f9f6cce17490\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w" Nov 26 01:00:00 crc kubenswrapper[4766]: I1126 01:00:00.421564 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a13a895b-28a8-46d1-895d-f9f6cce17490-secret-volume\") pod \"collect-profiles-29401980-n2x8w\" (UID: \"a13a895b-28a8-46d1-895d-f9f6cce17490\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w" Nov 26 01:00:00 crc kubenswrapper[4766]: I1126 01:00:00.442099 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pt54k\" (UniqueName: \"kubernetes.io/projected/a13a895b-28a8-46d1-895d-f9f6cce17490-kube-api-access-pt54k\") pod \"collect-profiles-29401980-n2x8w\" (UID: \"a13a895b-28a8-46d1-895d-f9f6cce17490\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w" Nov 26 01:00:00 crc kubenswrapper[4766]: I1126 01:00:00.548291 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w" Nov 26 01:00:01 crc kubenswrapper[4766]: I1126 01:00:01.042615 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w"] Nov 26 01:00:01 crc kubenswrapper[4766]: W1126 01:00:01.044070 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda13a895b_28a8_46d1_895d_f9f6cce17490.slice/crio-d89c05e61365f2342746671ef4504b4639f90ed227faadf20e4e690321a14703 WatchSource:0}: Error finding container d89c05e61365f2342746671ef4504b4639f90ed227faadf20e4e690321a14703: Status 404 returned error can't find the container with id d89c05e61365f2342746671ef4504b4639f90ed227faadf20e4e690321a14703 Nov 26 01:00:01 crc kubenswrapper[4766]: I1126 01:00:01.579766 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w" event={"ID":"a13a895b-28a8-46d1-895d-f9f6cce17490","Type":"ContainerStarted","Data":"f6ac10ea28c25a969ab68c19eb363ff24c8ccc6bdd1e6a40648eca0ae7b16f5a"} Nov 26 01:00:01 crc kubenswrapper[4766]: I1126 01:00:01.580123 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w" event={"ID":"a13a895b-28a8-46d1-895d-f9f6cce17490","Type":"ContainerStarted","Data":"d89c05e61365f2342746671ef4504b4639f90ed227faadf20e4e690321a14703"} Nov 26 01:00:01 crc kubenswrapper[4766]: I1126 01:00:01.608447 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w" podStartSLOduration=1.608424516 podStartE2EDuration="1.608424516s" podCreationTimestamp="2025-11-26 01:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 01:00:01.595410195 +0000 UTC m=+2182.444180645" watchObservedRunningTime="2025-11-26 01:00:01.608424516 +0000 UTC m=+2182.457194966" Nov 26 01:00:02 crc kubenswrapper[4766]: I1126 01:00:02.594039 4766 generic.go:334] "Generic (PLEG): container finished" podID="a13a895b-28a8-46d1-895d-f9f6cce17490" containerID="f6ac10ea28c25a969ab68c19eb363ff24c8ccc6bdd1e6a40648eca0ae7b16f5a" exitCode=0 Nov 26 01:00:02 crc kubenswrapper[4766]: I1126 01:00:02.594104 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w" event={"ID":"a13a895b-28a8-46d1-895d-f9f6cce17490","Type":"ContainerDied","Data":"f6ac10ea28c25a969ab68c19eb363ff24c8ccc6bdd1e6a40648eca0ae7b16f5a"} Nov 26 01:00:04 crc kubenswrapper[4766]: I1126 01:00:04.157184 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w" Nov 26 01:00:04 crc kubenswrapper[4766]: I1126 01:00:04.227195 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a13a895b-28a8-46d1-895d-f9f6cce17490-config-volume\") pod \"a13a895b-28a8-46d1-895d-f9f6cce17490\" (UID: \"a13a895b-28a8-46d1-895d-f9f6cce17490\") " Nov 26 01:00:04 crc kubenswrapper[4766]: I1126 01:00:04.227341 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pt54k\" (UniqueName: \"kubernetes.io/projected/a13a895b-28a8-46d1-895d-f9f6cce17490-kube-api-access-pt54k\") pod \"a13a895b-28a8-46d1-895d-f9f6cce17490\" (UID: \"a13a895b-28a8-46d1-895d-f9f6cce17490\") " Nov 26 01:00:04 crc kubenswrapper[4766]: I1126 01:00:04.227409 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a13a895b-28a8-46d1-895d-f9f6cce17490-secret-volume\") pod \"a13a895b-28a8-46d1-895d-f9f6cce17490\" (UID: \"a13a895b-28a8-46d1-895d-f9f6cce17490\") " Nov 26 01:00:04 crc kubenswrapper[4766]: I1126 01:00:04.227797 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a13a895b-28a8-46d1-895d-f9f6cce17490-config-volume" (OuterVolumeSpecName: "config-volume") pod "a13a895b-28a8-46d1-895d-f9f6cce17490" (UID: "a13a895b-28a8-46d1-895d-f9f6cce17490"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:00:04 crc kubenswrapper[4766]: I1126 01:00:04.228293 4766 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a13a895b-28a8-46d1-895d-f9f6cce17490-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 01:00:04 crc kubenswrapper[4766]: I1126 01:00:04.236593 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a13a895b-28a8-46d1-895d-f9f6cce17490-kube-api-access-pt54k" (OuterVolumeSpecName: "kube-api-access-pt54k") pod "a13a895b-28a8-46d1-895d-f9f6cce17490" (UID: "a13a895b-28a8-46d1-895d-f9f6cce17490"). InnerVolumeSpecName "kube-api-access-pt54k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:00:04 crc kubenswrapper[4766]: I1126 01:00:04.243548 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a13a895b-28a8-46d1-895d-f9f6cce17490-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a13a895b-28a8-46d1-895d-f9f6cce17490" (UID: "a13a895b-28a8-46d1-895d-f9f6cce17490"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:00:04 crc kubenswrapper[4766]: I1126 01:00:04.331450 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pt54k\" (UniqueName: \"kubernetes.io/projected/a13a895b-28a8-46d1-895d-f9f6cce17490-kube-api-access-pt54k\") on node \"crc\" DevicePath \"\"" Nov 26 01:00:04 crc kubenswrapper[4766]: I1126 01:00:04.331505 4766 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a13a895b-28a8-46d1-895d-f9f6cce17490-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 26 01:00:04 crc kubenswrapper[4766]: I1126 01:00:04.622372 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w" event={"ID":"a13a895b-28a8-46d1-895d-f9f6cce17490","Type":"ContainerDied","Data":"d89c05e61365f2342746671ef4504b4639f90ed227faadf20e4e690321a14703"} Nov 26 01:00:04 crc kubenswrapper[4766]: I1126 01:00:04.622413 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d89c05e61365f2342746671ef4504b4639f90ed227faadf20e4e690321a14703" Nov 26 01:00:04 crc kubenswrapper[4766]: I1126 01:00:04.622476 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w" Nov 26 01:00:04 crc kubenswrapper[4766]: I1126 01:00:04.705204 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9"] Nov 26 01:00:04 crc kubenswrapper[4766]: I1126 01:00:04.714197 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401935-hhtv9"] Nov 26 01:00:05 crc kubenswrapper[4766]: I1126 01:00:05.841391 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dea338f1-f870-484b-ac43-7da254eac47a" path="/var/lib/kubelet/pods/dea338f1-f870-484b-ac43-7da254eac47a/volumes" Nov 26 01:00:09 crc kubenswrapper[4766]: I1126 01:00:09.714026 4766 generic.go:334] "Generic (PLEG): container finished" podID="f264bace-32cb-4fb6-acab-40d59db68d6e" containerID="c1e94194b85eda72901f1729372aa218cc8883221639975d9e52f2a1c7979027" exitCode=0 Nov 26 01:00:09 crc kubenswrapper[4766]: I1126 01:00:09.714161 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" event={"ID":"f264bace-32cb-4fb6-acab-40d59db68d6e","Type":"ContainerDied","Data":"c1e94194b85eda72901f1729372aa218cc8883221639975d9e52f2a1c7979027"} Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.479615 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.480236 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.644849 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.720029 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-inventory\") pod \"f264bace-32cb-4fb6-acab-40d59db68d6e\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.720381 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-telemetry-combined-ca-bundle\") pod \"f264bace-32cb-4fb6-acab-40d59db68d6e\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.720414 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g75k2\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-kube-api-access-g75k2\") pod \"f264bace-32cb-4fb6-acab-40d59db68d6e\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.720472 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"f264bace-32cb-4fb6-acab-40d59db68d6e\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.720492 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-bootstrap-combined-ca-bundle\") pod \"f264bace-32cb-4fb6-acab-40d59db68d6e\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.720523 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-telemetry-power-monitoring-combined-ca-bundle\") pod \"f264bace-32cb-4fb6-acab-40d59db68d6e\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.720553 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"f264bace-32cb-4fb6-acab-40d59db68d6e\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.720598 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"f264bace-32cb-4fb6-acab-40d59db68d6e\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.720715 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-ssh-key\") pod \"f264bace-32cb-4fb6-acab-40d59db68d6e\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.720732 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-libvirt-combined-ca-bundle\") pod \"f264bace-32cb-4fb6-acab-40d59db68d6e\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.720763 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-repo-setup-combined-ca-bundle\") pod \"f264bace-32cb-4fb6-acab-40d59db68d6e\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.720793 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"f264bace-32cb-4fb6-acab-40d59db68d6e\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.720824 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-ovn-combined-ca-bundle\") pod \"f264bace-32cb-4fb6-acab-40d59db68d6e\" (UID: \"f264bace-32cb-4fb6-acab-40d59db68d6e\") " Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.727380 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "f264bace-32cb-4fb6-acab-40d59db68d6e" (UID: "f264bace-32cb-4fb6-acab-40d59db68d6e"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.727574 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-kube-api-access-g75k2" (OuterVolumeSpecName: "kube-api-access-g75k2") pod "f264bace-32cb-4fb6-acab-40d59db68d6e" (UID: "f264bace-32cb-4fb6-acab-40d59db68d6e"). InnerVolumeSpecName "kube-api-access-g75k2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.728911 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0") pod "f264bace-32cb-4fb6-acab-40d59db68d6e" (UID: "f264bace-32cb-4fb6-acab-40d59db68d6e"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.728994 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "f264bace-32cb-4fb6-acab-40d59db68d6e" (UID: "f264bace-32cb-4fb6-acab-40d59db68d6e"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.729796 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "f264bace-32cb-4fb6-acab-40d59db68d6e" (UID: "f264bace-32cb-4fb6-acab-40d59db68d6e"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.730417 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "f264bace-32cb-4fb6-acab-40d59db68d6e" (UID: "f264bace-32cb-4fb6-acab-40d59db68d6e"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.731202 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "f264bace-32cb-4fb6-acab-40d59db68d6e" (UID: "f264bace-32cb-4fb6-acab-40d59db68d6e"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.731264 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "f264bace-32cb-4fb6-acab-40d59db68d6e" (UID: "f264bace-32cb-4fb6-acab-40d59db68d6e"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.731899 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "f264bace-32cb-4fb6-acab-40d59db68d6e" (UID: "f264bace-32cb-4fb6-acab-40d59db68d6e"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.735232 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "f264bace-32cb-4fb6-acab-40d59db68d6e" (UID: "f264bace-32cb-4fb6-acab-40d59db68d6e"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.735489 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "f264bace-32cb-4fb6-acab-40d59db68d6e" (UID: "f264bace-32cb-4fb6-acab-40d59db68d6e"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.742001 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" event={"ID":"f264bace-32cb-4fb6-acab-40d59db68d6e","Type":"ContainerDied","Data":"eeeac981735f1493788c1301ce1e4ede3bc32615bb83626f316ea2ed8aab3ac0"} Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.742215 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eeeac981735f1493788c1301ce1e4ede3bc32615bb83626f316ea2ed8aab3ac0" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.742075 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.762009 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f264bace-32cb-4fb6-acab-40d59db68d6e" (UID: "f264bace-32cb-4fb6-acab-40d59db68d6e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.764318 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-inventory" (OuterVolumeSpecName: "inventory") pod "f264bace-32cb-4fb6-acab-40d59db68d6e" (UID: "f264bace-32cb-4fb6-acab-40d59db68d6e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.826247 4766 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.826297 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.826314 4766 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.826330 4766 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.826352 4766 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.826368 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.826385 4766 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.826397 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g75k2\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-kube-api-access-g75k2\") on node \"crc\" DevicePath \"\"" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.826409 4766 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.826420 4766 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.826435 4766 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f264bace-32cb-4fb6-acab-40d59db68d6e-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.826447 4766 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.826459 4766 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f264bace-32cb-4fb6-acab-40d59db68d6e-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.869600 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z"] Nov 26 01:00:11 crc kubenswrapper[4766]: E1126 01:00:11.870074 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a13a895b-28a8-46d1-895d-f9f6cce17490" containerName="collect-profiles" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.870094 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="a13a895b-28a8-46d1-895d-f9f6cce17490" containerName="collect-profiles" Nov 26 01:00:11 crc kubenswrapper[4766]: E1126 01:00:11.870145 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f264bace-32cb-4fb6-acab-40d59db68d6e" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.870152 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="f264bace-32cb-4fb6-acab-40d59db68d6e" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.870347 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="a13a895b-28a8-46d1-895d-f9f6cce17490" containerName="collect-profiles" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.870367 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="f264bace-32cb-4fb6-acab-40d59db68d6e" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.871104 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.875438 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 26 01:00:11 crc kubenswrapper[4766]: I1126 01:00:11.884350 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z"] Nov 26 01:00:12 crc kubenswrapper[4766]: I1126 01:00:12.031268 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d456c126-aca3-47d2-b4fb-7ab506d116fb-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dzc8z\" (UID: \"d456c126-aca3-47d2-b4fb-7ab506d116fb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" Nov 26 01:00:12 crc kubenswrapper[4766]: I1126 01:00:12.031591 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mxbt\" (UniqueName: \"kubernetes.io/projected/d456c126-aca3-47d2-b4fb-7ab506d116fb-kube-api-access-8mxbt\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dzc8z\" (UID: \"d456c126-aca3-47d2-b4fb-7ab506d116fb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" Nov 26 01:00:12 crc kubenswrapper[4766]: I1126 01:00:12.032089 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d456c126-aca3-47d2-b4fb-7ab506d116fb-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dzc8z\" (UID: \"d456c126-aca3-47d2-b4fb-7ab506d116fb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" Nov 26 01:00:12 crc kubenswrapper[4766]: I1126 01:00:12.032301 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d456c126-aca3-47d2-b4fb-7ab506d116fb-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dzc8z\" (UID: \"d456c126-aca3-47d2-b4fb-7ab506d116fb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" Nov 26 01:00:12 crc kubenswrapper[4766]: I1126 01:00:12.032379 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d456c126-aca3-47d2-b4fb-7ab506d116fb-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dzc8z\" (UID: \"d456c126-aca3-47d2-b4fb-7ab506d116fb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" Nov 26 01:00:12 crc kubenswrapper[4766]: I1126 01:00:12.134011 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d456c126-aca3-47d2-b4fb-7ab506d116fb-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dzc8z\" (UID: \"d456c126-aca3-47d2-b4fb-7ab506d116fb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" Nov 26 01:00:12 crc kubenswrapper[4766]: I1126 01:00:12.134087 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d456c126-aca3-47d2-b4fb-7ab506d116fb-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dzc8z\" (UID: \"d456c126-aca3-47d2-b4fb-7ab506d116fb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" Nov 26 01:00:12 crc kubenswrapper[4766]: I1126 01:00:12.134120 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d456c126-aca3-47d2-b4fb-7ab506d116fb-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dzc8z\" (UID: \"d456c126-aca3-47d2-b4fb-7ab506d116fb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" Nov 26 01:00:12 crc kubenswrapper[4766]: I1126 01:00:12.134196 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d456c126-aca3-47d2-b4fb-7ab506d116fb-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dzc8z\" (UID: \"d456c126-aca3-47d2-b4fb-7ab506d116fb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" Nov 26 01:00:12 crc kubenswrapper[4766]: I1126 01:00:12.134278 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mxbt\" (UniqueName: \"kubernetes.io/projected/d456c126-aca3-47d2-b4fb-7ab506d116fb-kube-api-access-8mxbt\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dzc8z\" (UID: \"d456c126-aca3-47d2-b4fb-7ab506d116fb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" Nov 26 01:00:12 crc kubenswrapper[4766]: I1126 01:00:12.135114 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d456c126-aca3-47d2-b4fb-7ab506d116fb-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dzc8z\" (UID: \"d456c126-aca3-47d2-b4fb-7ab506d116fb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" Nov 26 01:00:12 crc kubenswrapper[4766]: I1126 01:00:12.141645 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d456c126-aca3-47d2-b4fb-7ab506d116fb-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dzc8z\" (UID: \"d456c126-aca3-47d2-b4fb-7ab506d116fb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" Nov 26 01:00:12 crc kubenswrapper[4766]: I1126 01:00:12.142163 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d456c126-aca3-47d2-b4fb-7ab506d116fb-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dzc8z\" (UID: \"d456c126-aca3-47d2-b4fb-7ab506d116fb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" Nov 26 01:00:12 crc kubenswrapper[4766]: I1126 01:00:12.143092 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d456c126-aca3-47d2-b4fb-7ab506d116fb-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dzc8z\" (UID: \"d456c126-aca3-47d2-b4fb-7ab506d116fb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" Nov 26 01:00:12 crc kubenswrapper[4766]: I1126 01:00:12.161969 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mxbt\" (UniqueName: \"kubernetes.io/projected/d456c126-aca3-47d2-b4fb-7ab506d116fb-kube-api-access-8mxbt\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dzc8z\" (UID: \"d456c126-aca3-47d2-b4fb-7ab506d116fb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" Nov 26 01:00:12 crc kubenswrapper[4766]: I1126 01:00:12.210162 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" Nov 26 01:00:12 crc kubenswrapper[4766]: I1126 01:00:12.814284 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z"] Nov 26 01:00:12 crc kubenswrapper[4766]: I1126 01:00:12.821886 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 01:00:13 crc kubenswrapper[4766]: I1126 01:00:13.784888 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" event={"ID":"d456c126-aca3-47d2-b4fb-7ab506d116fb","Type":"ContainerStarted","Data":"22fff6cb2b6b099a0362c698eb539824d4f86ce71060d1a9f8b3e798f72badfa"} Nov 26 01:00:13 crc kubenswrapper[4766]: I1126 01:00:13.785487 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" event={"ID":"d456c126-aca3-47d2-b4fb-7ab506d116fb","Type":"ContainerStarted","Data":"c495c01a66ab8a57acbe084b59d9e5e3555566ad8f06d0b17de2c95a1ae7ad1d"} Nov 26 01:00:13 crc kubenswrapper[4766]: I1126 01:00:13.808207 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" podStartSLOduration=2.25930097 podStartE2EDuration="2.808184715s" podCreationTimestamp="2025-11-26 01:00:11 +0000 UTC" firstStartedPulling="2025-11-26 01:00:12.821016501 +0000 UTC m=+2193.669786961" lastFinishedPulling="2025-11-26 01:00:13.369900276 +0000 UTC m=+2194.218670706" observedRunningTime="2025-11-26 01:00:13.802934195 +0000 UTC m=+2194.651704665" watchObservedRunningTime="2025-11-26 01:00:13.808184715 +0000 UTC m=+2194.656955155" Nov 26 01:00:19 crc kubenswrapper[4766]: I1126 01:00:19.082500 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c9rnq"] Nov 26 01:00:19 crc kubenswrapper[4766]: I1126 01:00:19.086334 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9rnq" Nov 26 01:00:19 crc kubenswrapper[4766]: I1126 01:00:19.130225 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9rnq"] Nov 26 01:00:19 crc kubenswrapper[4766]: I1126 01:00:19.209924 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdfpr\" (UniqueName: \"kubernetes.io/projected/e0685d87-9d72-45d7-b17e-378f035a9eea-kube-api-access-zdfpr\") pod \"redhat-marketplace-c9rnq\" (UID: \"e0685d87-9d72-45d7-b17e-378f035a9eea\") " pod="openshift-marketplace/redhat-marketplace-c9rnq" Nov 26 01:00:19 crc kubenswrapper[4766]: I1126 01:00:19.210311 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0685d87-9d72-45d7-b17e-378f035a9eea-utilities\") pod \"redhat-marketplace-c9rnq\" (UID: \"e0685d87-9d72-45d7-b17e-378f035a9eea\") " pod="openshift-marketplace/redhat-marketplace-c9rnq" Nov 26 01:00:19 crc kubenswrapper[4766]: I1126 01:00:19.210487 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0685d87-9d72-45d7-b17e-378f035a9eea-catalog-content\") pod \"redhat-marketplace-c9rnq\" (UID: \"e0685d87-9d72-45d7-b17e-378f035a9eea\") " pod="openshift-marketplace/redhat-marketplace-c9rnq" Nov 26 01:00:19 crc kubenswrapper[4766]: I1126 01:00:19.312594 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdfpr\" (UniqueName: \"kubernetes.io/projected/e0685d87-9d72-45d7-b17e-378f035a9eea-kube-api-access-zdfpr\") pod \"redhat-marketplace-c9rnq\" (UID: \"e0685d87-9d72-45d7-b17e-378f035a9eea\") " pod="openshift-marketplace/redhat-marketplace-c9rnq" Nov 26 01:00:19 crc kubenswrapper[4766]: I1126 01:00:19.312703 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0685d87-9d72-45d7-b17e-378f035a9eea-utilities\") pod \"redhat-marketplace-c9rnq\" (UID: \"e0685d87-9d72-45d7-b17e-378f035a9eea\") " pod="openshift-marketplace/redhat-marketplace-c9rnq" Nov 26 01:00:19 crc kubenswrapper[4766]: I1126 01:00:19.313352 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0685d87-9d72-45d7-b17e-378f035a9eea-utilities\") pod \"redhat-marketplace-c9rnq\" (UID: \"e0685d87-9d72-45d7-b17e-378f035a9eea\") " pod="openshift-marketplace/redhat-marketplace-c9rnq" Nov 26 01:00:19 crc kubenswrapper[4766]: I1126 01:00:19.313419 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0685d87-9d72-45d7-b17e-378f035a9eea-catalog-content\") pod \"redhat-marketplace-c9rnq\" (UID: \"e0685d87-9d72-45d7-b17e-378f035a9eea\") " pod="openshift-marketplace/redhat-marketplace-c9rnq" Nov 26 01:00:19 crc kubenswrapper[4766]: I1126 01:00:19.313684 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0685d87-9d72-45d7-b17e-378f035a9eea-catalog-content\") pod \"redhat-marketplace-c9rnq\" (UID: \"e0685d87-9d72-45d7-b17e-378f035a9eea\") " pod="openshift-marketplace/redhat-marketplace-c9rnq" Nov 26 01:00:19 crc kubenswrapper[4766]: I1126 01:00:19.334933 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdfpr\" (UniqueName: \"kubernetes.io/projected/e0685d87-9d72-45d7-b17e-378f035a9eea-kube-api-access-zdfpr\") pod \"redhat-marketplace-c9rnq\" (UID: \"e0685d87-9d72-45d7-b17e-378f035a9eea\") " pod="openshift-marketplace/redhat-marketplace-c9rnq" Nov 26 01:00:19 crc kubenswrapper[4766]: I1126 01:00:19.424105 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9rnq" Nov 26 01:00:19 crc kubenswrapper[4766]: I1126 01:00:19.962234 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9rnq"] Nov 26 01:00:20 crc kubenswrapper[4766]: I1126 01:00:20.886754 4766 generic.go:334] "Generic (PLEG): container finished" podID="e0685d87-9d72-45d7-b17e-378f035a9eea" containerID="ecce66f5ae15d355584b3218e57b85be972b82c19528fc178e98c6376966886e" exitCode=0 Nov 26 01:00:20 crc kubenswrapper[4766]: I1126 01:00:20.886868 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9rnq" event={"ID":"e0685d87-9d72-45d7-b17e-378f035a9eea","Type":"ContainerDied","Data":"ecce66f5ae15d355584b3218e57b85be972b82c19528fc178e98c6376966886e"} Nov 26 01:00:20 crc kubenswrapper[4766]: I1126 01:00:20.887297 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9rnq" event={"ID":"e0685d87-9d72-45d7-b17e-378f035a9eea","Type":"ContainerStarted","Data":"0154b19180d103a15f3088db4685597e9072f0081531fd7439485744aa0144c9"} Nov 26 01:00:34 crc kubenswrapper[4766]: I1126 01:00:34.058115 4766 generic.go:334] "Generic (PLEG): container finished" podID="e0685d87-9d72-45d7-b17e-378f035a9eea" containerID="92badd2b9c3ada4968a308d88c61acb6b1b08ff38bdf41ed809d16b1299dde1e" exitCode=0 Nov 26 01:00:34 crc kubenswrapper[4766]: I1126 01:00:34.058231 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9rnq" event={"ID":"e0685d87-9d72-45d7-b17e-378f035a9eea","Type":"ContainerDied","Data":"92badd2b9c3ada4968a308d88c61acb6b1b08ff38bdf41ed809d16b1299dde1e"} Nov 26 01:00:35 crc kubenswrapper[4766]: I1126 01:00:35.070538 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9rnq" event={"ID":"e0685d87-9d72-45d7-b17e-378f035a9eea","Type":"ContainerStarted","Data":"06be87534f55efa79388c71f527b415e99b4360f1f30d3c8e749bcdae68e5bd2"} Nov 26 01:00:35 crc kubenswrapper[4766]: I1126 01:00:35.095979 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c9rnq" podStartSLOduration=2.511067168 podStartE2EDuration="16.095952703s" podCreationTimestamp="2025-11-26 01:00:19 +0000 UTC" firstStartedPulling="2025-11-26 01:00:20.889777281 +0000 UTC m=+2201.738547741" lastFinishedPulling="2025-11-26 01:00:34.474662836 +0000 UTC m=+2215.323433276" observedRunningTime="2025-11-26 01:00:35.090922679 +0000 UTC m=+2215.939693139" watchObservedRunningTime="2025-11-26 01:00:35.095952703 +0000 UTC m=+2215.944723133" Nov 26 01:00:39 crc kubenswrapper[4766]: I1126 01:00:39.424806 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c9rnq" Nov 26 01:00:39 crc kubenswrapper[4766]: I1126 01:00:39.425404 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c9rnq" Nov 26 01:00:39 crc kubenswrapper[4766]: I1126 01:00:39.481763 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c9rnq" Nov 26 01:00:40 crc kubenswrapper[4766]: I1126 01:00:40.214846 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c9rnq" Nov 26 01:00:40 crc kubenswrapper[4766]: I1126 01:00:40.285569 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9rnq"] Nov 26 01:00:41 crc kubenswrapper[4766]: I1126 01:00:41.479460 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:00:41 crc kubenswrapper[4766]: I1126 01:00:41.479833 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:00:42 crc kubenswrapper[4766]: I1126 01:00:42.167506 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c9rnq" podUID="e0685d87-9d72-45d7-b17e-378f035a9eea" containerName="registry-server" containerID="cri-o://06be87534f55efa79388c71f527b415e99b4360f1f30d3c8e749bcdae68e5bd2" gracePeriod=2 Nov 26 01:00:42 crc kubenswrapper[4766]: E1126 01:00:42.326225 4766 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0685d87_9d72_45d7_b17e_378f035a9eea.slice/crio-conmon-06be87534f55efa79388c71f527b415e99b4360f1f30d3c8e749bcdae68e5bd2.scope\": RecentStats: unable to find data in memory cache]" Nov 26 01:00:42 crc kubenswrapper[4766]: I1126 01:00:42.746874 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9rnq" Nov 26 01:00:42 crc kubenswrapper[4766]: I1126 01:00:42.777367 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdfpr\" (UniqueName: \"kubernetes.io/projected/e0685d87-9d72-45d7-b17e-378f035a9eea-kube-api-access-zdfpr\") pod \"e0685d87-9d72-45d7-b17e-378f035a9eea\" (UID: \"e0685d87-9d72-45d7-b17e-378f035a9eea\") " Nov 26 01:00:42 crc kubenswrapper[4766]: I1126 01:00:42.777808 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0685d87-9d72-45d7-b17e-378f035a9eea-utilities\") pod \"e0685d87-9d72-45d7-b17e-378f035a9eea\" (UID: \"e0685d87-9d72-45d7-b17e-378f035a9eea\") " Nov 26 01:00:42 crc kubenswrapper[4766]: I1126 01:00:42.777941 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0685d87-9d72-45d7-b17e-378f035a9eea-catalog-content\") pod \"e0685d87-9d72-45d7-b17e-378f035a9eea\" (UID: \"e0685d87-9d72-45d7-b17e-378f035a9eea\") " Nov 26 01:00:42 crc kubenswrapper[4766]: I1126 01:00:42.783569 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0685d87-9d72-45d7-b17e-378f035a9eea-utilities" (OuterVolumeSpecName: "utilities") pod "e0685d87-9d72-45d7-b17e-378f035a9eea" (UID: "e0685d87-9d72-45d7-b17e-378f035a9eea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:00:42 crc kubenswrapper[4766]: I1126 01:00:42.794924 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0685d87-9d72-45d7-b17e-378f035a9eea-kube-api-access-zdfpr" (OuterVolumeSpecName: "kube-api-access-zdfpr") pod "e0685d87-9d72-45d7-b17e-378f035a9eea" (UID: "e0685d87-9d72-45d7-b17e-378f035a9eea"). InnerVolumeSpecName "kube-api-access-zdfpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:00:42 crc kubenswrapper[4766]: I1126 01:00:42.797677 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0685d87-9d72-45d7-b17e-378f035a9eea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0685d87-9d72-45d7-b17e-378f035a9eea" (UID: "e0685d87-9d72-45d7-b17e-378f035a9eea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:00:42 crc kubenswrapper[4766]: I1126 01:00:42.880598 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdfpr\" (UniqueName: \"kubernetes.io/projected/e0685d87-9d72-45d7-b17e-378f035a9eea-kube-api-access-zdfpr\") on node \"crc\" DevicePath \"\"" Nov 26 01:00:42 crc kubenswrapper[4766]: I1126 01:00:42.880629 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0685d87-9d72-45d7-b17e-378f035a9eea-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 01:00:42 crc kubenswrapper[4766]: I1126 01:00:42.880638 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0685d87-9d72-45d7-b17e-378f035a9eea-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 01:00:43 crc kubenswrapper[4766]: I1126 01:00:43.180504 4766 generic.go:334] "Generic (PLEG): container finished" podID="e0685d87-9d72-45d7-b17e-378f035a9eea" containerID="06be87534f55efa79388c71f527b415e99b4360f1f30d3c8e749bcdae68e5bd2" exitCode=0 Nov 26 01:00:43 crc kubenswrapper[4766]: I1126 01:00:43.180550 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9rnq" event={"ID":"e0685d87-9d72-45d7-b17e-378f035a9eea","Type":"ContainerDied","Data":"06be87534f55efa79388c71f527b415e99b4360f1f30d3c8e749bcdae68e5bd2"} Nov 26 01:00:43 crc kubenswrapper[4766]: I1126 01:00:43.180580 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9rnq" event={"ID":"e0685d87-9d72-45d7-b17e-378f035a9eea","Type":"ContainerDied","Data":"0154b19180d103a15f3088db4685597e9072f0081531fd7439485744aa0144c9"} Nov 26 01:00:43 crc kubenswrapper[4766]: I1126 01:00:43.180600 4766 scope.go:117] "RemoveContainer" containerID="06be87534f55efa79388c71f527b415e99b4360f1f30d3c8e749bcdae68e5bd2" Nov 26 01:00:43 crc kubenswrapper[4766]: I1126 01:00:43.180747 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9rnq" Nov 26 01:00:43 crc kubenswrapper[4766]: I1126 01:00:43.221711 4766 scope.go:117] "RemoveContainer" containerID="92badd2b9c3ada4968a308d88c61acb6b1b08ff38bdf41ed809d16b1299dde1e" Nov 26 01:00:43 crc kubenswrapper[4766]: I1126 01:00:43.250765 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9rnq"] Nov 26 01:00:43 crc kubenswrapper[4766]: I1126 01:00:43.258276 4766 scope.go:117] "RemoveContainer" containerID="ecce66f5ae15d355584b3218e57b85be972b82c19528fc178e98c6376966886e" Nov 26 01:00:43 crc kubenswrapper[4766]: I1126 01:00:43.266281 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9rnq"] Nov 26 01:00:43 crc kubenswrapper[4766]: I1126 01:00:43.310043 4766 scope.go:117] "RemoveContainer" containerID="06be87534f55efa79388c71f527b415e99b4360f1f30d3c8e749bcdae68e5bd2" Nov 26 01:00:43 crc kubenswrapper[4766]: E1126 01:00:43.310628 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06be87534f55efa79388c71f527b415e99b4360f1f30d3c8e749bcdae68e5bd2\": container with ID starting with 06be87534f55efa79388c71f527b415e99b4360f1f30d3c8e749bcdae68e5bd2 not found: ID does not exist" containerID="06be87534f55efa79388c71f527b415e99b4360f1f30d3c8e749bcdae68e5bd2" Nov 26 01:00:43 crc kubenswrapper[4766]: I1126 01:00:43.310793 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06be87534f55efa79388c71f527b415e99b4360f1f30d3c8e749bcdae68e5bd2"} err="failed to get container status \"06be87534f55efa79388c71f527b415e99b4360f1f30d3c8e749bcdae68e5bd2\": rpc error: code = NotFound desc = could not find container \"06be87534f55efa79388c71f527b415e99b4360f1f30d3c8e749bcdae68e5bd2\": container with ID starting with 06be87534f55efa79388c71f527b415e99b4360f1f30d3c8e749bcdae68e5bd2 not found: ID does not exist" Nov 26 01:00:43 crc kubenswrapper[4766]: I1126 01:00:43.310824 4766 scope.go:117] "RemoveContainer" containerID="92badd2b9c3ada4968a308d88c61acb6b1b08ff38bdf41ed809d16b1299dde1e" Nov 26 01:00:43 crc kubenswrapper[4766]: E1126 01:00:43.311331 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92badd2b9c3ada4968a308d88c61acb6b1b08ff38bdf41ed809d16b1299dde1e\": container with ID starting with 92badd2b9c3ada4968a308d88c61acb6b1b08ff38bdf41ed809d16b1299dde1e not found: ID does not exist" containerID="92badd2b9c3ada4968a308d88c61acb6b1b08ff38bdf41ed809d16b1299dde1e" Nov 26 01:00:43 crc kubenswrapper[4766]: I1126 01:00:43.311369 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92badd2b9c3ada4968a308d88c61acb6b1b08ff38bdf41ed809d16b1299dde1e"} err="failed to get container status \"92badd2b9c3ada4968a308d88c61acb6b1b08ff38bdf41ed809d16b1299dde1e\": rpc error: code = NotFound desc = could not find container \"92badd2b9c3ada4968a308d88c61acb6b1b08ff38bdf41ed809d16b1299dde1e\": container with ID starting with 92badd2b9c3ada4968a308d88c61acb6b1b08ff38bdf41ed809d16b1299dde1e not found: ID does not exist" Nov 26 01:00:43 crc kubenswrapper[4766]: I1126 01:00:43.311396 4766 scope.go:117] "RemoveContainer" containerID="ecce66f5ae15d355584b3218e57b85be972b82c19528fc178e98c6376966886e" Nov 26 01:00:43 crc kubenswrapper[4766]: E1126 01:00:43.311883 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecce66f5ae15d355584b3218e57b85be972b82c19528fc178e98c6376966886e\": container with ID starting with ecce66f5ae15d355584b3218e57b85be972b82c19528fc178e98c6376966886e not found: ID does not exist" containerID="ecce66f5ae15d355584b3218e57b85be972b82c19528fc178e98c6376966886e" Nov 26 01:00:43 crc kubenswrapper[4766]: I1126 01:00:43.311917 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecce66f5ae15d355584b3218e57b85be972b82c19528fc178e98c6376966886e"} err="failed to get container status \"ecce66f5ae15d355584b3218e57b85be972b82c19528fc178e98c6376966886e\": rpc error: code = NotFound desc = could not find container \"ecce66f5ae15d355584b3218e57b85be972b82c19528fc178e98c6376966886e\": container with ID starting with ecce66f5ae15d355584b3218e57b85be972b82c19528fc178e98c6376966886e not found: ID does not exist" Nov 26 01:00:43 crc kubenswrapper[4766]: I1126 01:00:43.839641 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0685d87-9d72-45d7-b17e-378f035a9eea" path="/var/lib/kubelet/pods/e0685d87-9d72-45d7-b17e-378f035a9eea/volumes" Nov 26 01:00:59 crc kubenswrapper[4766]: I1126 01:00:59.263984 4766 scope.go:117] "RemoveContainer" containerID="098f03a0a95f8d40a5860d0c453b6ff77121c9b9e68b9d956cda0c589be117cb" Nov 26 01:01:00 crc kubenswrapper[4766]: I1126 01:01:00.179914 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29401981-22f5z"] Nov 26 01:01:00 crc kubenswrapper[4766]: E1126 01:01:00.180936 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0685d87-9d72-45d7-b17e-378f035a9eea" containerName="extract-content" Nov 26 01:01:00 crc kubenswrapper[4766]: I1126 01:01:00.180964 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0685d87-9d72-45d7-b17e-378f035a9eea" containerName="extract-content" Nov 26 01:01:00 crc kubenswrapper[4766]: E1126 01:01:00.181009 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0685d87-9d72-45d7-b17e-378f035a9eea" containerName="extract-utilities" Nov 26 01:01:00 crc kubenswrapper[4766]: I1126 01:01:00.181023 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0685d87-9d72-45d7-b17e-378f035a9eea" containerName="extract-utilities" Nov 26 01:01:00 crc kubenswrapper[4766]: E1126 01:01:00.181047 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0685d87-9d72-45d7-b17e-378f035a9eea" containerName="registry-server" Nov 26 01:01:00 crc kubenswrapper[4766]: I1126 01:01:00.181059 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0685d87-9d72-45d7-b17e-378f035a9eea" containerName="registry-server" Nov 26 01:01:00 crc kubenswrapper[4766]: I1126 01:01:00.181618 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0685d87-9d72-45d7-b17e-378f035a9eea" containerName="registry-server" Nov 26 01:01:00 crc kubenswrapper[4766]: I1126 01:01:00.183727 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401981-22f5z" Nov 26 01:01:00 crc kubenswrapper[4766]: I1126 01:01:00.207302 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29401981-22f5z"] Nov 26 01:01:00 crc kubenswrapper[4766]: I1126 01:01:00.275182 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrb9k\" (UniqueName: \"kubernetes.io/projected/5e25beeb-e89d-4aed-9353-3755a0ef2efb-kube-api-access-rrb9k\") pod \"keystone-cron-29401981-22f5z\" (UID: \"5e25beeb-e89d-4aed-9353-3755a0ef2efb\") " pod="openstack/keystone-cron-29401981-22f5z" Nov 26 01:01:00 crc kubenswrapper[4766]: I1126 01:01:00.275231 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e25beeb-e89d-4aed-9353-3755a0ef2efb-config-data\") pod \"keystone-cron-29401981-22f5z\" (UID: \"5e25beeb-e89d-4aed-9353-3755a0ef2efb\") " pod="openstack/keystone-cron-29401981-22f5z" Nov 26 01:01:00 crc kubenswrapper[4766]: I1126 01:01:00.275312 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5e25beeb-e89d-4aed-9353-3755a0ef2efb-fernet-keys\") pod \"keystone-cron-29401981-22f5z\" (UID: \"5e25beeb-e89d-4aed-9353-3755a0ef2efb\") " pod="openstack/keystone-cron-29401981-22f5z" Nov 26 01:01:00 crc kubenswrapper[4766]: I1126 01:01:00.275338 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e25beeb-e89d-4aed-9353-3755a0ef2efb-combined-ca-bundle\") pod \"keystone-cron-29401981-22f5z\" (UID: \"5e25beeb-e89d-4aed-9353-3755a0ef2efb\") " pod="openstack/keystone-cron-29401981-22f5z" Nov 26 01:01:00 crc kubenswrapper[4766]: I1126 01:01:00.377305 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e25beeb-e89d-4aed-9353-3755a0ef2efb-config-data\") pod \"keystone-cron-29401981-22f5z\" (UID: \"5e25beeb-e89d-4aed-9353-3755a0ef2efb\") " pod="openstack/keystone-cron-29401981-22f5z" Nov 26 01:01:00 crc kubenswrapper[4766]: I1126 01:01:00.377852 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5e25beeb-e89d-4aed-9353-3755a0ef2efb-fernet-keys\") pod \"keystone-cron-29401981-22f5z\" (UID: \"5e25beeb-e89d-4aed-9353-3755a0ef2efb\") " pod="openstack/keystone-cron-29401981-22f5z" Nov 26 01:01:00 crc kubenswrapper[4766]: I1126 01:01:00.378709 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e25beeb-e89d-4aed-9353-3755a0ef2efb-combined-ca-bundle\") pod \"keystone-cron-29401981-22f5z\" (UID: \"5e25beeb-e89d-4aed-9353-3755a0ef2efb\") " pod="openstack/keystone-cron-29401981-22f5z" Nov 26 01:01:00 crc kubenswrapper[4766]: I1126 01:01:00.378936 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrb9k\" (UniqueName: \"kubernetes.io/projected/5e25beeb-e89d-4aed-9353-3755a0ef2efb-kube-api-access-rrb9k\") pod \"keystone-cron-29401981-22f5z\" (UID: \"5e25beeb-e89d-4aed-9353-3755a0ef2efb\") " pod="openstack/keystone-cron-29401981-22f5z" Nov 26 01:01:00 crc kubenswrapper[4766]: I1126 01:01:00.384872 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5e25beeb-e89d-4aed-9353-3755a0ef2efb-fernet-keys\") pod \"keystone-cron-29401981-22f5z\" (UID: \"5e25beeb-e89d-4aed-9353-3755a0ef2efb\") " pod="openstack/keystone-cron-29401981-22f5z" Nov 26 01:01:00 crc kubenswrapper[4766]: I1126 01:01:00.385097 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e25beeb-e89d-4aed-9353-3755a0ef2efb-config-data\") pod \"keystone-cron-29401981-22f5z\" (UID: \"5e25beeb-e89d-4aed-9353-3755a0ef2efb\") " pod="openstack/keystone-cron-29401981-22f5z" Nov 26 01:01:00 crc kubenswrapper[4766]: I1126 01:01:00.386948 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e25beeb-e89d-4aed-9353-3755a0ef2efb-combined-ca-bundle\") pod \"keystone-cron-29401981-22f5z\" (UID: \"5e25beeb-e89d-4aed-9353-3755a0ef2efb\") " pod="openstack/keystone-cron-29401981-22f5z" Nov 26 01:01:00 crc kubenswrapper[4766]: I1126 01:01:00.414532 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrb9k\" (UniqueName: \"kubernetes.io/projected/5e25beeb-e89d-4aed-9353-3755a0ef2efb-kube-api-access-rrb9k\") pod \"keystone-cron-29401981-22f5z\" (UID: \"5e25beeb-e89d-4aed-9353-3755a0ef2efb\") " pod="openstack/keystone-cron-29401981-22f5z" Nov 26 01:01:00 crc kubenswrapper[4766]: I1126 01:01:00.565968 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401981-22f5z" Nov 26 01:01:01 crc kubenswrapper[4766]: I1126 01:01:01.089258 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29401981-22f5z"] Nov 26 01:01:01 crc kubenswrapper[4766]: I1126 01:01:01.420307 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401981-22f5z" event={"ID":"5e25beeb-e89d-4aed-9353-3755a0ef2efb","Type":"ContainerStarted","Data":"7408cef8df2bdbf17be98ca19cc721f5e9b05372ac7e9851e7c5b46da233775c"} Nov 26 01:01:01 crc kubenswrapper[4766]: I1126 01:01:01.420358 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401981-22f5z" event={"ID":"5e25beeb-e89d-4aed-9353-3755a0ef2efb","Type":"ContainerStarted","Data":"76bbcb4cc80aa9794e7edd741e0c40bdf735b0e376f43315875fd3c2649bf575"} Nov 26 01:01:01 crc kubenswrapper[4766]: I1126 01:01:01.440320 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29401981-22f5z" podStartSLOduration=1.440299113 podStartE2EDuration="1.440299113s" podCreationTimestamp="2025-11-26 01:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 01:01:01.431718782 +0000 UTC m=+2242.280489212" watchObservedRunningTime="2025-11-26 01:01:01.440299113 +0000 UTC m=+2242.289069553" Nov 26 01:01:04 crc kubenswrapper[4766]: I1126 01:01:04.456881 4766 generic.go:334] "Generic (PLEG): container finished" podID="5e25beeb-e89d-4aed-9353-3755a0ef2efb" containerID="7408cef8df2bdbf17be98ca19cc721f5e9b05372ac7e9851e7c5b46da233775c" exitCode=0 Nov 26 01:01:04 crc kubenswrapper[4766]: I1126 01:01:04.456986 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401981-22f5z" event={"ID":"5e25beeb-e89d-4aed-9353-3755a0ef2efb","Type":"ContainerDied","Data":"7408cef8df2bdbf17be98ca19cc721f5e9b05372ac7e9851e7c5b46da233775c"} Nov 26 01:01:05 crc kubenswrapper[4766]: I1126 01:01:05.929925 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401981-22f5z" Nov 26 01:01:06 crc kubenswrapper[4766]: I1126 01:01:06.009083 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5e25beeb-e89d-4aed-9353-3755a0ef2efb-fernet-keys\") pod \"5e25beeb-e89d-4aed-9353-3755a0ef2efb\" (UID: \"5e25beeb-e89d-4aed-9353-3755a0ef2efb\") " Nov 26 01:01:06 crc kubenswrapper[4766]: I1126 01:01:06.009131 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e25beeb-e89d-4aed-9353-3755a0ef2efb-combined-ca-bundle\") pod \"5e25beeb-e89d-4aed-9353-3755a0ef2efb\" (UID: \"5e25beeb-e89d-4aed-9353-3755a0ef2efb\") " Nov 26 01:01:06 crc kubenswrapper[4766]: I1126 01:01:06.009207 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e25beeb-e89d-4aed-9353-3755a0ef2efb-config-data\") pod \"5e25beeb-e89d-4aed-9353-3755a0ef2efb\" (UID: \"5e25beeb-e89d-4aed-9353-3755a0ef2efb\") " Nov 26 01:01:06 crc kubenswrapper[4766]: I1126 01:01:06.009238 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrb9k\" (UniqueName: \"kubernetes.io/projected/5e25beeb-e89d-4aed-9353-3755a0ef2efb-kube-api-access-rrb9k\") pod \"5e25beeb-e89d-4aed-9353-3755a0ef2efb\" (UID: \"5e25beeb-e89d-4aed-9353-3755a0ef2efb\") " Nov 26 01:01:06 crc kubenswrapper[4766]: I1126 01:01:06.015903 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e25beeb-e89d-4aed-9353-3755a0ef2efb-kube-api-access-rrb9k" (OuterVolumeSpecName: "kube-api-access-rrb9k") pod "5e25beeb-e89d-4aed-9353-3755a0ef2efb" (UID: "5e25beeb-e89d-4aed-9353-3755a0ef2efb"). InnerVolumeSpecName "kube-api-access-rrb9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:01:06 crc kubenswrapper[4766]: I1126 01:01:06.017403 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e25beeb-e89d-4aed-9353-3755a0ef2efb-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5e25beeb-e89d-4aed-9353-3755a0ef2efb" (UID: "5e25beeb-e89d-4aed-9353-3755a0ef2efb"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:01:06 crc kubenswrapper[4766]: I1126 01:01:06.047846 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e25beeb-e89d-4aed-9353-3755a0ef2efb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e25beeb-e89d-4aed-9353-3755a0ef2efb" (UID: "5e25beeb-e89d-4aed-9353-3755a0ef2efb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:01:06 crc kubenswrapper[4766]: I1126 01:01:06.090170 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e25beeb-e89d-4aed-9353-3755a0ef2efb-config-data" (OuterVolumeSpecName: "config-data") pod "5e25beeb-e89d-4aed-9353-3755a0ef2efb" (UID: "5e25beeb-e89d-4aed-9353-3755a0ef2efb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:01:06 crc kubenswrapper[4766]: I1126 01:01:06.111439 4766 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5e25beeb-e89d-4aed-9353-3755a0ef2efb-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 26 01:01:06 crc kubenswrapper[4766]: I1126 01:01:06.111473 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e25beeb-e89d-4aed-9353-3755a0ef2efb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:01:06 crc kubenswrapper[4766]: I1126 01:01:06.111486 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e25beeb-e89d-4aed-9353-3755a0ef2efb-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 01:01:06 crc kubenswrapper[4766]: I1126 01:01:06.111499 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrb9k\" (UniqueName: \"kubernetes.io/projected/5e25beeb-e89d-4aed-9353-3755a0ef2efb-kube-api-access-rrb9k\") on node \"crc\" DevicePath \"\"" Nov 26 01:01:06 crc kubenswrapper[4766]: I1126 01:01:06.484688 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401981-22f5z" event={"ID":"5e25beeb-e89d-4aed-9353-3755a0ef2efb","Type":"ContainerDied","Data":"76bbcb4cc80aa9794e7edd741e0c40bdf735b0e376f43315875fd3c2649bf575"} Nov 26 01:01:06 crc kubenswrapper[4766]: I1126 01:01:06.485052 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76bbcb4cc80aa9794e7edd741e0c40bdf735b0e376f43315875fd3c2649bf575" Nov 26 01:01:06 crc kubenswrapper[4766]: I1126 01:01:06.484765 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401981-22f5z" Nov 26 01:01:11 crc kubenswrapper[4766]: I1126 01:01:11.479834 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:01:11 crc kubenswrapper[4766]: I1126 01:01:11.480386 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:01:11 crc kubenswrapper[4766]: I1126 01:01:11.480439 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 01:01:11 crc kubenswrapper[4766]: I1126 01:01:11.481032 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 01:01:11 crc kubenswrapper[4766]: I1126 01:01:11.481087 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" gracePeriod=600 Nov 26 01:01:11 crc kubenswrapper[4766]: E1126 01:01:11.611869 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:01:12 crc kubenswrapper[4766]: I1126 01:01:12.579801 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" exitCode=0 Nov 26 01:01:12 crc kubenswrapper[4766]: I1126 01:01:12.579856 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c"} Nov 26 01:01:12 crc kubenswrapper[4766]: I1126 01:01:12.579890 4766 scope.go:117] "RemoveContainer" containerID="31c5fe6cf4a7b0b0322cd91369e8b92e06d2251c1e99313b96d5eb55cc36b4e6" Nov 26 01:01:12 crc kubenswrapper[4766]: I1126 01:01:12.580705 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:01:12 crc kubenswrapper[4766]: E1126 01:01:12.581006 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:01:26 crc kubenswrapper[4766]: I1126 01:01:26.826551 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:01:26 crc kubenswrapper[4766]: E1126 01:01:26.828461 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:01:27 crc kubenswrapper[4766]: I1126 01:01:27.037984 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-8vfjx"] Nov 26 01:01:27 crc kubenswrapper[4766]: I1126 01:01:27.046363 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-8vfjx"] Nov 26 01:01:27 crc kubenswrapper[4766]: I1126 01:01:27.839341 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75040a0b-cc48-4e77-9956-ca849299ff7b" path="/var/lib/kubelet/pods/75040a0b-cc48-4e77-9956-ca849299ff7b/volumes" Nov 26 01:01:28 crc kubenswrapper[4766]: I1126 01:01:28.300760 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hnspp"] Nov 26 01:01:28 crc kubenswrapper[4766]: E1126 01:01:28.301239 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e25beeb-e89d-4aed-9353-3755a0ef2efb" containerName="keystone-cron" Nov 26 01:01:28 crc kubenswrapper[4766]: I1126 01:01:28.301255 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e25beeb-e89d-4aed-9353-3755a0ef2efb" containerName="keystone-cron" Nov 26 01:01:28 crc kubenswrapper[4766]: I1126 01:01:28.301504 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e25beeb-e89d-4aed-9353-3755a0ef2efb" containerName="keystone-cron" Nov 26 01:01:28 crc kubenswrapper[4766]: I1126 01:01:28.303026 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hnspp" Nov 26 01:01:28 crc kubenswrapper[4766]: I1126 01:01:28.313132 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hnspp"] Nov 26 01:01:28 crc kubenswrapper[4766]: I1126 01:01:28.451012 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d325de0-2816-4577-a116-1c5eaa39c7a8-catalog-content\") pod \"redhat-operators-hnspp\" (UID: \"6d325de0-2816-4577-a116-1c5eaa39c7a8\") " pod="openshift-marketplace/redhat-operators-hnspp" Nov 26 01:01:28 crc kubenswrapper[4766]: I1126 01:01:28.451315 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d325de0-2816-4577-a116-1c5eaa39c7a8-utilities\") pod \"redhat-operators-hnspp\" (UID: \"6d325de0-2816-4577-a116-1c5eaa39c7a8\") " pod="openshift-marketplace/redhat-operators-hnspp" Nov 26 01:01:28 crc kubenswrapper[4766]: I1126 01:01:28.451688 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9282\" (UniqueName: \"kubernetes.io/projected/6d325de0-2816-4577-a116-1c5eaa39c7a8-kube-api-access-q9282\") pod \"redhat-operators-hnspp\" (UID: \"6d325de0-2816-4577-a116-1c5eaa39c7a8\") " pod="openshift-marketplace/redhat-operators-hnspp" Nov 26 01:01:28 crc kubenswrapper[4766]: I1126 01:01:28.553724 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d325de0-2816-4577-a116-1c5eaa39c7a8-utilities\") pod \"redhat-operators-hnspp\" (UID: \"6d325de0-2816-4577-a116-1c5eaa39c7a8\") " pod="openshift-marketplace/redhat-operators-hnspp" Nov 26 01:01:28 crc kubenswrapper[4766]: I1126 01:01:28.553870 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9282\" (UniqueName: \"kubernetes.io/projected/6d325de0-2816-4577-a116-1c5eaa39c7a8-kube-api-access-q9282\") pod \"redhat-operators-hnspp\" (UID: \"6d325de0-2816-4577-a116-1c5eaa39c7a8\") " pod="openshift-marketplace/redhat-operators-hnspp" Nov 26 01:01:28 crc kubenswrapper[4766]: I1126 01:01:28.553939 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d325de0-2816-4577-a116-1c5eaa39c7a8-catalog-content\") pod \"redhat-operators-hnspp\" (UID: \"6d325de0-2816-4577-a116-1c5eaa39c7a8\") " pod="openshift-marketplace/redhat-operators-hnspp" Nov 26 01:01:28 crc kubenswrapper[4766]: I1126 01:01:28.554235 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d325de0-2816-4577-a116-1c5eaa39c7a8-utilities\") pod \"redhat-operators-hnspp\" (UID: \"6d325de0-2816-4577-a116-1c5eaa39c7a8\") " pod="openshift-marketplace/redhat-operators-hnspp" Nov 26 01:01:28 crc kubenswrapper[4766]: I1126 01:01:28.554271 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d325de0-2816-4577-a116-1c5eaa39c7a8-catalog-content\") pod \"redhat-operators-hnspp\" (UID: \"6d325de0-2816-4577-a116-1c5eaa39c7a8\") " pod="openshift-marketplace/redhat-operators-hnspp" Nov 26 01:01:28 crc kubenswrapper[4766]: I1126 01:01:28.577032 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9282\" (UniqueName: \"kubernetes.io/projected/6d325de0-2816-4577-a116-1c5eaa39c7a8-kube-api-access-q9282\") pod \"redhat-operators-hnspp\" (UID: \"6d325de0-2816-4577-a116-1c5eaa39c7a8\") " pod="openshift-marketplace/redhat-operators-hnspp" Nov 26 01:01:28 crc kubenswrapper[4766]: I1126 01:01:28.631003 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hnspp" Nov 26 01:01:29 crc kubenswrapper[4766]: I1126 01:01:29.110142 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hnspp"] Nov 26 01:01:29 crc kubenswrapper[4766]: W1126 01:01:29.125156 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d325de0_2816_4577_a116_1c5eaa39c7a8.slice/crio-a85cc0109a691e2026d68f2fd87e4045df0195562c99d9228b228a05bed73cb7 WatchSource:0}: Error finding container a85cc0109a691e2026d68f2fd87e4045df0195562c99d9228b228a05bed73cb7: Status 404 returned error can't find the container with id a85cc0109a691e2026d68f2fd87e4045df0195562c99d9228b228a05bed73cb7 Nov 26 01:01:29 crc kubenswrapper[4766]: I1126 01:01:29.782690 4766 generic.go:334] "Generic (PLEG): container finished" podID="6d325de0-2816-4577-a116-1c5eaa39c7a8" containerID="363d3905a1f81e548237632cde938b1c6ba1ff9e8b1025b1157f98f923218bd8" exitCode=0 Nov 26 01:01:29 crc kubenswrapper[4766]: I1126 01:01:29.782782 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnspp" event={"ID":"6d325de0-2816-4577-a116-1c5eaa39c7a8","Type":"ContainerDied","Data":"363d3905a1f81e548237632cde938b1c6ba1ff9e8b1025b1157f98f923218bd8"} Nov 26 01:01:29 crc kubenswrapper[4766]: I1126 01:01:29.783089 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnspp" event={"ID":"6d325de0-2816-4577-a116-1c5eaa39c7a8","Type":"ContainerStarted","Data":"a85cc0109a691e2026d68f2fd87e4045df0195562c99d9228b228a05bed73cb7"} Nov 26 01:01:32 crc kubenswrapper[4766]: I1126 01:01:32.817424 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnspp" event={"ID":"6d325de0-2816-4577-a116-1c5eaa39c7a8","Type":"ContainerStarted","Data":"332c0533ce94893f74026bef1cf3b07b755ce5d0e0f6194a8726e3777850d962"} Nov 26 01:01:33 crc kubenswrapper[4766]: I1126 01:01:33.835754 4766 generic.go:334] "Generic (PLEG): container finished" podID="d456c126-aca3-47d2-b4fb-7ab506d116fb" containerID="22fff6cb2b6b099a0362c698eb539824d4f86ce71060d1a9f8b3e798f72badfa" exitCode=0 Nov 26 01:01:33 crc kubenswrapper[4766]: I1126 01:01:33.853598 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" event={"ID":"d456c126-aca3-47d2-b4fb-7ab506d116fb","Type":"ContainerDied","Data":"22fff6cb2b6b099a0362c698eb539824d4f86ce71060d1a9f8b3e798f72badfa"} Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.377639 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.548374 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d456c126-aca3-47d2-b4fb-7ab506d116fb-inventory\") pod \"d456c126-aca3-47d2-b4fb-7ab506d116fb\" (UID: \"d456c126-aca3-47d2-b4fb-7ab506d116fb\") " Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.548487 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d456c126-aca3-47d2-b4fb-7ab506d116fb-ssh-key\") pod \"d456c126-aca3-47d2-b4fb-7ab506d116fb\" (UID: \"d456c126-aca3-47d2-b4fb-7ab506d116fb\") " Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.548515 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mxbt\" (UniqueName: \"kubernetes.io/projected/d456c126-aca3-47d2-b4fb-7ab506d116fb-kube-api-access-8mxbt\") pod \"d456c126-aca3-47d2-b4fb-7ab506d116fb\" (UID: \"d456c126-aca3-47d2-b4fb-7ab506d116fb\") " Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.548557 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d456c126-aca3-47d2-b4fb-7ab506d116fb-ovn-combined-ca-bundle\") pod \"d456c126-aca3-47d2-b4fb-7ab506d116fb\" (UID: \"d456c126-aca3-47d2-b4fb-7ab506d116fb\") " Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.548590 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d456c126-aca3-47d2-b4fb-7ab506d116fb-ovncontroller-config-0\") pod \"d456c126-aca3-47d2-b4fb-7ab506d116fb\" (UID: \"d456c126-aca3-47d2-b4fb-7ab506d116fb\") " Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.584150 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d456c126-aca3-47d2-b4fb-7ab506d116fb-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d456c126-aca3-47d2-b4fb-7ab506d116fb" (UID: "d456c126-aca3-47d2-b4fb-7ab506d116fb"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.590939 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d456c126-aca3-47d2-b4fb-7ab506d116fb-kube-api-access-8mxbt" (OuterVolumeSpecName: "kube-api-access-8mxbt") pod "d456c126-aca3-47d2-b4fb-7ab506d116fb" (UID: "d456c126-aca3-47d2-b4fb-7ab506d116fb"). InnerVolumeSpecName "kube-api-access-8mxbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.621075 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d456c126-aca3-47d2-b4fb-7ab506d116fb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d456c126-aca3-47d2-b4fb-7ab506d116fb" (UID: "d456c126-aca3-47d2-b4fb-7ab506d116fb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.628525 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d456c126-aca3-47d2-b4fb-7ab506d116fb-inventory" (OuterVolumeSpecName: "inventory") pod "d456c126-aca3-47d2-b4fb-7ab506d116fb" (UID: "d456c126-aca3-47d2-b4fb-7ab506d116fb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.628603 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d456c126-aca3-47d2-b4fb-7ab506d116fb-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "d456c126-aca3-47d2-b4fb-7ab506d116fb" (UID: "d456c126-aca3-47d2-b4fb-7ab506d116fb"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.651510 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d456c126-aca3-47d2-b4fb-7ab506d116fb-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.651546 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mxbt\" (UniqueName: \"kubernetes.io/projected/d456c126-aca3-47d2-b4fb-7ab506d116fb-kube-api-access-8mxbt\") on node \"crc\" DevicePath \"\"" Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.651557 4766 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d456c126-aca3-47d2-b4fb-7ab506d116fb-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.651566 4766 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d456c126-aca3-47d2-b4fb-7ab506d116fb-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.651578 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d456c126-aca3-47d2-b4fb-7ab506d116fb-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.860345 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" event={"ID":"d456c126-aca3-47d2-b4fb-7ab506d116fb","Type":"ContainerDied","Data":"c495c01a66ab8a57acbe084b59d9e5e3555566ad8f06d0b17de2c95a1ae7ad1d"} Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.860404 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c495c01a66ab8a57acbe084b59d9e5e3555566ad8f06d0b17de2c95a1ae7ad1d" Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.860521 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z" Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.989156 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm"] Nov 26 01:01:35 crc kubenswrapper[4766]: E1126 01:01:35.989983 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d456c126-aca3-47d2-b4fb-7ab506d116fb" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.990015 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d456c126-aca3-47d2-b4fb-7ab506d116fb" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.990367 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="d456c126-aca3-47d2-b4fb-7ab506d116fb" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.991684 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.993586 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.994057 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.994876 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.995009 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:01:35 crc kubenswrapper[4766]: I1126 01:01:35.995097 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:01:36 crc kubenswrapper[4766]: I1126 01:01:36.004333 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm"] Nov 26 01:01:36 crc kubenswrapper[4766]: I1126 01:01:36.163108 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm\" (UID: \"d2a8f94b-256b-4c29-83de-24b15b0284f2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" Nov 26 01:01:36 crc kubenswrapper[4766]: I1126 01:01:36.163945 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm\" (UID: \"d2a8f94b-256b-4c29-83de-24b15b0284f2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" Nov 26 01:01:36 crc kubenswrapper[4766]: I1126 01:01:36.164320 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjgc7\" (UniqueName: \"kubernetes.io/projected/d2a8f94b-256b-4c29-83de-24b15b0284f2-kube-api-access-zjgc7\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm\" (UID: \"d2a8f94b-256b-4c29-83de-24b15b0284f2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" Nov 26 01:01:36 crc kubenswrapper[4766]: I1126 01:01:36.164435 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm\" (UID: \"d2a8f94b-256b-4c29-83de-24b15b0284f2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" Nov 26 01:01:36 crc kubenswrapper[4766]: I1126 01:01:36.164556 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm\" (UID: \"d2a8f94b-256b-4c29-83de-24b15b0284f2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" Nov 26 01:01:36 crc kubenswrapper[4766]: I1126 01:01:36.266505 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjgc7\" (UniqueName: \"kubernetes.io/projected/d2a8f94b-256b-4c29-83de-24b15b0284f2-kube-api-access-zjgc7\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm\" (UID: \"d2a8f94b-256b-4c29-83de-24b15b0284f2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" Nov 26 01:01:36 crc kubenswrapper[4766]: I1126 01:01:36.266575 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm\" (UID: \"d2a8f94b-256b-4c29-83de-24b15b0284f2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" Nov 26 01:01:36 crc kubenswrapper[4766]: I1126 01:01:36.266633 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm\" (UID: \"d2a8f94b-256b-4c29-83de-24b15b0284f2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" Nov 26 01:01:36 crc kubenswrapper[4766]: I1126 01:01:36.266699 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm\" (UID: \"d2a8f94b-256b-4c29-83de-24b15b0284f2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" Nov 26 01:01:36 crc kubenswrapper[4766]: I1126 01:01:36.266751 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm\" (UID: \"d2a8f94b-256b-4c29-83de-24b15b0284f2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" Nov 26 01:01:36 crc kubenswrapper[4766]: I1126 01:01:36.271453 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm\" (UID: \"d2a8f94b-256b-4c29-83de-24b15b0284f2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" Nov 26 01:01:36 crc kubenswrapper[4766]: I1126 01:01:36.271951 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm\" (UID: \"d2a8f94b-256b-4c29-83de-24b15b0284f2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" Nov 26 01:01:36 crc kubenswrapper[4766]: I1126 01:01:36.271978 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm\" (UID: \"d2a8f94b-256b-4c29-83de-24b15b0284f2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" Nov 26 01:01:36 crc kubenswrapper[4766]: I1126 01:01:36.272738 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm\" (UID: \"d2a8f94b-256b-4c29-83de-24b15b0284f2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" Nov 26 01:01:36 crc kubenswrapper[4766]: I1126 01:01:36.300962 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjgc7\" (UniqueName: \"kubernetes.io/projected/d2a8f94b-256b-4c29-83de-24b15b0284f2-kube-api-access-zjgc7\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm\" (UID: \"d2a8f94b-256b-4c29-83de-24b15b0284f2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" Nov 26 01:01:36 crc kubenswrapper[4766]: I1126 01:01:36.311951 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" Nov 26 01:01:36 crc kubenswrapper[4766]: I1126 01:01:36.957456 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm"] Nov 26 01:01:37 crc kubenswrapper[4766]: W1126 01:01:37.006969 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2a8f94b_256b_4c29_83de_24b15b0284f2.slice/crio-92099e29bd03eec20e760155aa2ec60f9ffbe5284bb887534d30080e6fc67308 WatchSource:0}: Error finding container 92099e29bd03eec20e760155aa2ec60f9ffbe5284bb887534d30080e6fc67308: Status 404 returned error can't find the container with id 92099e29bd03eec20e760155aa2ec60f9ffbe5284bb887534d30080e6fc67308 Nov 26 01:01:37 crc kubenswrapper[4766]: I1126 01:01:37.826866 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:01:37 crc kubenswrapper[4766]: E1126 01:01:37.827536 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:01:37 crc kubenswrapper[4766]: I1126 01:01:37.882687 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" event={"ID":"d2a8f94b-256b-4c29-83de-24b15b0284f2","Type":"ContainerStarted","Data":"45a3eee6d40c5c70f429b1e60ea753207c66d3e90e880a0b3db180839ab48d8b"} Nov 26 01:01:37 crc kubenswrapper[4766]: I1126 01:01:37.882941 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" event={"ID":"d2a8f94b-256b-4c29-83de-24b15b0284f2","Type":"ContainerStarted","Data":"92099e29bd03eec20e760155aa2ec60f9ffbe5284bb887534d30080e6fc67308"} Nov 26 01:01:37 crc kubenswrapper[4766]: I1126 01:01:37.887705 4766 generic.go:334] "Generic (PLEG): container finished" podID="6d325de0-2816-4577-a116-1c5eaa39c7a8" containerID="332c0533ce94893f74026bef1cf3b07b755ce5d0e0f6194a8726e3777850d962" exitCode=0 Nov 26 01:01:37 crc kubenswrapper[4766]: I1126 01:01:37.887816 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnspp" event={"ID":"6d325de0-2816-4577-a116-1c5eaa39c7a8","Type":"ContainerDied","Data":"332c0533ce94893f74026bef1cf3b07b755ce5d0e0f6194a8726e3777850d962"} Nov 26 01:01:37 crc kubenswrapper[4766]: I1126 01:01:37.904581 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" podStartSLOduration=2.398173203 podStartE2EDuration="2.904564628s" podCreationTimestamp="2025-11-26 01:01:35 +0000 UTC" firstStartedPulling="2025-11-26 01:01:37.011212399 +0000 UTC m=+2277.859982839" lastFinishedPulling="2025-11-26 01:01:37.517603804 +0000 UTC m=+2278.366374264" observedRunningTime="2025-11-26 01:01:37.897563285 +0000 UTC m=+2278.746333735" watchObservedRunningTime="2025-11-26 01:01:37.904564628 +0000 UTC m=+2278.753335058" Nov 26 01:01:38 crc kubenswrapper[4766]: I1126 01:01:38.924807 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnspp" event={"ID":"6d325de0-2816-4577-a116-1c5eaa39c7a8","Type":"ContainerStarted","Data":"66b20891e51e36defe7cce89abeaf6aa0898d50ecac72094db661b667dc7562e"} Nov 26 01:01:39 crc kubenswrapper[4766]: I1126 01:01:39.017047 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hnspp" podStartSLOduration=2.46925874 podStartE2EDuration="11.017025906s" podCreationTimestamp="2025-11-26 01:01:28 +0000 UTC" firstStartedPulling="2025-11-26 01:01:29.785149055 +0000 UTC m=+2270.633919535" lastFinishedPulling="2025-11-26 01:01:38.332916241 +0000 UTC m=+2279.181686701" observedRunningTime="2025-11-26 01:01:39.010088575 +0000 UTC m=+2279.858859005" watchObservedRunningTime="2025-11-26 01:01:39.017025906 +0000 UTC m=+2279.865796336" Nov 26 01:01:40 crc kubenswrapper[4766]: I1126 01:01:40.051180 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-tz998"] Nov 26 01:01:40 crc kubenswrapper[4766]: I1126 01:01:40.061462 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-tz998"] Nov 26 01:01:41 crc kubenswrapper[4766]: I1126 01:01:41.839330 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d6086e9-30aa-4a56-9233-182c811c85c2" path="/var/lib/kubelet/pods/6d6086e9-30aa-4a56-9233-182c811c85c2/volumes" Nov 26 01:01:48 crc kubenswrapper[4766]: I1126 01:01:48.631845 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hnspp" Nov 26 01:01:48 crc kubenswrapper[4766]: I1126 01:01:48.632621 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hnspp" Nov 26 01:01:48 crc kubenswrapper[4766]: I1126 01:01:48.709323 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hnspp" Nov 26 01:01:49 crc kubenswrapper[4766]: I1126 01:01:49.132606 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hnspp" Nov 26 01:01:49 crc kubenswrapper[4766]: I1126 01:01:49.253923 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hnspp"] Nov 26 01:01:49 crc kubenswrapper[4766]: I1126 01:01:49.841798 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:01:49 crc kubenswrapper[4766]: E1126 01:01:49.842396 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:01:51 crc kubenswrapper[4766]: I1126 01:01:51.055318 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hnspp" podUID="6d325de0-2816-4577-a116-1c5eaa39c7a8" containerName="registry-server" containerID="cri-o://66b20891e51e36defe7cce89abeaf6aa0898d50ecac72094db661b667dc7562e" gracePeriod=2 Nov 26 01:01:51 crc kubenswrapper[4766]: I1126 01:01:51.667843 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hnspp" Nov 26 01:01:51 crc kubenswrapper[4766]: I1126 01:01:51.781436 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9282\" (UniqueName: \"kubernetes.io/projected/6d325de0-2816-4577-a116-1c5eaa39c7a8-kube-api-access-q9282\") pod \"6d325de0-2816-4577-a116-1c5eaa39c7a8\" (UID: \"6d325de0-2816-4577-a116-1c5eaa39c7a8\") " Nov 26 01:01:51 crc kubenswrapper[4766]: I1126 01:01:51.781630 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d325de0-2816-4577-a116-1c5eaa39c7a8-catalog-content\") pod \"6d325de0-2816-4577-a116-1c5eaa39c7a8\" (UID: \"6d325de0-2816-4577-a116-1c5eaa39c7a8\") " Nov 26 01:01:51 crc kubenswrapper[4766]: I1126 01:01:51.781754 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d325de0-2816-4577-a116-1c5eaa39c7a8-utilities\") pod \"6d325de0-2816-4577-a116-1c5eaa39c7a8\" (UID: \"6d325de0-2816-4577-a116-1c5eaa39c7a8\") " Nov 26 01:01:51 crc kubenswrapper[4766]: I1126 01:01:51.783294 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d325de0-2816-4577-a116-1c5eaa39c7a8-utilities" (OuterVolumeSpecName: "utilities") pod "6d325de0-2816-4577-a116-1c5eaa39c7a8" (UID: "6d325de0-2816-4577-a116-1c5eaa39c7a8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:01:51 crc kubenswrapper[4766]: I1126 01:01:51.796020 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d325de0-2816-4577-a116-1c5eaa39c7a8-kube-api-access-q9282" (OuterVolumeSpecName: "kube-api-access-q9282") pod "6d325de0-2816-4577-a116-1c5eaa39c7a8" (UID: "6d325de0-2816-4577-a116-1c5eaa39c7a8"). InnerVolumeSpecName "kube-api-access-q9282". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:01:51 crc kubenswrapper[4766]: I1126 01:01:51.883103 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d325de0-2816-4577-a116-1c5eaa39c7a8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d325de0-2816-4577-a116-1c5eaa39c7a8" (UID: "6d325de0-2816-4577-a116-1c5eaa39c7a8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:01:51 crc kubenswrapper[4766]: I1126 01:01:51.884470 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d325de0-2816-4577-a116-1c5eaa39c7a8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 01:01:51 crc kubenswrapper[4766]: I1126 01:01:51.884576 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d325de0-2816-4577-a116-1c5eaa39c7a8-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 01:01:51 crc kubenswrapper[4766]: I1126 01:01:51.884642 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9282\" (UniqueName: \"kubernetes.io/projected/6d325de0-2816-4577-a116-1c5eaa39c7a8-kube-api-access-q9282\") on node \"crc\" DevicePath \"\"" Nov 26 01:01:52 crc kubenswrapper[4766]: I1126 01:01:52.069885 4766 generic.go:334] "Generic (PLEG): container finished" podID="6d325de0-2816-4577-a116-1c5eaa39c7a8" containerID="66b20891e51e36defe7cce89abeaf6aa0898d50ecac72094db661b667dc7562e" exitCode=0 Nov 26 01:01:52 crc kubenswrapper[4766]: I1126 01:01:52.069936 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnspp" event={"ID":"6d325de0-2816-4577-a116-1c5eaa39c7a8","Type":"ContainerDied","Data":"66b20891e51e36defe7cce89abeaf6aa0898d50ecac72094db661b667dc7562e"} Nov 26 01:01:52 crc kubenswrapper[4766]: I1126 01:01:52.069950 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hnspp" Nov 26 01:01:52 crc kubenswrapper[4766]: I1126 01:01:52.069966 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnspp" event={"ID":"6d325de0-2816-4577-a116-1c5eaa39c7a8","Type":"ContainerDied","Data":"a85cc0109a691e2026d68f2fd87e4045df0195562c99d9228b228a05bed73cb7"} Nov 26 01:01:52 crc kubenswrapper[4766]: I1126 01:01:52.069988 4766 scope.go:117] "RemoveContainer" containerID="66b20891e51e36defe7cce89abeaf6aa0898d50ecac72094db661b667dc7562e" Nov 26 01:01:52 crc kubenswrapper[4766]: I1126 01:01:52.097954 4766 scope.go:117] "RemoveContainer" containerID="332c0533ce94893f74026bef1cf3b07b755ce5d0e0f6194a8726e3777850d962" Nov 26 01:01:52 crc kubenswrapper[4766]: I1126 01:01:52.108755 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hnspp"] Nov 26 01:01:52 crc kubenswrapper[4766]: I1126 01:01:52.119983 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hnspp"] Nov 26 01:01:52 crc kubenswrapper[4766]: I1126 01:01:52.148149 4766 scope.go:117] "RemoveContainer" containerID="363d3905a1f81e548237632cde938b1c6ba1ff9e8b1025b1157f98f923218bd8" Nov 26 01:01:52 crc kubenswrapper[4766]: I1126 01:01:52.199052 4766 scope.go:117] "RemoveContainer" containerID="66b20891e51e36defe7cce89abeaf6aa0898d50ecac72094db661b667dc7562e" Nov 26 01:01:52 crc kubenswrapper[4766]: E1126 01:01:52.199472 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66b20891e51e36defe7cce89abeaf6aa0898d50ecac72094db661b667dc7562e\": container with ID starting with 66b20891e51e36defe7cce89abeaf6aa0898d50ecac72094db661b667dc7562e not found: ID does not exist" containerID="66b20891e51e36defe7cce89abeaf6aa0898d50ecac72094db661b667dc7562e" Nov 26 01:01:52 crc kubenswrapper[4766]: I1126 01:01:52.199507 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66b20891e51e36defe7cce89abeaf6aa0898d50ecac72094db661b667dc7562e"} err="failed to get container status \"66b20891e51e36defe7cce89abeaf6aa0898d50ecac72094db661b667dc7562e\": rpc error: code = NotFound desc = could not find container \"66b20891e51e36defe7cce89abeaf6aa0898d50ecac72094db661b667dc7562e\": container with ID starting with 66b20891e51e36defe7cce89abeaf6aa0898d50ecac72094db661b667dc7562e not found: ID does not exist" Nov 26 01:01:52 crc kubenswrapper[4766]: I1126 01:01:52.199523 4766 scope.go:117] "RemoveContainer" containerID="332c0533ce94893f74026bef1cf3b07b755ce5d0e0f6194a8726e3777850d962" Nov 26 01:01:52 crc kubenswrapper[4766]: E1126 01:01:52.199842 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"332c0533ce94893f74026bef1cf3b07b755ce5d0e0f6194a8726e3777850d962\": container with ID starting with 332c0533ce94893f74026bef1cf3b07b755ce5d0e0f6194a8726e3777850d962 not found: ID does not exist" containerID="332c0533ce94893f74026bef1cf3b07b755ce5d0e0f6194a8726e3777850d962" Nov 26 01:01:52 crc kubenswrapper[4766]: I1126 01:01:52.199896 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"332c0533ce94893f74026bef1cf3b07b755ce5d0e0f6194a8726e3777850d962"} err="failed to get container status \"332c0533ce94893f74026bef1cf3b07b755ce5d0e0f6194a8726e3777850d962\": rpc error: code = NotFound desc = could not find container \"332c0533ce94893f74026bef1cf3b07b755ce5d0e0f6194a8726e3777850d962\": container with ID starting with 332c0533ce94893f74026bef1cf3b07b755ce5d0e0f6194a8726e3777850d962 not found: ID does not exist" Nov 26 01:01:52 crc kubenswrapper[4766]: I1126 01:01:52.199926 4766 scope.go:117] "RemoveContainer" containerID="363d3905a1f81e548237632cde938b1c6ba1ff9e8b1025b1157f98f923218bd8" Nov 26 01:01:52 crc kubenswrapper[4766]: E1126 01:01:52.200230 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"363d3905a1f81e548237632cde938b1c6ba1ff9e8b1025b1157f98f923218bd8\": container with ID starting with 363d3905a1f81e548237632cde938b1c6ba1ff9e8b1025b1157f98f923218bd8 not found: ID does not exist" containerID="363d3905a1f81e548237632cde938b1c6ba1ff9e8b1025b1157f98f923218bd8" Nov 26 01:01:52 crc kubenswrapper[4766]: I1126 01:01:52.200272 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"363d3905a1f81e548237632cde938b1c6ba1ff9e8b1025b1157f98f923218bd8"} err="failed to get container status \"363d3905a1f81e548237632cde938b1c6ba1ff9e8b1025b1157f98f923218bd8\": rpc error: code = NotFound desc = could not find container \"363d3905a1f81e548237632cde938b1c6ba1ff9e8b1025b1157f98f923218bd8\": container with ID starting with 363d3905a1f81e548237632cde938b1c6ba1ff9e8b1025b1157f98f923218bd8 not found: ID does not exist" Nov 26 01:01:53 crc kubenswrapper[4766]: I1126 01:01:53.859838 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d325de0-2816-4577-a116-1c5eaa39c7a8" path="/var/lib/kubelet/pods/6d325de0-2816-4577-a116-1c5eaa39c7a8/volumes" Nov 26 01:01:59 crc kubenswrapper[4766]: I1126 01:01:59.414725 4766 scope.go:117] "RemoveContainer" containerID="b0221c71033eb66079348bb9d44485237f3dfeeeb515394898f7d0d95d0b38bd" Nov 26 01:01:59 crc kubenswrapper[4766]: I1126 01:01:59.458147 4766 scope.go:117] "RemoveContainer" containerID="17a7b3c12e36d0510101c272211eeed5cbabed5ed57c708c65189e195d9a53d8" Nov 26 01:02:03 crc kubenswrapper[4766]: I1126 01:02:03.829257 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:02:03 crc kubenswrapper[4766]: E1126 01:02:03.830556 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:02:14 crc kubenswrapper[4766]: I1126 01:02:14.826448 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:02:14 crc kubenswrapper[4766]: E1126 01:02:14.827176 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:02:26 crc kubenswrapper[4766]: I1126 01:02:26.826912 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:02:26 crc kubenswrapper[4766]: E1126 01:02:26.830335 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:02:40 crc kubenswrapper[4766]: I1126 01:02:40.827002 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:02:40 crc kubenswrapper[4766]: E1126 01:02:40.827941 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:02:53 crc kubenswrapper[4766]: I1126 01:02:53.827957 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:02:53 crc kubenswrapper[4766]: E1126 01:02:53.829464 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:03:04 crc kubenswrapper[4766]: I1126 01:03:04.827446 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:03:04 crc kubenswrapper[4766]: E1126 01:03:04.828239 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:03:16 crc kubenswrapper[4766]: I1126 01:03:16.827076 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:03:16 crc kubenswrapper[4766]: E1126 01:03:16.827812 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:03:28 crc kubenswrapper[4766]: I1126 01:03:28.830213 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:03:28 crc kubenswrapper[4766]: E1126 01:03:28.831178 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:03:43 crc kubenswrapper[4766]: I1126 01:03:43.827614 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:03:43 crc kubenswrapper[4766]: E1126 01:03:43.828673 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:03:57 crc kubenswrapper[4766]: I1126 01:03:57.831548 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:03:57 crc kubenswrapper[4766]: E1126 01:03:57.832647 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:04:11 crc kubenswrapper[4766]: I1126 01:04:11.826831 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:04:11 crc kubenswrapper[4766]: E1126 01:04:11.827691 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:04:22 crc kubenswrapper[4766]: I1126 01:04:22.828125 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:04:22 crc kubenswrapper[4766]: E1126 01:04:22.829340 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:04:33 crc kubenswrapper[4766]: I1126 01:04:33.827430 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:04:33 crc kubenswrapper[4766]: E1126 01:04:33.828171 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:04:48 crc kubenswrapper[4766]: I1126 01:04:48.827806 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:04:48 crc kubenswrapper[4766]: E1126 01:04:48.829158 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:05:00 crc kubenswrapper[4766]: I1126 01:05:00.827630 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:05:00 crc kubenswrapper[4766]: E1126 01:05:00.829929 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:05:13 crc kubenswrapper[4766]: I1126 01:05:13.827116 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:05:13 crc kubenswrapper[4766]: E1126 01:05:13.831279 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:05:28 crc kubenswrapper[4766]: I1126 01:05:28.827448 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:05:28 crc kubenswrapper[4766]: E1126 01:05:28.828529 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:05:40 crc kubenswrapper[4766]: I1126 01:05:40.827015 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:05:40 crc kubenswrapper[4766]: E1126 01:05:40.827718 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:05:51 crc kubenswrapper[4766]: I1126 01:05:51.827963 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:05:51 crc kubenswrapper[4766]: E1126 01:05:51.828972 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:06:04 crc kubenswrapper[4766]: I1126 01:06:04.827880 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:06:04 crc kubenswrapper[4766]: E1126 01:06:04.828925 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:06:16 crc kubenswrapper[4766]: I1126 01:06:16.826952 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:06:17 crc kubenswrapper[4766]: I1126 01:06:17.528917 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"3f97e605694d4755ce21053291eee5a113d9b02cc30e9ddf766f118e2989829b"} Nov 26 01:06:29 crc kubenswrapper[4766]: I1126 01:06:29.674372 4766 generic.go:334] "Generic (PLEG): container finished" podID="d2a8f94b-256b-4c29-83de-24b15b0284f2" containerID="45a3eee6d40c5c70f429b1e60ea753207c66d3e90e880a0b3db180839ab48d8b" exitCode=0 Nov 26 01:06:29 crc kubenswrapper[4766]: I1126 01:06:29.674484 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" event={"ID":"d2a8f94b-256b-4c29-83de-24b15b0284f2","Type":"ContainerDied","Data":"45a3eee6d40c5c70f429b1e60ea753207c66d3e90e880a0b3db180839ab48d8b"} Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.237226 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.262522 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-ssh-key\") pod \"d2a8f94b-256b-4c29-83de-24b15b0284f2\" (UID: \"d2a8f94b-256b-4c29-83de-24b15b0284f2\") " Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.262631 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-libvirt-combined-ca-bundle\") pod \"d2a8f94b-256b-4c29-83de-24b15b0284f2\" (UID: \"d2a8f94b-256b-4c29-83de-24b15b0284f2\") " Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.262750 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjgc7\" (UniqueName: \"kubernetes.io/projected/d2a8f94b-256b-4c29-83de-24b15b0284f2-kube-api-access-zjgc7\") pod \"d2a8f94b-256b-4c29-83de-24b15b0284f2\" (UID: \"d2a8f94b-256b-4c29-83de-24b15b0284f2\") " Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.262781 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-inventory\") pod \"d2a8f94b-256b-4c29-83de-24b15b0284f2\" (UID: \"d2a8f94b-256b-4c29-83de-24b15b0284f2\") " Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.263004 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-libvirt-secret-0\") pod \"d2a8f94b-256b-4c29-83de-24b15b0284f2\" (UID: \"d2a8f94b-256b-4c29-83de-24b15b0284f2\") " Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.269749 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "d2a8f94b-256b-4c29-83de-24b15b0284f2" (UID: "d2a8f94b-256b-4c29-83de-24b15b0284f2"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.281439 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2a8f94b-256b-4c29-83de-24b15b0284f2-kube-api-access-zjgc7" (OuterVolumeSpecName: "kube-api-access-zjgc7") pod "d2a8f94b-256b-4c29-83de-24b15b0284f2" (UID: "d2a8f94b-256b-4c29-83de-24b15b0284f2"). InnerVolumeSpecName "kube-api-access-zjgc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.297028 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-inventory" (OuterVolumeSpecName: "inventory") pod "d2a8f94b-256b-4c29-83de-24b15b0284f2" (UID: "d2a8f94b-256b-4c29-83de-24b15b0284f2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.312395 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d2a8f94b-256b-4c29-83de-24b15b0284f2" (UID: "d2a8f94b-256b-4c29-83de-24b15b0284f2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.335773 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "d2a8f94b-256b-4c29-83de-24b15b0284f2" (UID: "d2a8f94b-256b-4c29-83de-24b15b0284f2"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.366245 4766 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.366461 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.366536 4766 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.366602 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjgc7\" (UniqueName: \"kubernetes.io/projected/d2a8f94b-256b-4c29-83de-24b15b0284f2-kube-api-access-zjgc7\") on node \"crc\" DevicePath \"\"" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.366689 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2a8f94b-256b-4c29-83de-24b15b0284f2-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.698865 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" event={"ID":"d2a8f94b-256b-4c29-83de-24b15b0284f2","Type":"ContainerDied","Data":"92099e29bd03eec20e760155aa2ec60f9ffbe5284bb887534d30080e6fc67308"} Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.698903 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92099e29bd03eec20e760155aa2ec60f9ffbe5284bb887534d30080e6fc67308" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.698980 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.885598 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx"] Nov 26 01:06:31 crc kubenswrapper[4766]: E1126 01:06:31.886029 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2a8f94b-256b-4c29-83de-24b15b0284f2" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.886050 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2a8f94b-256b-4c29-83de-24b15b0284f2" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 26 01:06:31 crc kubenswrapper[4766]: E1126 01:06:31.886105 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d325de0-2816-4577-a116-1c5eaa39c7a8" containerName="extract-content" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.886116 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d325de0-2816-4577-a116-1c5eaa39c7a8" containerName="extract-content" Nov 26 01:06:31 crc kubenswrapper[4766]: E1126 01:06:31.886138 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d325de0-2816-4577-a116-1c5eaa39c7a8" containerName="extract-utilities" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.886147 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d325de0-2816-4577-a116-1c5eaa39c7a8" containerName="extract-utilities" Nov 26 01:06:31 crc kubenswrapper[4766]: E1126 01:06:31.886167 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d325de0-2816-4577-a116-1c5eaa39c7a8" containerName="registry-server" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.886176 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d325de0-2816-4577-a116-1c5eaa39c7a8" containerName="registry-server" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.886433 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d325de0-2816-4577-a116-1c5eaa39c7a8" containerName="registry-server" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.886455 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2a8f94b-256b-4c29-83de-24b15b0284f2" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.887320 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx"] Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.887414 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.897224 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.898409 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.898812 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.898966 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.901308 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.981314 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smpcp\" (UniqueName: \"kubernetes.io/projected/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-kube-api-access-smpcp\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.981562 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.981666 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.981820 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.981865 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.981902 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:31 crc kubenswrapper[4766]: I1126 01:06:31.982004 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:32 crc kubenswrapper[4766]: I1126 01:06:32.084111 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smpcp\" (UniqueName: \"kubernetes.io/projected/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-kube-api-access-smpcp\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:32 crc kubenswrapper[4766]: I1126 01:06:32.084227 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:32 crc kubenswrapper[4766]: I1126 01:06:32.084257 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:32 crc kubenswrapper[4766]: I1126 01:06:32.084319 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:32 crc kubenswrapper[4766]: I1126 01:06:32.084348 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:32 crc kubenswrapper[4766]: I1126 01:06:32.084372 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:32 crc kubenswrapper[4766]: I1126 01:06:32.084419 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:32 crc kubenswrapper[4766]: I1126 01:06:32.088522 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:32 crc kubenswrapper[4766]: I1126 01:06:32.089712 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:32 crc kubenswrapper[4766]: I1126 01:06:32.089790 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:32 crc kubenswrapper[4766]: I1126 01:06:32.090853 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:32 crc kubenswrapper[4766]: I1126 01:06:32.093051 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:32 crc kubenswrapper[4766]: I1126 01:06:32.094169 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:32 crc kubenswrapper[4766]: I1126 01:06:32.111705 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smpcp\" (UniqueName: \"kubernetes.io/projected/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-kube-api-access-smpcp\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:32 crc kubenswrapper[4766]: I1126 01:06:32.223978 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:06:32 crc kubenswrapper[4766]: W1126 01:06:32.778379 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6acb6255_ec9f_4c75_ac41_cb7e1ec842a9.slice/crio-59243fb0332cc19e487b24f13e5cb41be615c3c1f338ac837cc737a3f6735c6d WatchSource:0}: Error finding container 59243fb0332cc19e487b24f13e5cb41be615c3c1f338ac837cc737a3f6735c6d: Status 404 returned error can't find the container with id 59243fb0332cc19e487b24f13e5cb41be615c3c1f338ac837cc737a3f6735c6d Nov 26 01:06:32 crc kubenswrapper[4766]: I1126 01:06:32.780973 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx"] Nov 26 01:06:32 crc kubenswrapper[4766]: I1126 01:06:32.781027 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 01:06:33 crc kubenswrapper[4766]: I1126 01:06:33.721312 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" event={"ID":"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9","Type":"ContainerStarted","Data":"6440a835d32a787feaf06d14ed60fa3af91eec1a1e2e84f51ae6e6c6caad0bcd"} Nov 26 01:06:33 crc kubenswrapper[4766]: I1126 01:06:33.721718 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" event={"ID":"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9","Type":"ContainerStarted","Data":"59243fb0332cc19e487b24f13e5cb41be615c3c1f338ac837cc737a3f6735c6d"} Nov 26 01:06:33 crc kubenswrapper[4766]: I1126 01:06:33.761264 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" podStartSLOduration=2.246860519 podStartE2EDuration="2.761227914s" podCreationTimestamp="2025-11-26 01:06:31 +0000 UTC" firstStartedPulling="2025-11-26 01:06:32.780811582 +0000 UTC m=+2573.629582012" lastFinishedPulling="2025-11-26 01:06:33.295178967 +0000 UTC m=+2574.143949407" observedRunningTime="2025-11-26 01:06:33.745382584 +0000 UTC m=+2574.594153044" watchObservedRunningTime="2025-11-26 01:06:33.761227914 +0000 UTC m=+2574.609998394" Nov 26 01:08:41 crc kubenswrapper[4766]: I1126 01:08:41.479463 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:08:41 crc kubenswrapper[4766]: I1126 01:08:41.480904 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:09:11 crc kubenswrapper[4766]: I1126 01:09:11.482691 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:09:11 crc kubenswrapper[4766]: I1126 01:09:11.484770 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:09:27 crc kubenswrapper[4766]: I1126 01:09:27.981920 4766 generic.go:334] "Generic (PLEG): container finished" podID="6acb6255-ec9f-4c75-ac41-cb7e1ec842a9" containerID="6440a835d32a787feaf06d14ed60fa3af91eec1a1e2e84f51ae6e6c6caad0bcd" exitCode=0 Nov 26 01:09:27 crc kubenswrapper[4766]: I1126 01:09:27.982040 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" event={"ID":"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9","Type":"ContainerDied","Data":"6440a835d32a787feaf06d14ed60fa3af91eec1a1e2e84f51ae6e6c6caad0bcd"} Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.561084 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.679720 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-telemetry-combined-ca-bundle\") pod \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.679808 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ceilometer-compute-config-data-1\") pod \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.681122 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smpcp\" (UniqueName: \"kubernetes.io/projected/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-kube-api-access-smpcp\") pod \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.681178 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ceilometer-compute-config-data-2\") pod \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.681285 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ssh-key\") pod \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.681376 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-inventory\") pod \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.681473 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ceilometer-compute-config-data-0\") pod \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\" (UID: \"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9\") " Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.691526 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "6acb6255-ec9f-4c75-ac41-cb7e1ec842a9" (UID: "6acb6255-ec9f-4c75-ac41-cb7e1ec842a9"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.691561 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-kube-api-access-smpcp" (OuterVolumeSpecName: "kube-api-access-smpcp") pod "6acb6255-ec9f-4c75-ac41-cb7e1ec842a9" (UID: "6acb6255-ec9f-4c75-ac41-cb7e1ec842a9"). InnerVolumeSpecName "kube-api-access-smpcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.716457 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "6acb6255-ec9f-4c75-ac41-cb7e1ec842a9" (UID: "6acb6255-ec9f-4c75-ac41-cb7e1ec842a9"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.725293 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "6acb6255-ec9f-4c75-ac41-cb7e1ec842a9" (UID: "6acb6255-ec9f-4c75-ac41-cb7e1ec842a9"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.739080 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6acb6255-ec9f-4c75-ac41-cb7e1ec842a9" (UID: "6acb6255-ec9f-4c75-ac41-cb7e1ec842a9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.751213 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-inventory" (OuterVolumeSpecName: "inventory") pod "6acb6255-ec9f-4c75-ac41-cb7e1ec842a9" (UID: "6acb6255-ec9f-4c75-ac41-cb7e1ec842a9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.751938 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "6acb6255-ec9f-4c75-ac41-cb7e1ec842a9" (UID: "6acb6255-ec9f-4c75-ac41-cb7e1ec842a9"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.784702 4766 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.784746 4766 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.784761 4766 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.784776 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smpcp\" (UniqueName: \"kubernetes.io/projected/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-kube-api-access-smpcp\") on node \"crc\" DevicePath \"\"" Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.784788 4766 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.784802 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:09:29 crc kubenswrapper[4766]: I1126 01:09:29.784815 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.013898 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" event={"ID":"6acb6255-ec9f-4c75-ac41-cb7e1ec842a9","Type":"ContainerDied","Data":"59243fb0332cc19e487b24f13e5cb41be615c3c1f338ac837cc737a3f6735c6d"} Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.013937 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59243fb0332cc19e487b24f13e5cb41be615c3c1f338ac837cc737a3f6735c6d" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.014447 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.116966 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp"] Nov 26 01:09:30 crc kubenswrapper[4766]: E1126 01:09:30.118285 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6acb6255-ec9f-4c75-ac41-cb7e1ec842a9" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.118456 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6acb6255-ec9f-4c75-ac41-cb7e1ec842a9" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.119078 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="6acb6255-ec9f-4c75-ac41-cb7e1ec842a9" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.122850 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.125988 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-ipmi-config-data" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.126431 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.127266 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.127737 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.128160 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.133530 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp"] Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.294968 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx5qz\" (UniqueName: \"kubernetes.io/projected/afe63c23-a92e-4701-a422-d431ed80d42d-kube-api-access-qx5qz\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.296137 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.296455 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.296764 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.296993 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.297209 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.297437 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.399058 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.399152 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.399193 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.399229 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.399261 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.399298 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.399337 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx5qz\" (UniqueName: \"kubernetes.io/projected/afe63c23-a92e-4701-a422-d431ed80d42d-kube-api-access-qx5qz\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.405159 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.405471 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.406413 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.409403 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.413636 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.415122 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.427403 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx5qz\" (UniqueName: \"kubernetes.io/projected/afe63c23-a92e-4701-a422-d431ed80d42d-kube-api-access-qx5qz\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:30 crc kubenswrapper[4766]: I1126 01:09:30.467486 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:09:31 crc kubenswrapper[4766]: I1126 01:09:31.064377 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp"] Nov 26 01:09:31 crc kubenswrapper[4766]: W1126 01:09:31.065830 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafe63c23_a92e_4701_a422_d431ed80d42d.slice/crio-5397693f9a4f7d427ffb4d87f47831d4ff1d3a41fec6592c5eff0db7849116b6 WatchSource:0}: Error finding container 5397693f9a4f7d427ffb4d87f47831d4ff1d3a41fec6592c5eff0db7849116b6: Status 404 returned error can't find the container with id 5397693f9a4f7d427ffb4d87f47831d4ff1d3a41fec6592c5eff0db7849116b6 Nov 26 01:09:32 crc kubenswrapper[4766]: I1126 01:09:32.035363 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" event={"ID":"afe63c23-a92e-4701-a422-d431ed80d42d","Type":"ContainerStarted","Data":"62544c5a4746cf33b2e88c463d3bb5e09fe39de843456d753c9aa80b23f092bb"} Nov 26 01:09:32 crc kubenswrapper[4766]: I1126 01:09:32.035706 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" event={"ID":"afe63c23-a92e-4701-a422-d431ed80d42d","Type":"ContainerStarted","Data":"5397693f9a4f7d427ffb4d87f47831d4ff1d3a41fec6592c5eff0db7849116b6"} Nov 26 01:09:32 crc kubenswrapper[4766]: I1126 01:09:32.060078 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" podStartSLOduration=1.521718921 podStartE2EDuration="2.060056697s" podCreationTimestamp="2025-11-26 01:09:30 +0000 UTC" firstStartedPulling="2025-11-26 01:09:31.069031752 +0000 UTC m=+2751.917802182" lastFinishedPulling="2025-11-26 01:09:31.607369528 +0000 UTC m=+2752.456139958" observedRunningTime="2025-11-26 01:09:32.052769917 +0000 UTC m=+2752.901540347" watchObservedRunningTime="2025-11-26 01:09:32.060056697 +0000 UTC m=+2752.908827137" Nov 26 01:09:41 crc kubenswrapper[4766]: I1126 01:09:41.481993 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:09:41 crc kubenswrapper[4766]: I1126 01:09:41.482539 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:09:41 crc kubenswrapper[4766]: I1126 01:09:41.482589 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 01:09:41 crc kubenswrapper[4766]: I1126 01:09:41.483609 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3f97e605694d4755ce21053291eee5a113d9b02cc30e9ddf766f118e2989829b"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 01:09:41 crc kubenswrapper[4766]: I1126 01:09:41.483688 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://3f97e605694d4755ce21053291eee5a113d9b02cc30e9ddf766f118e2989829b" gracePeriod=600 Nov 26 01:09:42 crc kubenswrapper[4766]: I1126 01:09:42.162380 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="3f97e605694d4755ce21053291eee5a113d9b02cc30e9ddf766f118e2989829b" exitCode=0 Nov 26 01:09:42 crc kubenswrapper[4766]: I1126 01:09:42.162452 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"3f97e605694d4755ce21053291eee5a113d9b02cc30e9ddf766f118e2989829b"} Nov 26 01:09:42 crc kubenswrapper[4766]: I1126 01:09:42.163078 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50"} Nov 26 01:09:42 crc kubenswrapper[4766]: I1126 01:09:42.163106 4766 scope.go:117] "RemoveContainer" containerID="857db0389942935161b49d4ed667d794ea1b95d13d5e905b2112a2b19062ad0c" Nov 26 01:09:56 crc kubenswrapper[4766]: I1126 01:09:56.970336 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6zwj9"] Nov 26 01:09:56 crc kubenswrapper[4766]: I1126 01:09:56.976401 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6zwj9" Nov 26 01:09:57 crc kubenswrapper[4766]: I1126 01:09:57.035110 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6zwj9"] Nov 26 01:09:57 crc kubenswrapper[4766]: I1126 01:09:57.072825 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfls7\" (UniqueName: \"kubernetes.io/projected/023496b5-7ea6-4bcf-b43d-150910eb3fc9-kube-api-access-bfls7\") pod \"certified-operators-6zwj9\" (UID: \"023496b5-7ea6-4bcf-b43d-150910eb3fc9\") " pod="openshift-marketplace/certified-operators-6zwj9" Nov 26 01:09:57 crc kubenswrapper[4766]: I1126 01:09:57.073372 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/023496b5-7ea6-4bcf-b43d-150910eb3fc9-utilities\") pod \"certified-operators-6zwj9\" (UID: \"023496b5-7ea6-4bcf-b43d-150910eb3fc9\") " pod="openshift-marketplace/certified-operators-6zwj9" Nov 26 01:09:57 crc kubenswrapper[4766]: I1126 01:09:57.073418 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/023496b5-7ea6-4bcf-b43d-150910eb3fc9-catalog-content\") pod \"certified-operators-6zwj9\" (UID: \"023496b5-7ea6-4bcf-b43d-150910eb3fc9\") " pod="openshift-marketplace/certified-operators-6zwj9" Nov 26 01:09:57 crc kubenswrapper[4766]: I1126 01:09:57.175360 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfls7\" (UniqueName: \"kubernetes.io/projected/023496b5-7ea6-4bcf-b43d-150910eb3fc9-kube-api-access-bfls7\") pod \"certified-operators-6zwj9\" (UID: \"023496b5-7ea6-4bcf-b43d-150910eb3fc9\") " pod="openshift-marketplace/certified-operators-6zwj9" Nov 26 01:09:57 crc kubenswrapper[4766]: I1126 01:09:57.175801 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/023496b5-7ea6-4bcf-b43d-150910eb3fc9-utilities\") pod \"certified-operators-6zwj9\" (UID: \"023496b5-7ea6-4bcf-b43d-150910eb3fc9\") " pod="openshift-marketplace/certified-operators-6zwj9" Nov 26 01:09:57 crc kubenswrapper[4766]: I1126 01:09:57.175857 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/023496b5-7ea6-4bcf-b43d-150910eb3fc9-catalog-content\") pod \"certified-operators-6zwj9\" (UID: \"023496b5-7ea6-4bcf-b43d-150910eb3fc9\") " pod="openshift-marketplace/certified-operators-6zwj9" Nov 26 01:09:57 crc kubenswrapper[4766]: I1126 01:09:57.178521 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/023496b5-7ea6-4bcf-b43d-150910eb3fc9-catalog-content\") pod \"certified-operators-6zwj9\" (UID: \"023496b5-7ea6-4bcf-b43d-150910eb3fc9\") " pod="openshift-marketplace/certified-operators-6zwj9" Nov 26 01:09:57 crc kubenswrapper[4766]: I1126 01:09:57.178576 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/023496b5-7ea6-4bcf-b43d-150910eb3fc9-utilities\") pod \"certified-operators-6zwj9\" (UID: \"023496b5-7ea6-4bcf-b43d-150910eb3fc9\") " pod="openshift-marketplace/certified-operators-6zwj9" Nov 26 01:09:57 crc kubenswrapper[4766]: I1126 01:09:57.206272 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfls7\" (UniqueName: \"kubernetes.io/projected/023496b5-7ea6-4bcf-b43d-150910eb3fc9-kube-api-access-bfls7\") pod \"certified-operators-6zwj9\" (UID: \"023496b5-7ea6-4bcf-b43d-150910eb3fc9\") " pod="openshift-marketplace/certified-operators-6zwj9" Nov 26 01:09:57 crc kubenswrapper[4766]: I1126 01:09:57.355398 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6zwj9" Nov 26 01:09:57 crc kubenswrapper[4766]: I1126 01:09:57.857102 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6zwj9"] Nov 26 01:09:58 crc kubenswrapper[4766]: I1126 01:09:58.365499 4766 generic.go:334] "Generic (PLEG): container finished" podID="023496b5-7ea6-4bcf-b43d-150910eb3fc9" containerID="0bf7a1b94e1a20a15d5402a9cca80662e3a96dd754d482e3d36191dc562cdf44" exitCode=0 Nov 26 01:09:58 crc kubenswrapper[4766]: I1126 01:09:58.365899 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6zwj9" event={"ID":"023496b5-7ea6-4bcf-b43d-150910eb3fc9","Type":"ContainerDied","Data":"0bf7a1b94e1a20a15d5402a9cca80662e3a96dd754d482e3d36191dc562cdf44"} Nov 26 01:09:58 crc kubenswrapper[4766]: I1126 01:09:58.367177 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6zwj9" event={"ID":"023496b5-7ea6-4bcf-b43d-150910eb3fc9","Type":"ContainerStarted","Data":"71bc306beadbb228cfa3dd84200caaf0145b0d8e993843be98e79c25483332c0"} Nov 26 01:09:59 crc kubenswrapper[4766]: I1126 01:09:59.376801 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6zwj9" event={"ID":"023496b5-7ea6-4bcf-b43d-150910eb3fc9","Type":"ContainerStarted","Data":"b7061ba77b30776eb86c7b0cacffcbaed2b08728ebe8373861b5bc6a1f5f6cf6"} Nov 26 01:10:01 crc kubenswrapper[4766]: I1126 01:10:01.397465 4766 generic.go:334] "Generic (PLEG): container finished" podID="023496b5-7ea6-4bcf-b43d-150910eb3fc9" containerID="b7061ba77b30776eb86c7b0cacffcbaed2b08728ebe8373861b5bc6a1f5f6cf6" exitCode=0 Nov 26 01:10:01 crc kubenswrapper[4766]: I1126 01:10:01.397543 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6zwj9" event={"ID":"023496b5-7ea6-4bcf-b43d-150910eb3fc9","Type":"ContainerDied","Data":"b7061ba77b30776eb86c7b0cacffcbaed2b08728ebe8373861b5bc6a1f5f6cf6"} Nov 26 01:10:02 crc kubenswrapper[4766]: I1126 01:10:02.412783 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6zwj9" event={"ID":"023496b5-7ea6-4bcf-b43d-150910eb3fc9","Type":"ContainerStarted","Data":"97cb88a99b1ef8977332f2a8e7765fbb5a5e27b4aa235cf912c887f30fe73c36"} Nov 26 01:10:02 crc kubenswrapper[4766]: I1126 01:10:02.445108 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6zwj9" podStartSLOduration=3.003241553 podStartE2EDuration="6.445084224s" podCreationTimestamp="2025-11-26 01:09:56 +0000 UTC" firstStartedPulling="2025-11-26 01:09:58.373336204 +0000 UTC m=+2779.222106644" lastFinishedPulling="2025-11-26 01:10:01.815178875 +0000 UTC m=+2782.663949315" observedRunningTime="2025-11-26 01:10:02.430477283 +0000 UTC m=+2783.279247743" watchObservedRunningTime="2025-11-26 01:10:02.445084224 +0000 UTC m=+2783.293854664" Nov 26 01:10:07 crc kubenswrapper[4766]: I1126 01:10:07.356594 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6zwj9" Nov 26 01:10:07 crc kubenswrapper[4766]: I1126 01:10:07.357240 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6zwj9" Nov 26 01:10:07 crc kubenswrapper[4766]: I1126 01:10:07.462507 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6zwj9" Nov 26 01:10:07 crc kubenswrapper[4766]: I1126 01:10:07.541346 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6zwj9" Nov 26 01:10:07 crc kubenswrapper[4766]: I1126 01:10:07.732698 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6zwj9"] Nov 26 01:10:09 crc kubenswrapper[4766]: I1126 01:10:09.492378 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6zwj9" podUID="023496b5-7ea6-4bcf-b43d-150910eb3fc9" containerName="registry-server" containerID="cri-o://97cb88a99b1ef8977332f2a8e7765fbb5a5e27b4aa235cf912c887f30fe73c36" gracePeriod=2 Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.043592 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6zwj9" Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.062183 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfls7\" (UniqueName: \"kubernetes.io/projected/023496b5-7ea6-4bcf-b43d-150910eb3fc9-kube-api-access-bfls7\") pod \"023496b5-7ea6-4bcf-b43d-150910eb3fc9\" (UID: \"023496b5-7ea6-4bcf-b43d-150910eb3fc9\") " Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.062241 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/023496b5-7ea6-4bcf-b43d-150910eb3fc9-utilities\") pod \"023496b5-7ea6-4bcf-b43d-150910eb3fc9\" (UID: \"023496b5-7ea6-4bcf-b43d-150910eb3fc9\") " Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.062397 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/023496b5-7ea6-4bcf-b43d-150910eb3fc9-catalog-content\") pod \"023496b5-7ea6-4bcf-b43d-150910eb3fc9\" (UID: \"023496b5-7ea6-4bcf-b43d-150910eb3fc9\") " Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.062965 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/023496b5-7ea6-4bcf-b43d-150910eb3fc9-utilities" (OuterVolumeSpecName: "utilities") pod "023496b5-7ea6-4bcf-b43d-150910eb3fc9" (UID: "023496b5-7ea6-4bcf-b43d-150910eb3fc9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.067764 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/023496b5-7ea6-4bcf-b43d-150910eb3fc9-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.072349 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/023496b5-7ea6-4bcf-b43d-150910eb3fc9-kube-api-access-bfls7" (OuterVolumeSpecName: "kube-api-access-bfls7") pod "023496b5-7ea6-4bcf-b43d-150910eb3fc9" (UID: "023496b5-7ea6-4bcf-b43d-150910eb3fc9"). InnerVolumeSpecName "kube-api-access-bfls7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.137152 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/023496b5-7ea6-4bcf-b43d-150910eb3fc9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "023496b5-7ea6-4bcf-b43d-150910eb3fc9" (UID: "023496b5-7ea6-4bcf-b43d-150910eb3fc9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.169618 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/023496b5-7ea6-4bcf-b43d-150910eb3fc9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.169664 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfls7\" (UniqueName: \"kubernetes.io/projected/023496b5-7ea6-4bcf-b43d-150910eb3fc9-kube-api-access-bfls7\") on node \"crc\" DevicePath \"\"" Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.511365 4766 generic.go:334] "Generic (PLEG): container finished" podID="023496b5-7ea6-4bcf-b43d-150910eb3fc9" containerID="97cb88a99b1ef8977332f2a8e7765fbb5a5e27b4aa235cf912c887f30fe73c36" exitCode=0 Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.511447 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6zwj9" Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.511467 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6zwj9" event={"ID":"023496b5-7ea6-4bcf-b43d-150910eb3fc9","Type":"ContainerDied","Data":"97cb88a99b1ef8977332f2a8e7765fbb5a5e27b4aa235cf912c887f30fe73c36"} Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.511970 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6zwj9" event={"ID":"023496b5-7ea6-4bcf-b43d-150910eb3fc9","Type":"ContainerDied","Data":"71bc306beadbb228cfa3dd84200caaf0145b0d8e993843be98e79c25483332c0"} Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.512005 4766 scope.go:117] "RemoveContainer" containerID="97cb88a99b1ef8977332f2a8e7765fbb5a5e27b4aa235cf912c887f30fe73c36" Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.533435 4766 scope.go:117] "RemoveContainer" containerID="b7061ba77b30776eb86c7b0cacffcbaed2b08728ebe8373861b5bc6a1f5f6cf6" Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.556103 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6zwj9"] Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.567167 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6zwj9"] Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.573119 4766 scope.go:117] "RemoveContainer" containerID="0bf7a1b94e1a20a15d5402a9cca80662e3a96dd754d482e3d36191dc562cdf44" Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.610445 4766 scope.go:117] "RemoveContainer" containerID="97cb88a99b1ef8977332f2a8e7765fbb5a5e27b4aa235cf912c887f30fe73c36" Nov 26 01:10:10 crc kubenswrapper[4766]: E1126 01:10:10.611338 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97cb88a99b1ef8977332f2a8e7765fbb5a5e27b4aa235cf912c887f30fe73c36\": container with ID starting with 97cb88a99b1ef8977332f2a8e7765fbb5a5e27b4aa235cf912c887f30fe73c36 not found: ID does not exist" containerID="97cb88a99b1ef8977332f2a8e7765fbb5a5e27b4aa235cf912c887f30fe73c36" Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.611397 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97cb88a99b1ef8977332f2a8e7765fbb5a5e27b4aa235cf912c887f30fe73c36"} err="failed to get container status \"97cb88a99b1ef8977332f2a8e7765fbb5a5e27b4aa235cf912c887f30fe73c36\": rpc error: code = NotFound desc = could not find container \"97cb88a99b1ef8977332f2a8e7765fbb5a5e27b4aa235cf912c887f30fe73c36\": container with ID starting with 97cb88a99b1ef8977332f2a8e7765fbb5a5e27b4aa235cf912c887f30fe73c36 not found: ID does not exist" Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.611431 4766 scope.go:117] "RemoveContainer" containerID="b7061ba77b30776eb86c7b0cacffcbaed2b08728ebe8373861b5bc6a1f5f6cf6" Nov 26 01:10:10 crc kubenswrapper[4766]: E1126 01:10:10.611866 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7061ba77b30776eb86c7b0cacffcbaed2b08728ebe8373861b5bc6a1f5f6cf6\": container with ID starting with b7061ba77b30776eb86c7b0cacffcbaed2b08728ebe8373861b5bc6a1f5f6cf6 not found: ID does not exist" containerID="b7061ba77b30776eb86c7b0cacffcbaed2b08728ebe8373861b5bc6a1f5f6cf6" Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.611904 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7061ba77b30776eb86c7b0cacffcbaed2b08728ebe8373861b5bc6a1f5f6cf6"} err="failed to get container status \"b7061ba77b30776eb86c7b0cacffcbaed2b08728ebe8373861b5bc6a1f5f6cf6\": rpc error: code = NotFound desc = could not find container \"b7061ba77b30776eb86c7b0cacffcbaed2b08728ebe8373861b5bc6a1f5f6cf6\": container with ID starting with b7061ba77b30776eb86c7b0cacffcbaed2b08728ebe8373861b5bc6a1f5f6cf6 not found: ID does not exist" Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.611929 4766 scope.go:117] "RemoveContainer" containerID="0bf7a1b94e1a20a15d5402a9cca80662e3a96dd754d482e3d36191dc562cdf44" Nov 26 01:10:10 crc kubenswrapper[4766]: E1126 01:10:10.612248 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bf7a1b94e1a20a15d5402a9cca80662e3a96dd754d482e3d36191dc562cdf44\": container with ID starting with 0bf7a1b94e1a20a15d5402a9cca80662e3a96dd754d482e3d36191dc562cdf44 not found: ID does not exist" containerID="0bf7a1b94e1a20a15d5402a9cca80662e3a96dd754d482e3d36191dc562cdf44" Nov 26 01:10:10 crc kubenswrapper[4766]: I1126 01:10:10.612292 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bf7a1b94e1a20a15d5402a9cca80662e3a96dd754d482e3d36191dc562cdf44"} err="failed to get container status \"0bf7a1b94e1a20a15d5402a9cca80662e3a96dd754d482e3d36191dc562cdf44\": rpc error: code = NotFound desc = could not find container \"0bf7a1b94e1a20a15d5402a9cca80662e3a96dd754d482e3d36191dc562cdf44\": container with ID starting with 0bf7a1b94e1a20a15d5402a9cca80662e3a96dd754d482e3d36191dc562cdf44 not found: ID does not exist" Nov 26 01:10:11 crc kubenswrapper[4766]: I1126 01:10:11.940202 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="023496b5-7ea6-4bcf-b43d-150910eb3fc9" path="/var/lib/kubelet/pods/023496b5-7ea6-4bcf-b43d-150910eb3fc9/volumes" Nov 26 01:10:22 crc kubenswrapper[4766]: I1126 01:10:22.187474 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hz9h4"] Nov 26 01:10:22 crc kubenswrapper[4766]: E1126 01:10:22.188794 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="023496b5-7ea6-4bcf-b43d-150910eb3fc9" containerName="extract-utilities" Nov 26 01:10:22 crc kubenswrapper[4766]: I1126 01:10:22.188819 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="023496b5-7ea6-4bcf-b43d-150910eb3fc9" containerName="extract-utilities" Nov 26 01:10:22 crc kubenswrapper[4766]: E1126 01:10:22.188845 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="023496b5-7ea6-4bcf-b43d-150910eb3fc9" containerName="extract-content" Nov 26 01:10:22 crc kubenswrapper[4766]: I1126 01:10:22.188858 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="023496b5-7ea6-4bcf-b43d-150910eb3fc9" containerName="extract-content" Nov 26 01:10:22 crc kubenswrapper[4766]: E1126 01:10:22.188886 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="023496b5-7ea6-4bcf-b43d-150910eb3fc9" containerName="registry-server" Nov 26 01:10:22 crc kubenswrapper[4766]: I1126 01:10:22.188899 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="023496b5-7ea6-4bcf-b43d-150910eb3fc9" containerName="registry-server" Nov 26 01:10:22 crc kubenswrapper[4766]: I1126 01:10:22.189290 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="023496b5-7ea6-4bcf-b43d-150910eb3fc9" containerName="registry-server" Nov 26 01:10:22 crc kubenswrapper[4766]: I1126 01:10:22.191254 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hz9h4" Nov 26 01:10:22 crc kubenswrapper[4766]: I1126 01:10:22.264767 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hz9h4"] Nov 26 01:10:22 crc kubenswrapper[4766]: I1126 01:10:22.346994 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a275746-1b85-48ac-b2fe-73245940c1f2-utilities\") pod \"community-operators-hz9h4\" (UID: \"6a275746-1b85-48ac-b2fe-73245940c1f2\") " pod="openshift-marketplace/community-operators-hz9h4" Nov 26 01:10:22 crc kubenswrapper[4766]: I1126 01:10:22.347067 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swx5s\" (UniqueName: \"kubernetes.io/projected/6a275746-1b85-48ac-b2fe-73245940c1f2-kube-api-access-swx5s\") pod \"community-operators-hz9h4\" (UID: \"6a275746-1b85-48ac-b2fe-73245940c1f2\") " pod="openshift-marketplace/community-operators-hz9h4" Nov 26 01:10:22 crc kubenswrapper[4766]: I1126 01:10:22.347212 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a275746-1b85-48ac-b2fe-73245940c1f2-catalog-content\") pod \"community-operators-hz9h4\" (UID: \"6a275746-1b85-48ac-b2fe-73245940c1f2\") " pod="openshift-marketplace/community-operators-hz9h4" Nov 26 01:10:22 crc kubenswrapper[4766]: I1126 01:10:22.449862 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a275746-1b85-48ac-b2fe-73245940c1f2-catalog-content\") pod \"community-operators-hz9h4\" (UID: \"6a275746-1b85-48ac-b2fe-73245940c1f2\") " pod="openshift-marketplace/community-operators-hz9h4" Nov 26 01:10:22 crc kubenswrapper[4766]: I1126 01:10:22.450892 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a275746-1b85-48ac-b2fe-73245940c1f2-utilities\") pod \"community-operators-hz9h4\" (UID: \"6a275746-1b85-48ac-b2fe-73245940c1f2\") " pod="openshift-marketplace/community-operators-hz9h4" Nov 26 01:10:22 crc kubenswrapper[4766]: I1126 01:10:22.450697 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a275746-1b85-48ac-b2fe-73245940c1f2-catalog-content\") pod \"community-operators-hz9h4\" (UID: \"6a275746-1b85-48ac-b2fe-73245940c1f2\") " pod="openshift-marketplace/community-operators-hz9h4" Nov 26 01:10:22 crc kubenswrapper[4766]: I1126 01:10:22.451010 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swx5s\" (UniqueName: \"kubernetes.io/projected/6a275746-1b85-48ac-b2fe-73245940c1f2-kube-api-access-swx5s\") pod \"community-operators-hz9h4\" (UID: \"6a275746-1b85-48ac-b2fe-73245940c1f2\") " pod="openshift-marketplace/community-operators-hz9h4" Nov 26 01:10:22 crc kubenswrapper[4766]: I1126 01:10:22.451504 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a275746-1b85-48ac-b2fe-73245940c1f2-utilities\") pod \"community-operators-hz9h4\" (UID: \"6a275746-1b85-48ac-b2fe-73245940c1f2\") " pod="openshift-marketplace/community-operators-hz9h4" Nov 26 01:10:22 crc kubenswrapper[4766]: I1126 01:10:22.472482 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swx5s\" (UniqueName: \"kubernetes.io/projected/6a275746-1b85-48ac-b2fe-73245940c1f2-kube-api-access-swx5s\") pod \"community-operators-hz9h4\" (UID: \"6a275746-1b85-48ac-b2fe-73245940c1f2\") " pod="openshift-marketplace/community-operators-hz9h4" Nov 26 01:10:22 crc kubenswrapper[4766]: I1126 01:10:22.559272 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hz9h4" Nov 26 01:10:23 crc kubenswrapper[4766]: I1126 01:10:23.111171 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hz9h4"] Nov 26 01:10:23 crc kubenswrapper[4766]: I1126 01:10:23.701538 4766 generic.go:334] "Generic (PLEG): container finished" podID="6a275746-1b85-48ac-b2fe-73245940c1f2" containerID="2ae8c9c1a306371d61c288e91c8b8bc26cb00b36172baa9aae547417c4864a16" exitCode=0 Nov 26 01:10:23 crc kubenswrapper[4766]: I1126 01:10:23.701629 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hz9h4" event={"ID":"6a275746-1b85-48ac-b2fe-73245940c1f2","Type":"ContainerDied","Data":"2ae8c9c1a306371d61c288e91c8b8bc26cb00b36172baa9aae547417c4864a16"} Nov 26 01:10:23 crc kubenswrapper[4766]: I1126 01:10:23.701739 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hz9h4" event={"ID":"6a275746-1b85-48ac-b2fe-73245940c1f2","Type":"ContainerStarted","Data":"3e86252aee445ae7d6b7895505fc9dfc458fd00bc706d6512a084e2d818f9c52"} Nov 26 01:10:24 crc kubenswrapper[4766]: I1126 01:10:24.716460 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hz9h4" event={"ID":"6a275746-1b85-48ac-b2fe-73245940c1f2","Type":"ContainerStarted","Data":"f70ef3570bb88eb8b1a1967a7338ba152bd90d21a70942a8232b0a1b2745a7a4"} Nov 26 01:10:26 crc kubenswrapper[4766]: I1126 01:10:26.745213 4766 generic.go:334] "Generic (PLEG): container finished" podID="6a275746-1b85-48ac-b2fe-73245940c1f2" containerID="f70ef3570bb88eb8b1a1967a7338ba152bd90d21a70942a8232b0a1b2745a7a4" exitCode=0 Nov 26 01:10:26 crc kubenswrapper[4766]: I1126 01:10:26.745300 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hz9h4" event={"ID":"6a275746-1b85-48ac-b2fe-73245940c1f2","Type":"ContainerDied","Data":"f70ef3570bb88eb8b1a1967a7338ba152bd90d21a70942a8232b0a1b2745a7a4"} Nov 26 01:10:27 crc kubenswrapper[4766]: I1126 01:10:27.768949 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hz9h4" event={"ID":"6a275746-1b85-48ac-b2fe-73245940c1f2","Type":"ContainerStarted","Data":"afeac045e16a0ae7860a537597879154032b7fd0d630688e10a86bc543be0a21"} Nov 26 01:10:27 crc kubenswrapper[4766]: I1126 01:10:27.813279 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hz9h4" podStartSLOduration=2.382548136 podStartE2EDuration="5.813256861s" podCreationTimestamp="2025-11-26 01:10:22 +0000 UTC" firstStartedPulling="2025-11-26 01:10:23.70585781 +0000 UTC m=+2804.554628250" lastFinishedPulling="2025-11-26 01:10:27.136566515 +0000 UTC m=+2807.985336975" observedRunningTime="2025-11-26 01:10:27.805268193 +0000 UTC m=+2808.654038693" watchObservedRunningTime="2025-11-26 01:10:27.813256861 +0000 UTC m=+2808.662027301" Nov 26 01:10:32 crc kubenswrapper[4766]: I1126 01:10:32.560422 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hz9h4" Nov 26 01:10:32 crc kubenswrapper[4766]: I1126 01:10:32.560876 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hz9h4" Nov 26 01:10:32 crc kubenswrapper[4766]: I1126 01:10:32.649175 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hz9h4" Nov 26 01:10:32 crc kubenswrapper[4766]: I1126 01:10:32.927606 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hz9h4" Nov 26 01:10:33 crc kubenswrapper[4766]: I1126 01:10:33.006875 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hz9h4"] Nov 26 01:10:34 crc kubenswrapper[4766]: I1126 01:10:34.867032 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hz9h4" podUID="6a275746-1b85-48ac-b2fe-73245940c1f2" containerName="registry-server" containerID="cri-o://afeac045e16a0ae7860a537597879154032b7fd0d630688e10a86bc543be0a21" gracePeriod=2 Nov 26 01:10:35 crc kubenswrapper[4766]: I1126 01:10:35.416373 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hz9h4" Nov 26 01:10:35 crc kubenswrapper[4766]: I1126 01:10:35.597830 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a275746-1b85-48ac-b2fe-73245940c1f2-catalog-content\") pod \"6a275746-1b85-48ac-b2fe-73245940c1f2\" (UID: \"6a275746-1b85-48ac-b2fe-73245940c1f2\") " Nov 26 01:10:35 crc kubenswrapper[4766]: I1126 01:10:35.597994 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a275746-1b85-48ac-b2fe-73245940c1f2-utilities\") pod \"6a275746-1b85-48ac-b2fe-73245940c1f2\" (UID: \"6a275746-1b85-48ac-b2fe-73245940c1f2\") " Nov 26 01:10:35 crc kubenswrapper[4766]: I1126 01:10:35.598106 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swx5s\" (UniqueName: \"kubernetes.io/projected/6a275746-1b85-48ac-b2fe-73245940c1f2-kube-api-access-swx5s\") pod \"6a275746-1b85-48ac-b2fe-73245940c1f2\" (UID: \"6a275746-1b85-48ac-b2fe-73245940c1f2\") " Nov 26 01:10:35 crc kubenswrapper[4766]: I1126 01:10:35.599472 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a275746-1b85-48ac-b2fe-73245940c1f2-utilities" (OuterVolumeSpecName: "utilities") pod "6a275746-1b85-48ac-b2fe-73245940c1f2" (UID: "6a275746-1b85-48ac-b2fe-73245940c1f2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:10:35 crc kubenswrapper[4766]: I1126 01:10:35.611832 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a275746-1b85-48ac-b2fe-73245940c1f2-kube-api-access-swx5s" (OuterVolumeSpecName: "kube-api-access-swx5s") pod "6a275746-1b85-48ac-b2fe-73245940c1f2" (UID: "6a275746-1b85-48ac-b2fe-73245940c1f2"). InnerVolumeSpecName "kube-api-access-swx5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:10:35 crc kubenswrapper[4766]: I1126 01:10:35.678024 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a275746-1b85-48ac-b2fe-73245940c1f2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a275746-1b85-48ac-b2fe-73245940c1f2" (UID: "6a275746-1b85-48ac-b2fe-73245940c1f2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:10:35 crc kubenswrapper[4766]: I1126 01:10:35.702494 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a275746-1b85-48ac-b2fe-73245940c1f2-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 01:10:35 crc kubenswrapper[4766]: I1126 01:10:35.702564 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swx5s\" (UniqueName: \"kubernetes.io/projected/6a275746-1b85-48ac-b2fe-73245940c1f2-kube-api-access-swx5s\") on node \"crc\" DevicePath \"\"" Nov 26 01:10:35 crc kubenswrapper[4766]: I1126 01:10:35.702588 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a275746-1b85-48ac-b2fe-73245940c1f2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 01:10:35 crc kubenswrapper[4766]: I1126 01:10:35.881458 4766 generic.go:334] "Generic (PLEG): container finished" podID="6a275746-1b85-48ac-b2fe-73245940c1f2" containerID="afeac045e16a0ae7860a537597879154032b7fd0d630688e10a86bc543be0a21" exitCode=0 Nov 26 01:10:35 crc kubenswrapper[4766]: I1126 01:10:35.881548 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hz9h4" event={"ID":"6a275746-1b85-48ac-b2fe-73245940c1f2","Type":"ContainerDied","Data":"afeac045e16a0ae7860a537597879154032b7fd0d630688e10a86bc543be0a21"} Nov 26 01:10:35 crc kubenswrapper[4766]: I1126 01:10:35.881598 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hz9h4" event={"ID":"6a275746-1b85-48ac-b2fe-73245940c1f2","Type":"ContainerDied","Data":"3e86252aee445ae7d6b7895505fc9dfc458fd00bc706d6512a084e2d818f9c52"} Nov 26 01:10:35 crc kubenswrapper[4766]: I1126 01:10:35.881610 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hz9h4" Nov 26 01:10:35 crc kubenswrapper[4766]: I1126 01:10:35.881631 4766 scope.go:117] "RemoveContainer" containerID="afeac045e16a0ae7860a537597879154032b7fd0d630688e10a86bc543be0a21" Nov 26 01:10:35 crc kubenswrapper[4766]: I1126 01:10:35.919227 4766 scope.go:117] "RemoveContainer" containerID="f70ef3570bb88eb8b1a1967a7338ba152bd90d21a70942a8232b0a1b2745a7a4" Nov 26 01:10:35 crc kubenswrapper[4766]: I1126 01:10:35.926893 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hz9h4"] Nov 26 01:10:35 crc kubenswrapper[4766]: I1126 01:10:35.949040 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hz9h4"] Nov 26 01:10:35 crc kubenswrapper[4766]: I1126 01:10:35.964164 4766 scope.go:117] "RemoveContainer" containerID="2ae8c9c1a306371d61c288e91c8b8bc26cb00b36172baa9aae547417c4864a16" Nov 26 01:10:36 crc kubenswrapper[4766]: I1126 01:10:36.013450 4766 scope.go:117] "RemoveContainer" containerID="afeac045e16a0ae7860a537597879154032b7fd0d630688e10a86bc543be0a21" Nov 26 01:10:36 crc kubenswrapper[4766]: E1126 01:10:36.014974 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afeac045e16a0ae7860a537597879154032b7fd0d630688e10a86bc543be0a21\": container with ID starting with afeac045e16a0ae7860a537597879154032b7fd0d630688e10a86bc543be0a21 not found: ID does not exist" containerID="afeac045e16a0ae7860a537597879154032b7fd0d630688e10a86bc543be0a21" Nov 26 01:10:36 crc kubenswrapper[4766]: I1126 01:10:36.015021 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afeac045e16a0ae7860a537597879154032b7fd0d630688e10a86bc543be0a21"} err="failed to get container status \"afeac045e16a0ae7860a537597879154032b7fd0d630688e10a86bc543be0a21\": rpc error: code = NotFound desc = could not find container \"afeac045e16a0ae7860a537597879154032b7fd0d630688e10a86bc543be0a21\": container with ID starting with afeac045e16a0ae7860a537597879154032b7fd0d630688e10a86bc543be0a21 not found: ID does not exist" Nov 26 01:10:36 crc kubenswrapper[4766]: I1126 01:10:36.015046 4766 scope.go:117] "RemoveContainer" containerID="f70ef3570bb88eb8b1a1967a7338ba152bd90d21a70942a8232b0a1b2745a7a4" Nov 26 01:10:36 crc kubenswrapper[4766]: E1126 01:10:36.015612 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f70ef3570bb88eb8b1a1967a7338ba152bd90d21a70942a8232b0a1b2745a7a4\": container with ID starting with f70ef3570bb88eb8b1a1967a7338ba152bd90d21a70942a8232b0a1b2745a7a4 not found: ID does not exist" containerID="f70ef3570bb88eb8b1a1967a7338ba152bd90d21a70942a8232b0a1b2745a7a4" Nov 26 01:10:36 crc kubenswrapper[4766]: I1126 01:10:36.015683 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f70ef3570bb88eb8b1a1967a7338ba152bd90d21a70942a8232b0a1b2745a7a4"} err="failed to get container status \"f70ef3570bb88eb8b1a1967a7338ba152bd90d21a70942a8232b0a1b2745a7a4\": rpc error: code = NotFound desc = could not find container \"f70ef3570bb88eb8b1a1967a7338ba152bd90d21a70942a8232b0a1b2745a7a4\": container with ID starting with f70ef3570bb88eb8b1a1967a7338ba152bd90d21a70942a8232b0a1b2745a7a4 not found: ID does not exist" Nov 26 01:10:36 crc kubenswrapper[4766]: I1126 01:10:36.015714 4766 scope.go:117] "RemoveContainer" containerID="2ae8c9c1a306371d61c288e91c8b8bc26cb00b36172baa9aae547417c4864a16" Nov 26 01:10:36 crc kubenswrapper[4766]: E1126 01:10:36.016103 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ae8c9c1a306371d61c288e91c8b8bc26cb00b36172baa9aae547417c4864a16\": container with ID starting with 2ae8c9c1a306371d61c288e91c8b8bc26cb00b36172baa9aae547417c4864a16 not found: ID does not exist" containerID="2ae8c9c1a306371d61c288e91c8b8bc26cb00b36172baa9aae547417c4864a16" Nov 26 01:10:36 crc kubenswrapper[4766]: I1126 01:10:36.016163 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ae8c9c1a306371d61c288e91c8b8bc26cb00b36172baa9aae547417c4864a16"} err="failed to get container status \"2ae8c9c1a306371d61c288e91c8b8bc26cb00b36172baa9aae547417c4864a16\": rpc error: code = NotFound desc = could not find container \"2ae8c9c1a306371d61c288e91c8b8bc26cb00b36172baa9aae547417c4864a16\": container with ID starting with 2ae8c9c1a306371d61c288e91c8b8bc26cb00b36172baa9aae547417c4864a16 not found: ID does not exist" Nov 26 01:10:37 crc kubenswrapper[4766]: I1126 01:10:37.849312 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a275746-1b85-48ac-b2fe-73245940c1f2" path="/var/lib/kubelet/pods/6a275746-1b85-48ac-b2fe-73245940c1f2/volumes" Nov 26 01:11:20 crc kubenswrapper[4766]: I1126 01:11:20.207923 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8hhfj"] Nov 26 01:11:20 crc kubenswrapper[4766]: E1126 01:11:20.208865 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a275746-1b85-48ac-b2fe-73245940c1f2" containerName="extract-content" Nov 26 01:11:20 crc kubenswrapper[4766]: I1126 01:11:20.208880 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a275746-1b85-48ac-b2fe-73245940c1f2" containerName="extract-content" Nov 26 01:11:20 crc kubenswrapper[4766]: E1126 01:11:20.208902 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a275746-1b85-48ac-b2fe-73245940c1f2" containerName="registry-server" Nov 26 01:11:20 crc kubenswrapper[4766]: I1126 01:11:20.208908 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a275746-1b85-48ac-b2fe-73245940c1f2" containerName="registry-server" Nov 26 01:11:20 crc kubenswrapper[4766]: E1126 01:11:20.208934 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a275746-1b85-48ac-b2fe-73245940c1f2" containerName="extract-utilities" Nov 26 01:11:20 crc kubenswrapper[4766]: I1126 01:11:20.208941 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a275746-1b85-48ac-b2fe-73245940c1f2" containerName="extract-utilities" Nov 26 01:11:20 crc kubenswrapper[4766]: I1126 01:11:20.209161 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a275746-1b85-48ac-b2fe-73245940c1f2" containerName="registry-server" Nov 26 01:11:20 crc kubenswrapper[4766]: I1126 01:11:20.210636 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8hhfj" Nov 26 01:11:20 crc kubenswrapper[4766]: I1126 01:11:20.233868 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8hhfj"] Nov 26 01:11:20 crc kubenswrapper[4766]: I1126 01:11:20.238000 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc16bbba-993f-4376-98ee-8e7770ace54e-utilities\") pod \"redhat-marketplace-8hhfj\" (UID: \"fc16bbba-993f-4376-98ee-8e7770ace54e\") " pod="openshift-marketplace/redhat-marketplace-8hhfj" Nov 26 01:11:20 crc kubenswrapper[4766]: I1126 01:11:20.238046 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qc7n\" (UniqueName: \"kubernetes.io/projected/fc16bbba-993f-4376-98ee-8e7770ace54e-kube-api-access-9qc7n\") pod \"redhat-marketplace-8hhfj\" (UID: \"fc16bbba-993f-4376-98ee-8e7770ace54e\") " pod="openshift-marketplace/redhat-marketplace-8hhfj" Nov 26 01:11:20 crc kubenswrapper[4766]: I1126 01:11:20.238103 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc16bbba-993f-4376-98ee-8e7770ace54e-catalog-content\") pod \"redhat-marketplace-8hhfj\" (UID: \"fc16bbba-993f-4376-98ee-8e7770ace54e\") " pod="openshift-marketplace/redhat-marketplace-8hhfj" Nov 26 01:11:20 crc kubenswrapper[4766]: I1126 01:11:20.340670 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc16bbba-993f-4376-98ee-8e7770ace54e-utilities\") pod \"redhat-marketplace-8hhfj\" (UID: \"fc16bbba-993f-4376-98ee-8e7770ace54e\") " pod="openshift-marketplace/redhat-marketplace-8hhfj" Nov 26 01:11:20 crc kubenswrapper[4766]: I1126 01:11:20.340734 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qc7n\" (UniqueName: \"kubernetes.io/projected/fc16bbba-993f-4376-98ee-8e7770ace54e-kube-api-access-9qc7n\") pod \"redhat-marketplace-8hhfj\" (UID: \"fc16bbba-993f-4376-98ee-8e7770ace54e\") " pod="openshift-marketplace/redhat-marketplace-8hhfj" Nov 26 01:11:20 crc kubenswrapper[4766]: I1126 01:11:20.340791 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc16bbba-993f-4376-98ee-8e7770ace54e-catalog-content\") pod \"redhat-marketplace-8hhfj\" (UID: \"fc16bbba-993f-4376-98ee-8e7770ace54e\") " pod="openshift-marketplace/redhat-marketplace-8hhfj" Nov 26 01:11:20 crc kubenswrapper[4766]: I1126 01:11:20.341331 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc16bbba-993f-4376-98ee-8e7770ace54e-utilities\") pod \"redhat-marketplace-8hhfj\" (UID: \"fc16bbba-993f-4376-98ee-8e7770ace54e\") " pod="openshift-marketplace/redhat-marketplace-8hhfj" Nov 26 01:11:20 crc kubenswrapper[4766]: I1126 01:11:20.341362 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc16bbba-993f-4376-98ee-8e7770ace54e-catalog-content\") pod \"redhat-marketplace-8hhfj\" (UID: \"fc16bbba-993f-4376-98ee-8e7770ace54e\") " pod="openshift-marketplace/redhat-marketplace-8hhfj" Nov 26 01:11:20 crc kubenswrapper[4766]: I1126 01:11:20.379731 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qc7n\" (UniqueName: \"kubernetes.io/projected/fc16bbba-993f-4376-98ee-8e7770ace54e-kube-api-access-9qc7n\") pod \"redhat-marketplace-8hhfj\" (UID: \"fc16bbba-993f-4376-98ee-8e7770ace54e\") " pod="openshift-marketplace/redhat-marketplace-8hhfj" Nov 26 01:11:20 crc kubenswrapper[4766]: I1126 01:11:20.555597 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8hhfj" Nov 26 01:11:21 crc kubenswrapper[4766]: I1126 01:11:21.663578 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8hhfj"] Nov 26 01:11:22 crc kubenswrapper[4766]: I1126 01:11:22.489270 4766 generic.go:334] "Generic (PLEG): container finished" podID="fc16bbba-993f-4376-98ee-8e7770ace54e" containerID="c874ad8d0265cbc2129b6bd294403112deed9e8612ca551d26bb0884a0e1278f" exitCode=0 Nov 26 01:11:22 crc kubenswrapper[4766]: I1126 01:11:22.489394 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8hhfj" event={"ID":"fc16bbba-993f-4376-98ee-8e7770ace54e","Type":"ContainerDied","Data":"c874ad8d0265cbc2129b6bd294403112deed9e8612ca551d26bb0884a0e1278f"} Nov 26 01:11:22 crc kubenswrapper[4766]: I1126 01:11:22.489715 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8hhfj" event={"ID":"fc16bbba-993f-4376-98ee-8e7770ace54e","Type":"ContainerStarted","Data":"4b0580266a986973a63f9b0e40726b315e322c938ee77edc510c2347a4089ef5"} Nov 26 01:11:23 crc kubenswrapper[4766]: I1126 01:11:23.503259 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8hhfj" event={"ID":"fc16bbba-993f-4376-98ee-8e7770ace54e","Type":"ContainerStarted","Data":"435e79eeb56f5107a1e6e659ea5cc99ca8b8f2fd6bc9068bfeccf372f1165a9b"} Nov 26 01:11:24 crc kubenswrapper[4766]: I1126 01:11:24.516971 4766 generic.go:334] "Generic (PLEG): container finished" podID="fc16bbba-993f-4376-98ee-8e7770ace54e" containerID="435e79eeb56f5107a1e6e659ea5cc99ca8b8f2fd6bc9068bfeccf372f1165a9b" exitCode=0 Nov 26 01:11:24 crc kubenswrapper[4766]: I1126 01:11:24.517066 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8hhfj" event={"ID":"fc16bbba-993f-4376-98ee-8e7770ace54e","Type":"ContainerDied","Data":"435e79eeb56f5107a1e6e659ea5cc99ca8b8f2fd6bc9068bfeccf372f1165a9b"} Nov 26 01:11:25 crc kubenswrapper[4766]: I1126 01:11:25.544460 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8hhfj" event={"ID":"fc16bbba-993f-4376-98ee-8e7770ace54e","Type":"ContainerStarted","Data":"a6b09c8737a5e1a03c068689b393bc02e8ecf55e4ab270ef289f99e9e314b83c"} Nov 26 01:11:25 crc kubenswrapper[4766]: I1126 01:11:25.569263 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8hhfj" podStartSLOduration=3.157151626 podStartE2EDuration="5.569242953s" podCreationTimestamp="2025-11-26 01:11:20 +0000 UTC" firstStartedPulling="2025-11-26 01:11:22.492702812 +0000 UTC m=+2863.341473262" lastFinishedPulling="2025-11-26 01:11:24.904794119 +0000 UTC m=+2865.753564589" observedRunningTime="2025-11-26 01:11:25.562638541 +0000 UTC m=+2866.411408981" watchObservedRunningTime="2025-11-26 01:11:25.569242953 +0000 UTC m=+2866.418013393" Nov 26 01:11:30 crc kubenswrapper[4766]: I1126 01:11:30.556714 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8hhfj" Nov 26 01:11:30 crc kubenswrapper[4766]: I1126 01:11:30.557566 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8hhfj" Nov 26 01:11:30 crc kubenswrapper[4766]: I1126 01:11:30.634062 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8hhfj" Nov 26 01:11:30 crc kubenswrapper[4766]: I1126 01:11:30.711524 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8hhfj" Nov 26 01:11:30 crc kubenswrapper[4766]: I1126 01:11:30.887743 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8hhfj"] Nov 26 01:11:32 crc kubenswrapper[4766]: I1126 01:11:32.639470 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8hhfj" podUID="fc16bbba-993f-4376-98ee-8e7770ace54e" containerName="registry-server" containerID="cri-o://a6b09c8737a5e1a03c068689b393bc02e8ecf55e4ab270ef289f99e9e314b83c" gracePeriod=2 Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.295039 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8hhfj" Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.427780 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc16bbba-993f-4376-98ee-8e7770ace54e-catalog-content\") pod \"fc16bbba-993f-4376-98ee-8e7770ace54e\" (UID: \"fc16bbba-993f-4376-98ee-8e7770ace54e\") " Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.427916 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc16bbba-993f-4376-98ee-8e7770ace54e-utilities\") pod \"fc16bbba-993f-4376-98ee-8e7770ace54e\" (UID: \"fc16bbba-993f-4376-98ee-8e7770ace54e\") " Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.427986 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qc7n\" (UniqueName: \"kubernetes.io/projected/fc16bbba-993f-4376-98ee-8e7770ace54e-kube-api-access-9qc7n\") pod \"fc16bbba-993f-4376-98ee-8e7770ace54e\" (UID: \"fc16bbba-993f-4376-98ee-8e7770ace54e\") " Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.430561 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc16bbba-993f-4376-98ee-8e7770ace54e-utilities" (OuterVolumeSpecName: "utilities") pod "fc16bbba-993f-4376-98ee-8e7770ace54e" (UID: "fc16bbba-993f-4376-98ee-8e7770ace54e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.434111 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc16bbba-993f-4376-98ee-8e7770ace54e-kube-api-access-9qc7n" (OuterVolumeSpecName: "kube-api-access-9qc7n") pod "fc16bbba-993f-4376-98ee-8e7770ace54e" (UID: "fc16bbba-993f-4376-98ee-8e7770ace54e"). InnerVolumeSpecName "kube-api-access-9qc7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.456293 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc16bbba-993f-4376-98ee-8e7770ace54e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc16bbba-993f-4376-98ee-8e7770ace54e" (UID: "fc16bbba-993f-4376-98ee-8e7770ace54e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.530802 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qc7n\" (UniqueName: \"kubernetes.io/projected/fc16bbba-993f-4376-98ee-8e7770ace54e-kube-api-access-9qc7n\") on node \"crc\" DevicePath \"\"" Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.530847 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc16bbba-993f-4376-98ee-8e7770ace54e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.530860 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc16bbba-993f-4376-98ee-8e7770ace54e-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.654044 4766 generic.go:334] "Generic (PLEG): container finished" podID="fc16bbba-993f-4376-98ee-8e7770ace54e" containerID="a6b09c8737a5e1a03c068689b393bc02e8ecf55e4ab270ef289f99e9e314b83c" exitCode=0 Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.654105 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8hhfj" event={"ID":"fc16bbba-993f-4376-98ee-8e7770ace54e","Type":"ContainerDied","Data":"a6b09c8737a5e1a03c068689b393bc02e8ecf55e4ab270ef289f99e9e314b83c"} Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.654146 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8hhfj" event={"ID":"fc16bbba-993f-4376-98ee-8e7770ace54e","Type":"ContainerDied","Data":"4b0580266a986973a63f9b0e40726b315e322c938ee77edc510c2347a4089ef5"} Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.654155 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8hhfj" Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.654175 4766 scope.go:117] "RemoveContainer" containerID="a6b09c8737a5e1a03c068689b393bc02e8ecf55e4ab270ef289f99e9e314b83c" Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.675102 4766 scope.go:117] "RemoveContainer" containerID="435e79eeb56f5107a1e6e659ea5cc99ca8b8f2fd6bc9068bfeccf372f1165a9b" Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.710858 4766 scope.go:117] "RemoveContainer" containerID="c874ad8d0265cbc2129b6bd294403112deed9e8612ca551d26bb0884a0e1278f" Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.717871 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8hhfj"] Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.727789 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8hhfj"] Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.788897 4766 scope.go:117] "RemoveContainer" containerID="a6b09c8737a5e1a03c068689b393bc02e8ecf55e4ab270ef289f99e9e314b83c" Nov 26 01:11:33 crc kubenswrapper[4766]: E1126 01:11:33.789828 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6b09c8737a5e1a03c068689b393bc02e8ecf55e4ab270ef289f99e9e314b83c\": container with ID starting with a6b09c8737a5e1a03c068689b393bc02e8ecf55e4ab270ef289f99e9e314b83c not found: ID does not exist" containerID="a6b09c8737a5e1a03c068689b393bc02e8ecf55e4ab270ef289f99e9e314b83c" Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.789919 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6b09c8737a5e1a03c068689b393bc02e8ecf55e4ab270ef289f99e9e314b83c"} err="failed to get container status \"a6b09c8737a5e1a03c068689b393bc02e8ecf55e4ab270ef289f99e9e314b83c\": rpc error: code = NotFound desc = could not find container \"a6b09c8737a5e1a03c068689b393bc02e8ecf55e4ab270ef289f99e9e314b83c\": container with ID starting with a6b09c8737a5e1a03c068689b393bc02e8ecf55e4ab270ef289f99e9e314b83c not found: ID does not exist" Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.789971 4766 scope.go:117] "RemoveContainer" containerID="435e79eeb56f5107a1e6e659ea5cc99ca8b8f2fd6bc9068bfeccf372f1165a9b" Nov 26 01:11:33 crc kubenswrapper[4766]: E1126 01:11:33.790776 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"435e79eeb56f5107a1e6e659ea5cc99ca8b8f2fd6bc9068bfeccf372f1165a9b\": container with ID starting with 435e79eeb56f5107a1e6e659ea5cc99ca8b8f2fd6bc9068bfeccf372f1165a9b not found: ID does not exist" containerID="435e79eeb56f5107a1e6e659ea5cc99ca8b8f2fd6bc9068bfeccf372f1165a9b" Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.790830 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"435e79eeb56f5107a1e6e659ea5cc99ca8b8f2fd6bc9068bfeccf372f1165a9b"} err="failed to get container status \"435e79eeb56f5107a1e6e659ea5cc99ca8b8f2fd6bc9068bfeccf372f1165a9b\": rpc error: code = NotFound desc = could not find container \"435e79eeb56f5107a1e6e659ea5cc99ca8b8f2fd6bc9068bfeccf372f1165a9b\": container with ID starting with 435e79eeb56f5107a1e6e659ea5cc99ca8b8f2fd6bc9068bfeccf372f1165a9b not found: ID does not exist" Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.790869 4766 scope.go:117] "RemoveContainer" containerID="c874ad8d0265cbc2129b6bd294403112deed9e8612ca551d26bb0884a0e1278f" Nov 26 01:11:33 crc kubenswrapper[4766]: E1126 01:11:33.791320 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c874ad8d0265cbc2129b6bd294403112deed9e8612ca551d26bb0884a0e1278f\": container with ID starting with c874ad8d0265cbc2129b6bd294403112deed9e8612ca551d26bb0884a0e1278f not found: ID does not exist" containerID="c874ad8d0265cbc2129b6bd294403112deed9e8612ca551d26bb0884a0e1278f" Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.791374 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c874ad8d0265cbc2129b6bd294403112deed9e8612ca551d26bb0884a0e1278f"} err="failed to get container status \"c874ad8d0265cbc2129b6bd294403112deed9e8612ca551d26bb0884a0e1278f\": rpc error: code = NotFound desc = could not find container \"c874ad8d0265cbc2129b6bd294403112deed9e8612ca551d26bb0884a0e1278f\": container with ID starting with c874ad8d0265cbc2129b6bd294403112deed9e8612ca551d26bb0884a0e1278f not found: ID does not exist" Nov 26 01:11:33 crc kubenswrapper[4766]: I1126 01:11:33.844243 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc16bbba-993f-4376-98ee-8e7770ace54e" path="/var/lib/kubelet/pods/fc16bbba-993f-4376-98ee-8e7770ace54e/volumes" Nov 26 01:11:41 crc kubenswrapper[4766]: I1126 01:11:41.479773 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:11:41 crc kubenswrapper[4766]: I1126 01:11:41.480711 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:12:01 crc kubenswrapper[4766]: I1126 01:12:01.029252 4766 generic.go:334] "Generic (PLEG): container finished" podID="afe63c23-a92e-4701-a422-d431ed80d42d" containerID="62544c5a4746cf33b2e88c463d3bb5e09fe39de843456d753c9aa80b23f092bb" exitCode=0 Nov 26 01:12:01 crc kubenswrapper[4766]: I1126 01:12:01.029383 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" event={"ID":"afe63c23-a92e-4701-a422-d431ed80d42d","Type":"ContainerDied","Data":"62544c5a4746cf33b2e88c463d3bb5e09fe39de843456d753c9aa80b23f092bb"} Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.586063 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.724726 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ceilometer-ipmi-config-data-2\") pod \"afe63c23-a92e-4701-a422-d431ed80d42d\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.724982 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ceilometer-ipmi-config-data-1\") pod \"afe63c23-a92e-4701-a422-d431ed80d42d\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.725027 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-inventory\") pod \"afe63c23-a92e-4701-a422-d431ed80d42d\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.725049 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ssh-key\") pod \"afe63c23-a92e-4701-a422-d431ed80d42d\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.725122 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-telemetry-power-monitoring-combined-ca-bundle\") pod \"afe63c23-a92e-4701-a422-d431ed80d42d\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.725160 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ceilometer-ipmi-config-data-0\") pod \"afe63c23-a92e-4701-a422-d431ed80d42d\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.725279 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qx5qz\" (UniqueName: \"kubernetes.io/projected/afe63c23-a92e-4701-a422-d431ed80d42d-kube-api-access-qx5qz\") pod \"afe63c23-a92e-4701-a422-d431ed80d42d\" (UID: \"afe63c23-a92e-4701-a422-d431ed80d42d\") " Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.734074 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "afe63c23-a92e-4701-a422-d431ed80d42d" (UID: "afe63c23-a92e-4701-a422-d431ed80d42d"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.746380 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afe63c23-a92e-4701-a422-d431ed80d42d-kube-api-access-qx5qz" (OuterVolumeSpecName: "kube-api-access-qx5qz") pod "afe63c23-a92e-4701-a422-d431ed80d42d" (UID: "afe63c23-a92e-4701-a422-d431ed80d42d"). InnerVolumeSpecName "kube-api-access-qx5qz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.762142 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ceilometer-ipmi-config-data-1" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-1") pod "afe63c23-a92e-4701-a422-d431ed80d42d" (UID: "afe63c23-a92e-4701-a422-d431ed80d42d"). InnerVolumeSpecName "ceilometer-ipmi-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.763424 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "afe63c23-a92e-4701-a422-d431ed80d42d" (UID: "afe63c23-a92e-4701-a422-d431ed80d42d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.764166 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ceilometer-ipmi-config-data-0" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-0") pod "afe63c23-a92e-4701-a422-d431ed80d42d" (UID: "afe63c23-a92e-4701-a422-d431ed80d42d"). InnerVolumeSpecName "ceilometer-ipmi-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.767025 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-inventory" (OuterVolumeSpecName: "inventory") pod "afe63c23-a92e-4701-a422-d431ed80d42d" (UID: "afe63c23-a92e-4701-a422-d431ed80d42d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.776644 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ceilometer-ipmi-config-data-2" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-2") pod "afe63c23-a92e-4701-a422-d431ed80d42d" (UID: "afe63c23-a92e-4701-a422-d431ed80d42d"). InnerVolumeSpecName "ceilometer-ipmi-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.841982 4766 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ceilometer-ipmi-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.842029 4766 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ceilometer-ipmi-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.842050 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.842063 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.842078 4766 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.842093 4766 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/afe63c23-a92e-4701-a422-d431ed80d42d-ceilometer-ipmi-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:12:02 crc kubenswrapper[4766]: I1126 01:12:02.842106 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qx5qz\" (UniqueName: \"kubernetes.io/projected/afe63c23-a92e-4701-a422-d431ed80d42d-kube-api-access-qx5qz\") on node \"crc\" DevicePath \"\"" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.058011 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" event={"ID":"afe63c23-a92e-4701-a422-d431ed80d42d","Type":"ContainerDied","Data":"5397693f9a4f7d427ffb4d87f47831d4ff1d3a41fec6592c5eff0db7849116b6"} Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.058054 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5397693f9a4f7d427ffb4d87f47831d4ff1d3a41fec6592c5eff0db7849116b6" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.058093 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.201827 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr"] Nov 26 01:12:03 crc kubenswrapper[4766]: E1126 01:12:03.202379 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afe63c23-a92e-4701-a422-d431ed80d42d" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.202408 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="afe63c23-a92e-4701-a422-d431ed80d42d" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Nov 26 01:12:03 crc kubenswrapper[4766]: E1126 01:12:03.202434 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc16bbba-993f-4376-98ee-8e7770ace54e" containerName="extract-content" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.202442 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc16bbba-993f-4376-98ee-8e7770ace54e" containerName="extract-content" Nov 26 01:12:03 crc kubenswrapper[4766]: E1126 01:12:03.202471 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc16bbba-993f-4376-98ee-8e7770ace54e" containerName="registry-server" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.202479 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc16bbba-993f-4376-98ee-8e7770ace54e" containerName="registry-server" Nov 26 01:12:03 crc kubenswrapper[4766]: E1126 01:12:03.202509 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc16bbba-993f-4376-98ee-8e7770ace54e" containerName="extract-utilities" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.202519 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc16bbba-993f-4376-98ee-8e7770ace54e" containerName="extract-utilities" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.202876 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="afe63c23-a92e-4701-a422-d431ed80d42d" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.202916 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc16bbba-993f-4376-98ee-8e7770ace54e" containerName="registry-server" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.203766 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.209131 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.209197 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.209602 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.209993 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.213014 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-compute-config-data" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.215180 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr"] Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.355745 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsl79\" (UniqueName: \"kubernetes.io/projected/e5921e8a-b3ce-4840-8211-a96134855a2f-kube-api-access-zsl79\") pod \"logging-edpm-deployment-openstack-edpm-ipam-wm8zr\" (UID: \"e5921e8a-b3ce-4840-8211-a96134855a2f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.355806 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-wm8zr\" (UID: \"e5921e8a-b3ce-4840-8211-a96134855a2f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.355831 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-wm8zr\" (UID: \"e5921e8a-b3ce-4840-8211-a96134855a2f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.355956 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-wm8zr\" (UID: \"e5921e8a-b3ce-4840-8211-a96134855a2f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.356100 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-wm8zr\" (UID: \"e5921e8a-b3ce-4840-8211-a96134855a2f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.457465 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsl79\" (UniqueName: \"kubernetes.io/projected/e5921e8a-b3ce-4840-8211-a96134855a2f-kube-api-access-zsl79\") pod \"logging-edpm-deployment-openstack-edpm-ipam-wm8zr\" (UID: \"e5921e8a-b3ce-4840-8211-a96134855a2f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.457512 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-wm8zr\" (UID: \"e5921e8a-b3ce-4840-8211-a96134855a2f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.457536 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-wm8zr\" (UID: \"e5921e8a-b3ce-4840-8211-a96134855a2f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.457583 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-wm8zr\" (UID: \"e5921e8a-b3ce-4840-8211-a96134855a2f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.457745 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-wm8zr\" (UID: \"e5921e8a-b3ce-4840-8211-a96134855a2f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.470855 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-wm8zr\" (UID: \"e5921e8a-b3ce-4840-8211-a96134855a2f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.470955 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-wm8zr\" (UID: \"e5921e8a-b3ce-4840-8211-a96134855a2f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.471241 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-wm8zr\" (UID: \"e5921e8a-b3ce-4840-8211-a96134855a2f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.471935 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-wm8zr\" (UID: \"e5921e8a-b3ce-4840-8211-a96134855a2f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.473818 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsl79\" (UniqueName: \"kubernetes.io/projected/e5921e8a-b3ce-4840-8211-a96134855a2f-kube-api-access-zsl79\") pod \"logging-edpm-deployment-openstack-edpm-ipam-wm8zr\" (UID: \"e5921e8a-b3ce-4840-8211-a96134855a2f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" Nov 26 01:12:03 crc kubenswrapper[4766]: I1126 01:12:03.526420 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" Nov 26 01:12:04 crc kubenswrapper[4766]: W1126 01:12:04.208859 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5921e8a_b3ce_4840_8211_a96134855a2f.slice/crio-b80c6aed42bf814277c8443117d49bc4a4b64e694c01e46c2af1998f7bd2009d WatchSource:0}: Error finding container b80c6aed42bf814277c8443117d49bc4a4b64e694c01e46c2af1998f7bd2009d: Status 404 returned error can't find the container with id b80c6aed42bf814277c8443117d49bc4a4b64e694c01e46c2af1998f7bd2009d Nov 26 01:12:04 crc kubenswrapper[4766]: I1126 01:12:04.210769 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 01:12:04 crc kubenswrapper[4766]: I1126 01:12:04.212728 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr"] Nov 26 01:12:05 crc kubenswrapper[4766]: I1126 01:12:05.080834 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" event={"ID":"e5921e8a-b3ce-4840-8211-a96134855a2f","Type":"ContainerStarted","Data":"b80c6aed42bf814277c8443117d49bc4a4b64e694c01e46c2af1998f7bd2009d"} Nov 26 01:12:06 crc kubenswrapper[4766]: I1126 01:12:06.100065 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" event={"ID":"e5921e8a-b3ce-4840-8211-a96134855a2f","Type":"ContainerStarted","Data":"f8e1ff06e2703e15b1e93aff1dae5cd82d32771db4a0ca0aa591930ef4840c7f"} Nov 26 01:12:06 crc kubenswrapper[4766]: I1126 01:12:06.131063 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" podStartSLOduration=2.465106739 podStartE2EDuration="3.131039899s" podCreationTimestamp="2025-11-26 01:12:03 +0000 UTC" firstStartedPulling="2025-11-26 01:12:04.21055192 +0000 UTC m=+2905.059322350" lastFinishedPulling="2025-11-26 01:12:04.87648504 +0000 UTC m=+2905.725255510" observedRunningTime="2025-11-26 01:12:06.11889257 +0000 UTC m=+2906.967663020" watchObservedRunningTime="2025-11-26 01:12:06.131039899 +0000 UTC m=+2906.979810339" Nov 26 01:12:11 crc kubenswrapper[4766]: I1126 01:12:11.479857 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:12:11 crc kubenswrapper[4766]: I1126 01:12:11.480359 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:12:26 crc kubenswrapper[4766]: I1126 01:12:26.370154 4766 generic.go:334] "Generic (PLEG): container finished" podID="e5921e8a-b3ce-4840-8211-a96134855a2f" containerID="f8e1ff06e2703e15b1e93aff1dae5cd82d32771db4a0ca0aa591930ef4840c7f" exitCode=0 Nov 26 01:12:26 crc kubenswrapper[4766]: I1126 01:12:26.370292 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" event={"ID":"e5921e8a-b3ce-4840-8211-a96134855a2f","Type":"ContainerDied","Data":"f8e1ff06e2703e15b1e93aff1dae5cd82d32771db4a0ca0aa591930ef4840c7f"} Nov 26 01:12:27 crc kubenswrapper[4766]: I1126 01:12:27.909897 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" Nov 26 01:12:28 crc kubenswrapper[4766]: I1126 01:12:28.059807 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsl79\" (UniqueName: \"kubernetes.io/projected/e5921e8a-b3ce-4840-8211-a96134855a2f-kube-api-access-zsl79\") pod \"e5921e8a-b3ce-4840-8211-a96134855a2f\" (UID: \"e5921e8a-b3ce-4840-8211-a96134855a2f\") " Nov 26 01:12:28 crc kubenswrapper[4766]: I1126 01:12:28.059925 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-inventory\") pod \"e5921e8a-b3ce-4840-8211-a96134855a2f\" (UID: \"e5921e8a-b3ce-4840-8211-a96134855a2f\") " Nov 26 01:12:28 crc kubenswrapper[4766]: I1126 01:12:28.060003 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-logging-compute-config-data-0\") pod \"e5921e8a-b3ce-4840-8211-a96134855a2f\" (UID: \"e5921e8a-b3ce-4840-8211-a96134855a2f\") " Nov 26 01:12:28 crc kubenswrapper[4766]: I1126 01:12:28.060109 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-logging-compute-config-data-1\") pod \"e5921e8a-b3ce-4840-8211-a96134855a2f\" (UID: \"e5921e8a-b3ce-4840-8211-a96134855a2f\") " Nov 26 01:12:28 crc kubenswrapper[4766]: I1126 01:12:28.060169 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-ssh-key\") pod \"e5921e8a-b3ce-4840-8211-a96134855a2f\" (UID: \"e5921e8a-b3ce-4840-8211-a96134855a2f\") " Nov 26 01:12:28 crc kubenswrapper[4766]: I1126 01:12:28.068097 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5921e8a-b3ce-4840-8211-a96134855a2f-kube-api-access-zsl79" (OuterVolumeSpecName: "kube-api-access-zsl79") pod "e5921e8a-b3ce-4840-8211-a96134855a2f" (UID: "e5921e8a-b3ce-4840-8211-a96134855a2f"). InnerVolumeSpecName "kube-api-access-zsl79". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:12:28 crc kubenswrapper[4766]: I1126 01:12:28.091465 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-inventory" (OuterVolumeSpecName: "inventory") pod "e5921e8a-b3ce-4840-8211-a96134855a2f" (UID: "e5921e8a-b3ce-4840-8211-a96134855a2f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:12:28 crc kubenswrapper[4766]: I1126 01:12:28.117096 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-logging-compute-config-data-1" (OuterVolumeSpecName: "logging-compute-config-data-1") pod "e5921e8a-b3ce-4840-8211-a96134855a2f" (UID: "e5921e8a-b3ce-4840-8211-a96134855a2f"). InnerVolumeSpecName "logging-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:12:28 crc kubenswrapper[4766]: I1126 01:12:28.130488 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e5921e8a-b3ce-4840-8211-a96134855a2f" (UID: "e5921e8a-b3ce-4840-8211-a96134855a2f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:12:28 crc kubenswrapper[4766]: I1126 01:12:28.132183 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-logging-compute-config-data-0" (OuterVolumeSpecName: "logging-compute-config-data-0") pod "e5921e8a-b3ce-4840-8211-a96134855a2f" (UID: "e5921e8a-b3ce-4840-8211-a96134855a2f"). InnerVolumeSpecName "logging-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:12:28 crc kubenswrapper[4766]: I1126 01:12:28.163563 4766 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-logging-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 26 01:12:28 crc kubenswrapper[4766]: I1126 01:12:28.163615 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:12:28 crc kubenswrapper[4766]: I1126 01:12:28.163639 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsl79\" (UniqueName: \"kubernetes.io/projected/e5921e8a-b3ce-4840-8211-a96134855a2f-kube-api-access-zsl79\") on node \"crc\" DevicePath \"\"" Nov 26 01:12:28 crc kubenswrapper[4766]: I1126 01:12:28.163683 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:12:28 crc kubenswrapper[4766]: I1126 01:12:28.163703 4766 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e5921e8a-b3ce-4840-8211-a96134855a2f-logging-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:12:28 crc kubenswrapper[4766]: I1126 01:12:28.400484 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" event={"ID":"e5921e8a-b3ce-4840-8211-a96134855a2f","Type":"ContainerDied","Data":"b80c6aed42bf814277c8443117d49bc4a4b64e694c01e46c2af1998f7bd2009d"} Nov 26 01:12:28 crc kubenswrapper[4766]: I1126 01:12:28.400523 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b80c6aed42bf814277c8443117d49bc4a4b64e694c01e46c2af1998f7bd2009d" Nov 26 01:12:28 crc kubenswrapper[4766]: I1126 01:12:28.400580 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr" Nov 26 01:12:41 crc kubenswrapper[4766]: I1126 01:12:41.479865 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:12:41 crc kubenswrapper[4766]: I1126 01:12:41.480413 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:12:41 crc kubenswrapper[4766]: I1126 01:12:41.480463 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 01:12:41 crc kubenswrapper[4766]: I1126 01:12:41.481213 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 01:12:41 crc kubenswrapper[4766]: I1126 01:12:41.481268 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" gracePeriod=600 Nov 26 01:12:41 crc kubenswrapper[4766]: E1126 01:12:41.613144 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:12:42 crc kubenswrapper[4766]: I1126 01:12:42.581380 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" exitCode=0 Nov 26 01:12:42 crc kubenswrapper[4766]: I1126 01:12:42.581851 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50"} Nov 26 01:12:42 crc kubenswrapper[4766]: I1126 01:12:42.582062 4766 scope.go:117] "RemoveContainer" containerID="3f97e605694d4755ce21053291eee5a113d9b02cc30e9ddf766f118e2989829b" Nov 26 01:12:42 crc kubenswrapper[4766]: I1126 01:12:42.583287 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:12:42 crc kubenswrapper[4766]: E1126 01:12:42.583920 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:12:54 crc kubenswrapper[4766]: I1126 01:12:54.827686 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:12:54 crc kubenswrapper[4766]: E1126 01:12:54.828463 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:13:09 crc kubenswrapper[4766]: I1126 01:13:09.856252 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:13:09 crc kubenswrapper[4766]: E1126 01:13:09.857499 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:13:24 crc kubenswrapper[4766]: I1126 01:13:24.827277 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:13:24 crc kubenswrapper[4766]: E1126 01:13:24.828480 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:13:37 crc kubenswrapper[4766]: I1126 01:13:37.828386 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:13:37 crc kubenswrapper[4766]: E1126 01:13:37.829125 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:13:51 crc kubenswrapper[4766]: I1126 01:13:51.828064 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:13:51 crc kubenswrapper[4766]: E1126 01:13:51.829036 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:14:05 crc kubenswrapper[4766]: I1126 01:14:05.827538 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:14:05 crc kubenswrapper[4766]: E1126 01:14:05.828584 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:14:10 crc kubenswrapper[4766]: I1126 01:14:10.960120 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f5rgx"] Nov 26 01:14:10 crc kubenswrapper[4766]: E1126 01:14:10.961334 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5921e8a-b3ce-4840-8211-a96134855a2f" containerName="logging-edpm-deployment-openstack-edpm-ipam" Nov 26 01:14:10 crc kubenswrapper[4766]: I1126 01:14:10.961356 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5921e8a-b3ce-4840-8211-a96134855a2f" containerName="logging-edpm-deployment-openstack-edpm-ipam" Nov 26 01:14:10 crc kubenswrapper[4766]: I1126 01:14:10.961623 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5921e8a-b3ce-4840-8211-a96134855a2f" containerName="logging-edpm-deployment-openstack-edpm-ipam" Nov 26 01:14:10 crc kubenswrapper[4766]: I1126 01:14:10.963847 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f5rgx" Nov 26 01:14:11 crc kubenswrapper[4766]: I1126 01:14:11.005678 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f5rgx"] Nov 26 01:14:11 crc kubenswrapper[4766]: I1126 01:14:11.051093 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9xwv\" (UniqueName: \"kubernetes.io/projected/d39fa104-3076-45d1-92a1-3c0dcef1c7e3-kube-api-access-q9xwv\") pod \"redhat-operators-f5rgx\" (UID: \"d39fa104-3076-45d1-92a1-3c0dcef1c7e3\") " pod="openshift-marketplace/redhat-operators-f5rgx" Nov 26 01:14:11 crc kubenswrapper[4766]: I1126 01:14:11.051284 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d39fa104-3076-45d1-92a1-3c0dcef1c7e3-utilities\") pod \"redhat-operators-f5rgx\" (UID: \"d39fa104-3076-45d1-92a1-3c0dcef1c7e3\") " pod="openshift-marketplace/redhat-operators-f5rgx" Nov 26 01:14:11 crc kubenswrapper[4766]: I1126 01:14:11.051336 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d39fa104-3076-45d1-92a1-3c0dcef1c7e3-catalog-content\") pod \"redhat-operators-f5rgx\" (UID: \"d39fa104-3076-45d1-92a1-3c0dcef1c7e3\") " pod="openshift-marketplace/redhat-operators-f5rgx" Nov 26 01:14:11 crc kubenswrapper[4766]: I1126 01:14:11.152818 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d39fa104-3076-45d1-92a1-3c0dcef1c7e3-utilities\") pod \"redhat-operators-f5rgx\" (UID: \"d39fa104-3076-45d1-92a1-3c0dcef1c7e3\") " pod="openshift-marketplace/redhat-operators-f5rgx" Nov 26 01:14:11 crc kubenswrapper[4766]: I1126 01:14:11.152876 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d39fa104-3076-45d1-92a1-3c0dcef1c7e3-catalog-content\") pod \"redhat-operators-f5rgx\" (UID: \"d39fa104-3076-45d1-92a1-3c0dcef1c7e3\") " pod="openshift-marketplace/redhat-operators-f5rgx" Nov 26 01:14:11 crc kubenswrapper[4766]: I1126 01:14:11.153107 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9xwv\" (UniqueName: \"kubernetes.io/projected/d39fa104-3076-45d1-92a1-3c0dcef1c7e3-kube-api-access-q9xwv\") pod \"redhat-operators-f5rgx\" (UID: \"d39fa104-3076-45d1-92a1-3c0dcef1c7e3\") " pod="openshift-marketplace/redhat-operators-f5rgx" Nov 26 01:14:11 crc kubenswrapper[4766]: I1126 01:14:11.153429 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d39fa104-3076-45d1-92a1-3c0dcef1c7e3-utilities\") pod \"redhat-operators-f5rgx\" (UID: \"d39fa104-3076-45d1-92a1-3c0dcef1c7e3\") " pod="openshift-marketplace/redhat-operators-f5rgx" Nov 26 01:14:11 crc kubenswrapper[4766]: I1126 01:14:11.153458 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d39fa104-3076-45d1-92a1-3c0dcef1c7e3-catalog-content\") pod \"redhat-operators-f5rgx\" (UID: \"d39fa104-3076-45d1-92a1-3c0dcef1c7e3\") " pod="openshift-marketplace/redhat-operators-f5rgx" Nov 26 01:14:11 crc kubenswrapper[4766]: I1126 01:14:11.172635 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9xwv\" (UniqueName: \"kubernetes.io/projected/d39fa104-3076-45d1-92a1-3c0dcef1c7e3-kube-api-access-q9xwv\") pod \"redhat-operators-f5rgx\" (UID: \"d39fa104-3076-45d1-92a1-3c0dcef1c7e3\") " pod="openshift-marketplace/redhat-operators-f5rgx" Nov 26 01:14:11 crc kubenswrapper[4766]: I1126 01:14:11.340545 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f5rgx" Nov 26 01:14:11 crc kubenswrapper[4766]: I1126 01:14:11.846178 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f5rgx"] Nov 26 01:14:12 crc kubenswrapper[4766]: I1126 01:14:12.195013 4766 generic.go:334] "Generic (PLEG): container finished" podID="d39fa104-3076-45d1-92a1-3c0dcef1c7e3" containerID="5c91575c0c0fcdced52dfe6f7a2a8774b5a6a78f572724b4d5f719bcb220797c" exitCode=0 Nov 26 01:14:12 crc kubenswrapper[4766]: I1126 01:14:12.195107 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5rgx" event={"ID":"d39fa104-3076-45d1-92a1-3c0dcef1c7e3","Type":"ContainerDied","Data":"5c91575c0c0fcdced52dfe6f7a2a8774b5a6a78f572724b4d5f719bcb220797c"} Nov 26 01:14:12 crc kubenswrapper[4766]: I1126 01:14:12.195286 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5rgx" event={"ID":"d39fa104-3076-45d1-92a1-3c0dcef1c7e3","Type":"ContainerStarted","Data":"83cd57747186766aa32b58222180ed97a2f51f4991ec61fde3a06a8dfe733bac"} Nov 26 01:14:13 crc kubenswrapper[4766]: I1126 01:14:13.209208 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5rgx" event={"ID":"d39fa104-3076-45d1-92a1-3c0dcef1c7e3","Type":"ContainerStarted","Data":"01711d4aa0a3ac6a86faf541d17109aa609865428053be65c85141cd3849526f"} Nov 26 01:14:17 crc kubenswrapper[4766]: E1126 01:14:17.085862 4766 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd39fa104_3076_45d1_92a1_3c0dcef1c7e3.slice/crio-conmon-01711d4aa0a3ac6a86faf541d17109aa609865428053be65c85141cd3849526f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd39fa104_3076_45d1_92a1_3c0dcef1c7e3.slice/crio-01711d4aa0a3ac6a86faf541d17109aa609865428053be65c85141cd3849526f.scope\": RecentStats: unable to find data in memory cache]" Nov 26 01:14:17 crc kubenswrapper[4766]: I1126 01:14:17.278065 4766 generic.go:334] "Generic (PLEG): container finished" podID="d39fa104-3076-45d1-92a1-3c0dcef1c7e3" containerID="01711d4aa0a3ac6a86faf541d17109aa609865428053be65c85141cd3849526f" exitCode=0 Nov 26 01:14:17 crc kubenswrapper[4766]: I1126 01:14:17.278174 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5rgx" event={"ID":"d39fa104-3076-45d1-92a1-3c0dcef1c7e3","Type":"ContainerDied","Data":"01711d4aa0a3ac6a86faf541d17109aa609865428053be65c85141cd3849526f"} Nov 26 01:14:18 crc kubenswrapper[4766]: I1126 01:14:18.296296 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5rgx" event={"ID":"d39fa104-3076-45d1-92a1-3c0dcef1c7e3","Type":"ContainerStarted","Data":"9fd55883100d38b20adbf353ec41ef7f7864cd9ff23374f28ff6f6a3a793b26f"} Nov 26 01:14:18 crc kubenswrapper[4766]: I1126 01:14:18.326435 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f5rgx" podStartSLOduration=2.825179899 podStartE2EDuration="8.326410088s" podCreationTimestamp="2025-11-26 01:14:10 +0000 UTC" firstStartedPulling="2025-11-26 01:14:12.196732251 +0000 UTC m=+3033.045502681" lastFinishedPulling="2025-11-26 01:14:17.69796243 +0000 UTC m=+3038.546732870" observedRunningTime="2025-11-26 01:14:18.31227278 +0000 UTC m=+3039.161043250" watchObservedRunningTime="2025-11-26 01:14:18.326410088 +0000 UTC m=+3039.175180528" Nov 26 01:14:20 crc kubenswrapper[4766]: I1126 01:14:20.827941 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:14:20 crc kubenswrapper[4766]: E1126 01:14:20.829006 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:14:21 crc kubenswrapper[4766]: I1126 01:14:21.340979 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f5rgx" Nov 26 01:14:21 crc kubenswrapper[4766]: I1126 01:14:21.341357 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f5rgx" Nov 26 01:14:22 crc kubenswrapper[4766]: I1126 01:14:22.417072 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f5rgx" podUID="d39fa104-3076-45d1-92a1-3c0dcef1c7e3" containerName="registry-server" probeResult="failure" output=< Nov 26 01:14:22 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 01:14:22 crc kubenswrapper[4766]: > Nov 26 01:14:31 crc kubenswrapper[4766]: I1126 01:14:31.414763 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f5rgx" Nov 26 01:14:31 crc kubenswrapper[4766]: I1126 01:14:31.478974 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f5rgx" Nov 26 01:14:32 crc kubenswrapper[4766]: I1126 01:14:32.561414 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f5rgx"] Nov 26 01:14:32 crc kubenswrapper[4766]: I1126 01:14:32.561862 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f5rgx" podUID="d39fa104-3076-45d1-92a1-3c0dcef1c7e3" containerName="registry-server" containerID="cri-o://9fd55883100d38b20adbf353ec41ef7f7864cd9ff23374f28ff6f6a3a793b26f" gracePeriod=2 Nov 26 01:14:32 crc kubenswrapper[4766]: I1126 01:14:32.828794 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:14:32 crc kubenswrapper[4766]: E1126 01:14:32.829492 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.108952 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f5rgx" Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.179872 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d39fa104-3076-45d1-92a1-3c0dcef1c7e3-utilities\") pod \"d39fa104-3076-45d1-92a1-3c0dcef1c7e3\" (UID: \"d39fa104-3076-45d1-92a1-3c0dcef1c7e3\") " Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.180088 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9xwv\" (UniqueName: \"kubernetes.io/projected/d39fa104-3076-45d1-92a1-3c0dcef1c7e3-kube-api-access-q9xwv\") pod \"d39fa104-3076-45d1-92a1-3c0dcef1c7e3\" (UID: \"d39fa104-3076-45d1-92a1-3c0dcef1c7e3\") " Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.180143 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d39fa104-3076-45d1-92a1-3c0dcef1c7e3-catalog-content\") pod \"d39fa104-3076-45d1-92a1-3c0dcef1c7e3\" (UID: \"d39fa104-3076-45d1-92a1-3c0dcef1c7e3\") " Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.181679 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d39fa104-3076-45d1-92a1-3c0dcef1c7e3-utilities" (OuterVolumeSpecName: "utilities") pod "d39fa104-3076-45d1-92a1-3c0dcef1c7e3" (UID: "d39fa104-3076-45d1-92a1-3c0dcef1c7e3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.186237 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d39fa104-3076-45d1-92a1-3c0dcef1c7e3-kube-api-access-q9xwv" (OuterVolumeSpecName: "kube-api-access-q9xwv") pod "d39fa104-3076-45d1-92a1-3c0dcef1c7e3" (UID: "d39fa104-3076-45d1-92a1-3c0dcef1c7e3"). InnerVolumeSpecName "kube-api-access-q9xwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.282284 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d39fa104-3076-45d1-92a1-3c0dcef1c7e3-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.282317 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9xwv\" (UniqueName: \"kubernetes.io/projected/d39fa104-3076-45d1-92a1-3c0dcef1c7e3-kube-api-access-q9xwv\") on node \"crc\" DevicePath \"\"" Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.290443 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d39fa104-3076-45d1-92a1-3c0dcef1c7e3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d39fa104-3076-45d1-92a1-3c0dcef1c7e3" (UID: "d39fa104-3076-45d1-92a1-3c0dcef1c7e3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.383754 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d39fa104-3076-45d1-92a1-3c0dcef1c7e3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.496106 4766 generic.go:334] "Generic (PLEG): container finished" podID="d39fa104-3076-45d1-92a1-3c0dcef1c7e3" containerID="9fd55883100d38b20adbf353ec41ef7f7864cd9ff23374f28ff6f6a3a793b26f" exitCode=0 Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.496153 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5rgx" event={"ID":"d39fa104-3076-45d1-92a1-3c0dcef1c7e3","Type":"ContainerDied","Data":"9fd55883100d38b20adbf353ec41ef7f7864cd9ff23374f28ff6f6a3a793b26f"} Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.496187 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5rgx" event={"ID":"d39fa104-3076-45d1-92a1-3c0dcef1c7e3","Type":"ContainerDied","Data":"83cd57747186766aa32b58222180ed97a2f51f4991ec61fde3a06a8dfe733bac"} Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.496205 4766 scope.go:117] "RemoveContainer" containerID="9fd55883100d38b20adbf353ec41ef7f7864cd9ff23374f28ff6f6a3a793b26f" Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.496200 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f5rgx" Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.545144 4766 scope.go:117] "RemoveContainer" containerID="01711d4aa0a3ac6a86faf541d17109aa609865428053be65c85141cd3849526f" Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.557901 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f5rgx"] Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.567334 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f5rgx"] Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.584673 4766 scope.go:117] "RemoveContainer" containerID="5c91575c0c0fcdced52dfe6f7a2a8774b5a6a78f572724b4d5f719bcb220797c" Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.622380 4766 scope.go:117] "RemoveContainer" containerID="9fd55883100d38b20adbf353ec41ef7f7864cd9ff23374f28ff6f6a3a793b26f" Nov 26 01:14:33 crc kubenswrapper[4766]: E1126 01:14:33.666312 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fd55883100d38b20adbf353ec41ef7f7864cd9ff23374f28ff6f6a3a793b26f\": container with ID starting with 9fd55883100d38b20adbf353ec41ef7f7864cd9ff23374f28ff6f6a3a793b26f not found: ID does not exist" containerID="9fd55883100d38b20adbf353ec41ef7f7864cd9ff23374f28ff6f6a3a793b26f" Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.666381 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fd55883100d38b20adbf353ec41ef7f7864cd9ff23374f28ff6f6a3a793b26f"} err="failed to get container status \"9fd55883100d38b20adbf353ec41ef7f7864cd9ff23374f28ff6f6a3a793b26f\": rpc error: code = NotFound desc = could not find container \"9fd55883100d38b20adbf353ec41ef7f7864cd9ff23374f28ff6f6a3a793b26f\": container with ID starting with 9fd55883100d38b20adbf353ec41ef7f7864cd9ff23374f28ff6f6a3a793b26f not found: ID does not exist" Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.666425 4766 scope.go:117] "RemoveContainer" containerID="01711d4aa0a3ac6a86faf541d17109aa609865428053be65c85141cd3849526f" Nov 26 01:14:33 crc kubenswrapper[4766]: E1126 01:14:33.667093 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01711d4aa0a3ac6a86faf541d17109aa609865428053be65c85141cd3849526f\": container with ID starting with 01711d4aa0a3ac6a86faf541d17109aa609865428053be65c85141cd3849526f not found: ID does not exist" containerID="01711d4aa0a3ac6a86faf541d17109aa609865428053be65c85141cd3849526f" Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.667147 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01711d4aa0a3ac6a86faf541d17109aa609865428053be65c85141cd3849526f"} err="failed to get container status \"01711d4aa0a3ac6a86faf541d17109aa609865428053be65c85141cd3849526f\": rpc error: code = NotFound desc = could not find container \"01711d4aa0a3ac6a86faf541d17109aa609865428053be65c85141cd3849526f\": container with ID starting with 01711d4aa0a3ac6a86faf541d17109aa609865428053be65c85141cd3849526f not found: ID does not exist" Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.667177 4766 scope.go:117] "RemoveContainer" containerID="5c91575c0c0fcdced52dfe6f7a2a8774b5a6a78f572724b4d5f719bcb220797c" Nov 26 01:14:33 crc kubenswrapper[4766]: E1126 01:14:33.667851 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c91575c0c0fcdced52dfe6f7a2a8774b5a6a78f572724b4d5f719bcb220797c\": container with ID starting with 5c91575c0c0fcdced52dfe6f7a2a8774b5a6a78f572724b4d5f719bcb220797c not found: ID does not exist" containerID="5c91575c0c0fcdced52dfe6f7a2a8774b5a6a78f572724b4d5f719bcb220797c" Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.667903 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c91575c0c0fcdced52dfe6f7a2a8774b5a6a78f572724b4d5f719bcb220797c"} err="failed to get container status \"5c91575c0c0fcdced52dfe6f7a2a8774b5a6a78f572724b4d5f719bcb220797c\": rpc error: code = NotFound desc = could not find container \"5c91575c0c0fcdced52dfe6f7a2a8774b5a6a78f572724b4d5f719bcb220797c\": container with ID starting with 5c91575c0c0fcdced52dfe6f7a2a8774b5a6a78f572724b4d5f719bcb220797c not found: ID does not exist" Nov 26 01:14:33 crc kubenswrapper[4766]: I1126 01:14:33.850552 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d39fa104-3076-45d1-92a1-3c0dcef1c7e3" path="/var/lib/kubelet/pods/d39fa104-3076-45d1-92a1-3c0dcef1c7e3/volumes" Nov 26 01:14:46 crc kubenswrapper[4766]: I1126 01:14:46.828193 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:14:46 crc kubenswrapper[4766]: E1126 01:14:46.829369 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:15:00 crc kubenswrapper[4766]: I1126 01:15:00.160994 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401995-pd7kg"] Nov 26 01:15:00 crc kubenswrapper[4766]: E1126 01:15:00.162167 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d39fa104-3076-45d1-92a1-3c0dcef1c7e3" containerName="extract-content" Nov 26 01:15:00 crc kubenswrapper[4766]: I1126 01:15:00.162194 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d39fa104-3076-45d1-92a1-3c0dcef1c7e3" containerName="extract-content" Nov 26 01:15:00 crc kubenswrapper[4766]: E1126 01:15:00.162270 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d39fa104-3076-45d1-92a1-3c0dcef1c7e3" containerName="extract-utilities" Nov 26 01:15:00 crc kubenswrapper[4766]: I1126 01:15:00.162285 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d39fa104-3076-45d1-92a1-3c0dcef1c7e3" containerName="extract-utilities" Nov 26 01:15:00 crc kubenswrapper[4766]: E1126 01:15:00.162310 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d39fa104-3076-45d1-92a1-3c0dcef1c7e3" containerName="registry-server" Nov 26 01:15:00 crc kubenswrapper[4766]: I1126 01:15:00.162321 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d39fa104-3076-45d1-92a1-3c0dcef1c7e3" containerName="registry-server" Nov 26 01:15:00 crc kubenswrapper[4766]: I1126 01:15:00.162679 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="d39fa104-3076-45d1-92a1-3c0dcef1c7e3" containerName="registry-server" Nov 26 01:15:00 crc kubenswrapper[4766]: I1126 01:15:00.163688 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401995-pd7kg" Nov 26 01:15:00 crc kubenswrapper[4766]: I1126 01:15:00.166489 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 26 01:15:00 crc kubenswrapper[4766]: I1126 01:15:00.166786 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 26 01:15:00 crc kubenswrapper[4766]: I1126 01:15:00.173254 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401995-pd7kg"] Nov 26 01:15:00 crc kubenswrapper[4766]: I1126 01:15:00.275731 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ec347d52-c090-468b-8055-260ca66491bb-config-volume\") pod \"collect-profiles-29401995-pd7kg\" (UID: \"ec347d52-c090-468b-8055-260ca66491bb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401995-pd7kg" Nov 26 01:15:00 crc kubenswrapper[4766]: I1126 01:15:00.275984 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d5xn\" (UniqueName: \"kubernetes.io/projected/ec347d52-c090-468b-8055-260ca66491bb-kube-api-access-6d5xn\") pod \"collect-profiles-29401995-pd7kg\" (UID: \"ec347d52-c090-468b-8055-260ca66491bb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401995-pd7kg" Nov 26 01:15:00 crc kubenswrapper[4766]: I1126 01:15:00.276321 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ec347d52-c090-468b-8055-260ca66491bb-secret-volume\") pod \"collect-profiles-29401995-pd7kg\" (UID: \"ec347d52-c090-468b-8055-260ca66491bb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401995-pd7kg" Nov 26 01:15:00 crc kubenswrapper[4766]: I1126 01:15:00.377941 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ec347d52-c090-468b-8055-260ca66491bb-secret-volume\") pod \"collect-profiles-29401995-pd7kg\" (UID: \"ec347d52-c090-468b-8055-260ca66491bb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401995-pd7kg" Nov 26 01:15:00 crc kubenswrapper[4766]: I1126 01:15:00.378057 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ec347d52-c090-468b-8055-260ca66491bb-config-volume\") pod \"collect-profiles-29401995-pd7kg\" (UID: \"ec347d52-c090-468b-8055-260ca66491bb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401995-pd7kg" Nov 26 01:15:00 crc kubenswrapper[4766]: I1126 01:15:00.378196 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d5xn\" (UniqueName: \"kubernetes.io/projected/ec347d52-c090-468b-8055-260ca66491bb-kube-api-access-6d5xn\") pod \"collect-profiles-29401995-pd7kg\" (UID: \"ec347d52-c090-468b-8055-260ca66491bb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401995-pd7kg" Nov 26 01:15:00 crc kubenswrapper[4766]: I1126 01:15:00.379013 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ec347d52-c090-468b-8055-260ca66491bb-config-volume\") pod \"collect-profiles-29401995-pd7kg\" (UID: \"ec347d52-c090-468b-8055-260ca66491bb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401995-pd7kg" Nov 26 01:15:00 crc kubenswrapper[4766]: I1126 01:15:00.387300 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ec347d52-c090-468b-8055-260ca66491bb-secret-volume\") pod \"collect-profiles-29401995-pd7kg\" (UID: \"ec347d52-c090-468b-8055-260ca66491bb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401995-pd7kg" Nov 26 01:15:00 crc kubenswrapper[4766]: I1126 01:15:00.399366 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d5xn\" (UniqueName: \"kubernetes.io/projected/ec347d52-c090-468b-8055-260ca66491bb-kube-api-access-6d5xn\") pod \"collect-profiles-29401995-pd7kg\" (UID: \"ec347d52-c090-468b-8055-260ca66491bb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401995-pd7kg" Nov 26 01:15:00 crc kubenswrapper[4766]: I1126 01:15:00.486066 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401995-pd7kg" Nov 26 01:15:00 crc kubenswrapper[4766]: I1126 01:15:00.826880 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:15:00 crc kubenswrapper[4766]: E1126 01:15:00.827167 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:15:00 crc kubenswrapper[4766]: I1126 01:15:00.939999 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401995-pd7kg"] Nov 26 01:15:01 crc kubenswrapper[4766]: I1126 01:15:01.916960 4766 generic.go:334] "Generic (PLEG): container finished" podID="ec347d52-c090-468b-8055-260ca66491bb" containerID="47aed24158cb58dd47f194727b92c55ec5f1bca86c10d54bba7f3bdbc3a04684" exitCode=0 Nov 26 01:15:01 crc kubenswrapper[4766]: I1126 01:15:01.917445 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401995-pd7kg" event={"ID":"ec347d52-c090-468b-8055-260ca66491bb","Type":"ContainerDied","Data":"47aed24158cb58dd47f194727b92c55ec5f1bca86c10d54bba7f3bdbc3a04684"} Nov 26 01:15:01 crc kubenswrapper[4766]: I1126 01:15:01.917506 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401995-pd7kg" event={"ID":"ec347d52-c090-468b-8055-260ca66491bb","Type":"ContainerStarted","Data":"2a88274a40770d26769e4ffca318bc1fc3126963c32139748fef4473071235ee"} Nov 26 01:15:03 crc kubenswrapper[4766]: I1126 01:15:03.419947 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401995-pd7kg" Nov 26 01:15:03 crc kubenswrapper[4766]: I1126 01:15:03.552048 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ec347d52-c090-468b-8055-260ca66491bb-secret-volume\") pod \"ec347d52-c090-468b-8055-260ca66491bb\" (UID: \"ec347d52-c090-468b-8055-260ca66491bb\") " Nov 26 01:15:03 crc kubenswrapper[4766]: I1126 01:15:03.552231 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ec347d52-c090-468b-8055-260ca66491bb-config-volume\") pod \"ec347d52-c090-468b-8055-260ca66491bb\" (UID: \"ec347d52-c090-468b-8055-260ca66491bb\") " Nov 26 01:15:03 crc kubenswrapper[4766]: I1126 01:15:03.552340 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6d5xn\" (UniqueName: \"kubernetes.io/projected/ec347d52-c090-468b-8055-260ca66491bb-kube-api-access-6d5xn\") pod \"ec347d52-c090-468b-8055-260ca66491bb\" (UID: \"ec347d52-c090-468b-8055-260ca66491bb\") " Nov 26 01:15:03 crc kubenswrapper[4766]: I1126 01:15:03.552891 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec347d52-c090-468b-8055-260ca66491bb-config-volume" (OuterVolumeSpecName: "config-volume") pod "ec347d52-c090-468b-8055-260ca66491bb" (UID: "ec347d52-c090-468b-8055-260ca66491bb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:15:03 crc kubenswrapper[4766]: I1126 01:15:03.557600 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec347d52-c090-468b-8055-260ca66491bb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ec347d52-c090-468b-8055-260ca66491bb" (UID: "ec347d52-c090-468b-8055-260ca66491bb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:15:03 crc kubenswrapper[4766]: I1126 01:15:03.558033 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec347d52-c090-468b-8055-260ca66491bb-kube-api-access-6d5xn" (OuterVolumeSpecName: "kube-api-access-6d5xn") pod "ec347d52-c090-468b-8055-260ca66491bb" (UID: "ec347d52-c090-468b-8055-260ca66491bb"). InnerVolumeSpecName "kube-api-access-6d5xn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:15:03 crc kubenswrapper[4766]: I1126 01:15:03.654564 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6d5xn\" (UniqueName: \"kubernetes.io/projected/ec347d52-c090-468b-8055-260ca66491bb-kube-api-access-6d5xn\") on node \"crc\" DevicePath \"\"" Nov 26 01:15:03 crc kubenswrapper[4766]: I1126 01:15:03.654781 4766 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ec347d52-c090-468b-8055-260ca66491bb-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 26 01:15:03 crc kubenswrapper[4766]: I1126 01:15:03.654861 4766 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ec347d52-c090-468b-8055-260ca66491bb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 01:15:03 crc kubenswrapper[4766]: I1126 01:15:03.949879 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401995-pd7kg" event={"ID":"ec347d52-c090-468b-8055-260ca66491bb","Type":"ContainerDied","Data":"2a88274a40770d26769e4ffca318bc1fc3126963c32139748fef4473071235ee"} Nov 26 01:15:03 crc kubenswrapper[4766]: I1126 01:15:03.949921 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a88274a40770d26769e4ffca318bc1fc3126963c32139748fef4473071235ee" Nov 26 01:15:03 crc kubenswrapper[4766]: I1126 01:15:03.950278 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401995-pd7kg" Nov 26 01:15:04 crc kubenswrapper[4766]: I1126 01:15:04.514721 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z"] Nov 26 01:15:04 crc kubenswrapper[4766]: I1126 01:15:04.526145 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401950-66m4z"] Nov 26 01:15:05 crc kubenswrapper[4766]: I1126 01:15:05.847993 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="216d430f-aabb-4376-a764-0b1f07ff228b" path="/var/lib/kubelet/pods/216d430f-aabb-4376-a764-0b1f07ff228b/volumes" Nov 26 01:15:13 crc kubenswrapper[4766]: I1126 01:15:13.826999 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:15:13 crc kubenswrapper[4766]: E1126 01:15:13.827913 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:15:26 crc kubenswrapper[4766]: I1126 01:15:26.827298 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:15:26 crc kubenswrapper[4766]: E1126 01:15:26.828559 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:15:37 crc kubenswrapper[4766]: I1126 01:15:37.828490 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:15:37 crc kubenswrapper[4766]: E1126 01:15:37.829791 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:15:49 crc kubenswrapper[4766]: I1126 01:15:49.877132 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:15:49 crc kubenswrapper[4766]: E1126 01:15:49.878222 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:16:00 crc kubenswrapper[4766]: I1126 01:16:00.060942 4766 scope.go:117] "RemoveContainer" containerID="d79314a1058accc10891514c9da7003c771b394ca5edac26e6fa5b0c8e409794" Nov 26 01:16:04 crc kubenswrapper[4766]: I1126 01:16:04.827337 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:16:04 crc kubenswrapper[4766]: E1126 01:16:04.828344 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:16:17 crc kubenswrapper[4766]: I1126 01:16:17.827776 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:16:17 crc kubenswrapper[4766]: E1126 01:16:17.828836 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:16:31 crc kubenswrapper[4766]: I1126 01:16:31.826864 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:16:31 crc kubenswrapper[4766]: E1126 01:16:31.827753 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:16:42 crc kubenswrapper[4766]: I1126 01:16:42.826787 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:16:42 crc kubenswrapper[4766]: E1126 01:16:42.827885 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:16:57 crc kubenswrapper[4766]: I1126 01:16:57.828727 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:16:57 crc kubenswrapper[4766]: E1126 01:16:57.830989 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:17:12 crc kubenswrapper[4766]: I1126 01:17:12.827617 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:17:12 crc kubenswrapper[4766]: E1126 01:17:12.828884 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:17:26 crc kubenswrapper[4766]: I1126 01:17:26.827146 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:17:26 crc kubenswrapper[4766]: E1126 01:17:26.828258 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:17:41 crc kubenswrapper[4766]: I1126 01:17:41.827358 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:17:43 crc kubenswrapper[4766]: I1126 01:17:43.070489 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"207a49c35f1945eec7efffff2aa31a2a80326afc58102d8b59af0cccdc0adb5e"} Nov 26 01:17:46 crc kubenswrapper[4766]: I1126 01:17:46.800130 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p"] Nov 26 01:17:46 crc kubenswrapper[4766]: I1126 01:17:46.813168 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z"] Nov 26 01:17:46 crc kubenswrapper[4766]: I1126 01:17:46.826410 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cmb77"] Nov 26 01:17:46 crc kubenswrapper[4766]: I1126 01:17:46.838968 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr"] Nov 26 01:17:46 crc kubenswrapper[4766]: I1126 01:17:46.848429 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq"] Nov 26 01:17:46 crc kubenswrapper[4766]: I1126 01:17:46.858238 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-dzc8z"] Nov 26 01:17:46 crc kubenswrapper[4766]: I1126 01:17:46.868430 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp"] Nov 26 01:17:46 crc kubenswrapper[4766]: I1126 01:17:46.880669 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr"] Nov 26 01:17:46 crc kubenswrapper[4766]: I1126 01:17:46.907815 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx"] Nov 26 01:17:46 crc kubenswrapper[4766]: I1126 01:17:46.923036 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-57p4p"] Nov 26 01:17:46 crc kubenswrapper[4766]: I1126 01:17:46.936353 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4t4jr"] Nov 26 01:17:46 crc kubenswrapper[4766]: I1126 01:17:46.946380 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cmb77"] Nov 26 01:17:46 crc kubenswrapper[4766]: I1126 01:17:46.958249 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x"] Nov 26 01:17:46 crc kubenswrapper[4766]: I1126 01:17:46.966634 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x"] Nov 26 01:17:46 crc kubenswrapper[4766]: I1126 01:17:46.975618 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8"] Nov 26 01:17:46 crc kubenswrapper[4766]: I1126 01:17:46.984670 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp"] Nov 26 01:17:46 crc kubenswrapper[4766]: I1126 01:17:46.998040 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm"] Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.008479 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-p5snp"] Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.017553 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wnqnx"] Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.025663 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb"] Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.033519 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt"] Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.043528 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-sp6fp"] Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.052636 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8"] Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.061592 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rfgjq"] Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.070499 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dsr5x"] Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.079507 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-99tg8"] Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.089339 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9vd9x"] Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.098907 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-wm8zr"] Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.107569 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gwmm"] Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.119308 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-6t2cb"] Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.129179 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jltdt"] Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.138677 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-scsf8"] Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.843274 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ba3a09a-d108-4c92-b3db-4f531ba53427" path="/var/lib/kubelet/pods/0ba3a09a-d108-4c92-b3db-4f531ba53427/volumes" Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.846569 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fb35a35-5b73-4790-bd2f-b90260847990" path="/var/lib/kubelet/pods/1fb35a35-5b73-4790-bd2f-b90260847990/volumes" Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.847706 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e890aa1-53db-4e49-a3d2-f2ba570bfd7e" path="/var/lib/kubelet/pods/2e890aa1-53db-4e49-a3d2-f2ba570bfd7e/volumes" Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.848909 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e66227e-c68b-47cd-97e6-91a7e447ea81" path="/var/lib/kubelet/pods/3e66227e-c68b-47cd-97e6-91a7e447ea81/volumes" Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.852208 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5356fee3-3206-46eb-b466-4eb864ecb237" path="/var/lib/kubelet/pods/5356fee3-3206-46eb-b466-4eb864ecb237/volumes" Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.853380 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6acb6255-ec9f-4c75-ac41-cb7e1ec842a9" path="/var/lib/kubelet/pods/6acb6255-ec9f-4c75-ac41-cb7e1ec842a9/volumes" Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.854518 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e4f73e8-58b4-47c3-83c6-75cc10ae1866" path="/var/lib/kubelet/pods/7e4f73e8-58b4-47c3-83c6-75cc10ae1866/volumes" Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.856977 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fee6935-5d49-4559-93de-8b244b5a6261" path="/var/lib/kubelet/pods/7fee6935-5d49-4559-93de-8b244b5a6261/volumes" Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.858114 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2" path="/var/lib/kubelet/pods/9fa895f7-5ae7-4f30-9dfb-f610be9d6ab2/volumes" Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.859267 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac83f8e3-4d14-4f01-992d-ab5c7fcacaed" path="/var/lib/kubelet/pods/ac83f8e3-4d14-4f01-992d-ab5c7fcacaed/volumes" Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.860438 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afe63c23-a92e-4701-a422-d431ed80d42d" path="/var/lib/kubelet/pods/afe63c23-a92e-4701-a422-d431ed80d42d/volumes" Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.862396 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2a8f94b-256b-4c29-83de-24b15b0284f2" path="/var/lib/kubelet/pods/d2a8f94b-256b-4c29-83de-24b15b0284f2/volumes" Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.863596 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d456c126-aca3-47d2-b4fb-7ab506d116fb" path="/var/lib/kubelet/pods/d456c126-aca3-47d2-b4fb-7ab506d116fb/volumes" Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.864835 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5921e8a-b3ce-4840-8211-a96134855a2f" path="/var/lib/kubelet/pods/e5921e8a-b3ce-4840-8211-a96134855a2f/volumes" Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.866971 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecefedc9-a319-4ada-99e6-8e7bf2ad5da5" path="/var/lib/kubelet/pods/ecefedc9-a319-4ada-99e6-8e7bf2ad5da5/volumes" Nov 26 01:17:47 crc kubenswrapper[4766]: I1126 01:17:47.868273 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f264bace-32cb-4fb6-acab-40d59db68d6e" path="/var/lib/kubelet/pods/f264bace-32cb-4fb6-acab-40d59db68d6e/volumes" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.586829 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl"] Nov 26 01:17:51 crc kubenswrapper[4766]: E1126 01:17:51.587941 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec347d52-c090-468b-8055-260ca66491bb" containerName="collect-profiles" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.587957 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec347d52-c090-468b-8055-260ca66491bb" containerName="collect-profiles" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.588229 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec347d52-c090-468b-8055-260ca66491bb" containerName="collect-profiles" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.589047 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.591351 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.591926 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.592627 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.593581 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.593896 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.596299 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl"] Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.648046 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl\" (UID: \"7ce6b646-f313-478b-b425-bc3b56504c3f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.648210 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlj62\" (UniqueName: \"kubernetes.io/projected/7ce6b646-f313-478b-b425-bc3b56504c3f-kube-api-access-hlj62\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl\" (UID: \"7ce6b646-f313-478b-b425-bc3b56504c3f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.648283 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl\" (UID: \"7ce6b646-f313-478b-b425-bc3b56504c3f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.648324 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl\" (UID: \"7ce6b646-f313-478b-b425-bc3b56504c3f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.648344 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl\" (UID: \"7ce6b646-f313-478b-b425-bc3b56504c3f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.749825 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlj62\" (UniqueName: \"kubernetes.io/projected/7ce6b646-f313-478b-b425-bc3b56504c3f-kube-api-access-hlj62\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl\" (UID: \"7ce6b646-f313-478b-b425-bc3b56504c3f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.749926 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl\" (UID: \"7ce6b646-f313-478b-b425-bc3b56504c3f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.749978 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl\" (UID: \"7ce6b646-f313-478b-b425-bc3b56504c3f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.750006 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl\" (UID: \"7ce6b646-f313-478b-b425-bc3b56504c3f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.750038 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl\" (UID: \"7ce6b646-f313-478b-b425-bc3b56504c3f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.756055 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl\" (UID: \"7ce6b646-f313-478b-b425-bc3b56504c3f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.756175 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl\" (UID: \"7ce6b646-f313-478b-b425-bc3b56504c3f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.756941 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl\" (UID: \"7ce6b646-f313-478b-b425-bc3b56504c3f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.764303 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl\" (UID: \"7ce6b646-f313-478b-b425-bc3b56504c3f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.767627 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlj62\" (UniqueName: \"kubernetes.io/projected/7ce6b646-f313-478b-b425-bc3b56504c3f-kube-api-access-hlj62\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl\" (UID: \"7ce6b646-f313-478b-b425-bc3b56504c3f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" Nov 26 01:17:51 crc kubenswrapper[4766]: I1126 01:17:51.924141 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" Nov 26 01:17:52 crc kubenswrapper[4766]: I1126 01:17:52.513838 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl"] Nov 26 01:17:52 crc kubenswrapper[4766]: I1126 01:17:52.519841 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 01:17:53 crc kubenswrapper[4766]: I1126 01:17:53.181153 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" event={"ID":"7ce6b646-f313-478b-b425-bc3b56504c3f","Type":"ContainerStarted","Data":"e0652c690f395d4f4ceba4a8ec24f75c62605e1ff0e1a54a61da091438eb0a0e"} Nov 26 01:17:54 crc kubenswrapper[4766]: I1126 01:17:54.198209 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" event={"ID":"7ce6b646-f313-478b-b425-bc3b56504c3f","Type":"ContainerStarted","Data":"863e94ee35783b575e0510242f1aa1fadc03f49bb6bb2f84c5531fde82d56276"} Nov 26 01:17:54 crc kubenswrapper[4766]: I1126 01:17:54.219739 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" podStartSLOduration=2.710243536 podStartE2EDuration="3.219723676s" podCreationTimestamp="2025-11-26 01:17:51 +0000 UTC" firstStartedPulling="2025-11-26 01:17:52.519634127 +0000 UTC m=+3253.368404557" lastFinishedPulling="2025-11-26 01:17:53.029114267 +0000 UTC m=+3253.877884697" observedRunningTime="2025-11-26 01:17:54.219422639 +0000 UTC m=+3255.068193099" watchObservedRunningTime="2025-11-26 01:17:54.219723676 +0000 UTC m=+3255.068494106" Nov 26 01:18:00 crc kubenswrapper[4766]: I1126 01:18:00.188736 4766 scope.go:117] "RemoveContainer" containerID="6d9d72adc4a3064d64bf5e7372ee64711950dee9b1160e08bf2dca091eb825d5" Nov 26 01:18:00 crc kubenswrapper[4766]: I1126 01:18:00.258142 4766 scope.go:117] "RemoveContainer" containerID="7262c3217748cb4f2c388179da8f8d6b47d5240c1fe26815cc76eb98a2f6dec3" Nov 26 01:18:00 crc kubenswrapper[4766]: I1126 01:18:00.355818 4766 scope.go:117] "RemoveContainer" containerID="6440a835d32a787feaf06d14ed60fa3af91eec1a1e2e84f51ae6e6c6caad0bcd" Nov 26 01:18:00 crc kubenswrapper[4766]: I1126 01:18:00.468540 4766 scope.go:117] "RemoveContainer" containerID="bcf2656245fd5f4a1b2616954f9180c62d508dead67f1c3a7ab62a38bfbc9aa0" Nov 26 01:18:00 crc kubenswrapper[4766]: I1126 01:18:00.525630 4766 scope.go:117] "RemoveContainer" containerID="1b4c4650f175eb35d1916e6b05810a6a9f1a03d94cfb4e6f5312a3a56f6c4809" Nov 26 01:18:00 crc kubenswrapper[4766]: I1126 01:18:00.576282 4766 scope.go:117] "RemoveContainer" containerID="45a3eee6d40c5c70f429b1e60ea753207c66d3e90e880a0b3db180839ab48d8b" Nov 26 01:18:00 crc kubenswrapper[4766]: I1126 01:18:00.648823 4766 scope.go:117] "RemoveContainer" containerID="22fff6cb2b6b099a0362c698eb539824d4f86ce71060d1a9f8b3e798f72badfa" Nov 26 01:18:00 crc kubenswrapper[4766]: I1126 01:18:00.704981 4766 scope.go:117] "RemoveContainer" containerID="c1e94194b85eda72901f1729372aa218cc8883221639975d9e52f2a1c7979027" Nov 26 01:18:00 crc kubenswrapper[4766]: I1126 01:18:00.752984 4766 scope.go:117] "RemoveContainer" containerID="c20e1ac5e10091502b723502fb56034373de52996074b49ba01137cad6b8a1df" Nov 26 01:18:00 crc kubenswrapper[4766]: I1126 01:18:00.796315 4766 scope.go:117] "RemoveContainer" containerID="22db7d2b98931ff64bbf2313a66c843a87160a766dbd40e16b52adf6f125f17c" Nov 26 01:18:00 crc kubenswrapper[4766]: I1126 01:18:00.843672 4766 scope.go:117] "RemoveContainer" containerID="c95a0eaf6920e52eacf4489f1067b99e68b0b7651c58fd717451749e094d814e" Nov 26 01:18:00 crc kubenswrapper[4766]: I1126 01:18:00.906910 4766 scope.go:117] "RemoveContainer" containerID="62544c5a4746cf33b2e88c463d3bb5e09fe39de843456d753c9aa80b23f092bb" Nov 26 01:18:00 crc kubenswrapper[4766]: I1126 01:18:00.973035 4766 scope.go:117] "RemoveContainer" containerID="c898f90f982326804573682ce3aa916a45b1ec550c2211d44c67b313df2c79e0" Nov 26 01:18:01 crc kubenswrapper[4766]: I1126 01:18:01.010220 4766 scope.go:117] "RemoveContainer" containerID="68c7016a858c4a7a21c6443a85bc5d35e9693d290303733a56a2c665e77fb55f" Nov 26 01:18:01 crc kubenswrapper[4766]: I1126 01:18:01.045375 4766 scope.go:117] "RemoveContainer" containerID="aba7c1527460d45400164d3ba4df993725b996e9df22aebb71e574413d815763" Nov 26 01:18:06 crc kubenswrapper[4766]: I1126 01:18:06.383181 4766 generic.go:334] "Generic (PLEG): container finished" podID="7ce6b646-f313-478b-b425-bc3b56504c3f" containerID="863e94ee35783b575e0510242f1aa1fadc03f49bb6bb2f84c5531fde82d56276" exitCode=0 Nov 26 01:18:06 crc kubenswrapper[4766]: I1126 01:18:06.383273 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" event={"ID":"7ce6b646-f313-478b-b425-bc3b56504c3f","Type":"ContainerDied","Data":"863e94ee35783b575e0510242f1aa1fadc03f49bb6bb2f84c5531fde82d56276"} Nov 26 01:18:07 crc kubenswrapper[4766]: I1126 01:18:07.970045 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" Nov 26 01:18:07 crc kubenswrapper[4766]: I1126 01:18:07.977210 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-repo-setup-combined-ca-bundle\") pod \"7ce6b646-f313-478b-b425-bc3b56504c3f\" (UID: \"7ce6b646-f313-478b-b425-bc3b56504c3f\") " Nov 26 01:18:07 crc kubenswrapper[4766]: I1126 01:18:07.977521 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-inventory\") pod \"7ce6b646-f313-478b-b425-bc3b56504c3f\" (UID: \"7ce6b646-f313-478b-b425-bc3b56504c3f\") " Nov 26 01:18:07 crc kubenswrapper[4766]: I1126 01:18:07.978913 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-ceph\") pod \"7ce6b646-f313-478b-b425-bc3b56504c3f\" (UID: \"7ce6b646-f313-478b-b425-bc3b56504c3f\") " Nov 26 01:18:07 crc kubenswrapper[4766]: I1126 01:18:07.979597 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-ssh-key\") pod \"7ce6b646-f313-478b-b425-bc3b56504c3f\" (UID: \"7ce6b646-f313-478b-b425-bc3b56504c3f\") " Nov 26 01:18:07 crc kubenswrapper[4766]: I1126 01:18:07.979778 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlj62\" (UniqueName: \"kubernetes.io/projected/7ce6b646-f313-478b-b425-bc3b56504c3f-kube-api-access-hlj62\") pod \"7ce6b646-f313-478b-b425-bc3b56504c3f\" (UID: \"7ce6b646-f313-478b-b425-bc3b56504c3f\") " Nov 26 01:18:07 crc kubenswrapper[4766]: I1126 01:18:07.985729 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "7ce6b646-f313-478b-b425-bc3b56504c3f" (UID: "7ce6b646-f313-478b-b425-bc3b56504c3f"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:18:07 crc kubenswrapper[4766]: I1126 01:18:07.986015 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ce6b646-f313-478b-b425-bc3b56504c3f-kube-api-access-hlj62" (OuterVolumeSpecName: "kube-api-access-hlj62") pod "7ce6b646-f313-478b-b425-bc3b56504c3f" (UID: "7ce6b646-f313-478b-b425-bc3b56504c3f"). InnerVolumeSpecName "kube-api-access-hlj62". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:18:07 crc kubenswrapper[4766]: I1126 01:18:07.988567 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-ceph" (OuterVolumeSpecName: "ceph") pod "7ce6b646-f313-478b-b425-bc3b56504c3f" (UID: "7ce6b646-f313-478b-b425-bc3b56504c3f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.037692 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7ce6b646-f313-478b-b425-bc3b56504c3f" (UID: "7ce6b646-f313-478b-b425-bc3b56504c3f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.040835 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-inventory" (OuterVolumeSpecName: "inventory") pod "7ce6b646-f313-478b-b425-bc3b56504c3f" (UID: "7ce6b646-f313-478b-b425-bc3b56504c3f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.083920 4766 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.083964 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.083975 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.083984 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ce6b646-f313-478b-b425-bc3b56504c3f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.083994 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlj62\" (UniqueName: \"kubernetes.io/projected/7ce6b646-f313-478b-b425-bc3b56504c3f-kube-api-access-hlj62\") on node \"crc\" DevicePath \"\"" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.421553 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" event={"ID":"7ce6b646-f313-478b-b425-bc3b56504c3f","Type":"ContainerDied","Data":"e0652c690f395d4f4ceba4a8ec24f75c62605e1ff0e1a54a61da091438eb0a0e"} Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.421611 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0652c690f395d4f4ceba4a8ec24f75c62605e1ff0e1a54a61da091438eb0a0e" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.421693 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.489721 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl"] Nov 26 01:18:08 crc kubenswrapper[4766]: E1126 01:18:08.490369 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce6b646-f313-478b-b425-bc3b56504c3f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.490405 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce6b646-f313-478b-b425-bc3b56504c3f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.490725 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ce6b646-f313-478b-b425-bc3b56504c3f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.491893 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.494871 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.497109 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.497373 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.497381 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.498421 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.519646 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl"] Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.595135 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl\" (UID: \"973efc5a-bdcf-43af-baf5-11ae375d3494\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.595231 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl\" (UID: \"973efc5a-bdcf-43af-baf5-11ae375d3494\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.595401 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl\" (UID: \"973efc5a-bdcf-43af-baf5-11ae375d3494\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.595751 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mpq8\" (UniqueName: \"kubernetes.io/projected/973efc5a-bdcf-43af-baf5-11ae375d3494-kube-api-access-7mpq8\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl\" (UID: \"973efc5a-bdcf-43af-baf5-11ae375d3494\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.595873 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl\" (UID: \"973efc5a-bdcf-43af-baf5-11ae375d3494\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.698264 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mpq8\" (UniqueName: \"kubernetes.io/projected/973efc5a-bdcf-43af-baf5-11ae375d3494-kube-api-access-7mpq8\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl\" (UID: \"973efc5a-bdcf-43af-baf5-11ae375d3494\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.698694 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl\" (UID: \"973efc5a-bdcf-43af-baf5-11ae375d3494\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.698760 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl\" (UID: \"973efc5a-bdcf-43af-baf5-11ae375d3494\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.698826 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl\" (UID: \"973efc5a-bdcf-43af-baf5-11ae375d3494\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.698874 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl\" (UID: \"973efc5a-bdcf-43af-baf5-11ae375d3494\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.704831 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl\" (UID: \"973efc5a-bdcf-43af-baf5-11ae375d3494\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.706042 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl\" (UID: \"973efc5a-bdcf-43af-baf5-11ae375d3494\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.706236 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl\" (UID: \"973efc5a-bdcf-43af-baf5-11ae375d3494\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.707431 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl\" (UID: \"973efc5a-bdcf-43af-baf5-11ae375d3494\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.718155 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mpq8\" (UniqueName: \"kubernetes.io/projected/973efc5a-bdcf-43af-baf5-11ae375d3494-kube-api-access-7mpq8\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl\" (UID: \"973efc5a-bdcf-43af-baf5-11ae375d3494\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" Nov 26 01:18:08 crc kubenswrapper[4766]: I1126 01:18:08.819097 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" Nov 26 01:18:09 crc kubenswrapper[4766]: I1126 01:18:09.475384 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl"] Nov 26 01:18:09 crc kubenswrapper[4766]: W1126 01:18:09.491031 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod973efc5a_bdcf_43af_baf5_11ae375d3494.slice/crio-ac67ae246501f459b56096c565a07f130f543d15719ec60b22d449530b14684f WatchSource:0}: Error finding container ac67ae246501f459b56096c565a07f130f543d15719ec60b22d449530b14684f: Status 404 returned error can't find the container with id ac67ae246501f459b56096c565a07f130f543d15719ec60b22d449530b14684f Nov 26 01:18:10 crc kubenswrapper[4766]: I1126 01:18:10.450918 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" event={"ID":"973efc5a-bdcf-43af-baf5-11ae375d3494","Type":"ContainerStarted","Data":"4885ad85ce0a156b2ce5a0c8076000f9387c44f4fcd0269ff3b2dc2ff5776dea"} Nov 26 01:18:10 crc kubenswrapper[4766]: I1126 01:18:10.451242 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" event={"ID":"973efc5a-bdcf-43af-baf5-11ae375d3494","Type":"ContainerStarted","Data":"ac67ae246501f459b56096c565a07f130f543d15719ec60b22d449530b14684f"} Nov 26 01:18:10 crc kubenswrapper[4766]: I1126 01:18:10.496495 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" podStartSLOduration=2.096232985 podStartE2EDuration="2.496470595s" podCreationTimestamp="2025-11-26 01:18:08 +0000 UTC" firstStartedPulling="2025-11-26 01:18:09.499736261 +0000 UTC m=+3270.348506701" lastFinishedPulling="2025-11-26 01:18:09.899973881 +0000 UTC m=+3270.748744311" observedRunningTime="2025-11-26 01:18:10.477421138 +0000 UTC m=+3271.326191578" watchObservedRunningTime="2025-11-26 01:18:10.496470595 +0000 UTC m=+3271.345241025" Nov 26 01:19:01 crc kubenswrapper[4766]: I1126 01:19:01.468169 4766 scope.go:117] "RemoveContainer" containerID="f8e1ff06e2703e15b1e93aff1dae5cd82d32771db4a0ca0aa591930ef4840c7f" Nov 26 01:20:06 crc kubenswrapper[4766]: I1126 01:20:06.064301 4766 generic.go:334] "Generic (PLEG): container finished" podID="973efc5a-bdcf-43af-baf5-11ae375d3494" containerID="4885ad85ce0a156b2ce5a0c8076000f9387c44f4fcd0269ff3b2dc2ff5776dea" exitCode=0 Nov 26 01:20:06 crc kubenswrapper[4766]: I1126 01:20:06.064418 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" event={"ID":"973efc5a-bdcf-43af-baf5-11ae375d3494","Type":"ContainerDied","Data":"4885ad85ce0a156b2ce5a0c8076000f9387c44f4fcd0269ff3b2dc2ff5776dea"} Nov 26 01:20:07 crc kubenswrapper[4766]: I1126 01:20:07.688825 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" Nov 26 01:20:07 crc kubenswrapper[4766]: I1126 01:20:07.813942 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-ssh-key\") pod \"973efc5a-bdcf-43af-baf5-11ae375d3494\" (UID: \"973efc5a-bdcf-43af-baf5-11ae375d3494\") " Nov 26 01:20:07 crc kubenswrapper[4766]: I1126 01:20:07.814203 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-bootstrap-combined-ca-bundle\") pod \"973efc5a-bdcf-43af-baf5-11ae375d3494\" (UID: \"973efc5a-bdcf-43af-baf5-11ae375d3494\") " Nov 26 01:20:07 crc kubenswrapper[4766]: I1126 01:20:07.814460 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-ceph\") pod \"973efc5a-bdcf-43af-baf5-11ae375d3494\" (UID: \"973efc5a-bdcf-43af-baf5-11ae375d3494\") " Nov 26 01:20:07 crc kubenswrapper[4766]: I1126 01:20:07.814512 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-inventory\") pod \"973efc5a-bdcf-43af-baf5-11ae375d3494\" (UID: \"973efc5a-bdcf-43af-baf5-11ae375d3494\") " Nov 26 01:20:07 crc kubenswrapper[4766]: I1126 01:20:07.814613 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mpq8\" (UniqueName: \"kubernetes.io/projected/973efc5a-bdcf-43af-baf5-11ae375d3494-kube-api-access-7mpq8\") pod \"973efc5a-bdcf-43af-baf5-11ae375d3494\" (UID: \"973efc5a-bdcf-43af-baf5-11ae375d3494\") " Nov 26 01:20:07 crc kubenswrapper[4766]: I1126 01:20:07.820106 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "973efc5a-bdcf-43af-baf5-11ae375d3494" (UID: "973efc5a-bdcf-43af-baf5-11ae375d3494"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:20:07 crc kubenswrapper[4766]: I1126 01:20:07.825404 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-ceph" (OuterVolumeSpecName: "ceph") pod "973efc5a-bdcf-43af-baf5-11ae375d3494" (UID: "973efc5a-bdcf-43af-baf5-11ae375d3494"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:20:07 crc kubenswrapper[4766]: I1126 01:20:07.827221 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/973efc5a-bdcf-43af-baf5-11ae375d3494-kube-api-access-7mpq8" (OuterVolumeSpecName: "kube-api-access-7mpq8") pod "973efc5a-bdcf-43af-baf5-11ae375d3494" (UID: "973efc5a-bdcf-43af-baf5-11ae375d3494"). InnerVolumeSpecName "kube-api-access-7mpq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:20:07 crc kubenswrapper[4766]: I1126 01:20:07.879891 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-inventory" (OuterVolumeSpecName: "inventory") pod "973efc5a-bdcf-43af-baf5-11ae375d3494" (UID: "973efc5a-bdcf-43af-baf5-11ae375d3494"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:20:07 crc kubenswrapper[4766]: I1126 01:20:07.888688 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "973efc5a-bdcf-43af-baf5-11ae375d3494" (UID: "973efc5a-bdcf-43af-baf5-11ae375d3494"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:20:07 crc kubenswrapper[4766]: I1126 01:20:07.918501 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:20:07 crc kubenswrapper[4766]: I1126 01:20:07.918545 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:20:07 crc kubenswrapper[4766]: I1126 01:20:07.918559 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mpq8\" (UniqueName: \"kubernetes.io/projected/973efc5a-bdcf-43af-baf5-11ae375d3494-kube-api-access-7mpq8\") on node \"crc\" DevicePath \"\"" Nov 26 01:20:07 crc kubenswrapper[4766]: I1126 01:20:07.918572 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:20:07 crc kubenswrapper[4766]: I1126 01:20:07.918586 4766 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/973efc5a-bdcf-43af-baf5-11ae375d3494-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.095396 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" event={"ID":"973efc5a-bdcf-43af-baf5-11ae375d3494","Type":"ContainerDied","Data":"ac67ae246501f459b56096c565a07f130f543d15719ec60b22d449530b14684f"} Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.095443 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac67ae246501f459b56096c565a07f130f543d15719ec60b22d449530b14684f" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.095466 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.224944 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds"] Nov 26 01:20:08 crc kubenswrapper[4766]: E1126 01:20:08.225731 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="973efc5a-bdcf-43af-baf5-11ae375d3494" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.225763 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="973efc5a-bdcf-43af-baf5-11ae375d3494" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.226176 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="973efc5a-bdcf-43af-baf5-11ae375d3494" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.227468 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.229610 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.234593 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.235037 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.235336 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.235556 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.265296 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds"] Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.327716 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7ae4f62-184e-4093-9a0a-12d6188ee301-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4bmds\" (UID: \"a7ae4f62-184e-4093-9a0a-12d6188ee301\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.328086 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5sp9\" (UniqueName: \"kubernetes.io/projected/a7ae4f62-184e-4093-9a0a-12d6188ee301-kube-api-access-r5sp9\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4bmds\" (UID: \"a7ae4f62-184e-4093-9a0a-12d6188ee301\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.328328 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a7ae4f62-184e-4093-9a0a-12d6188ee301-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4bmds\" (UID: \"a7ae4f62-184e-4093-9a0a-12d6188ee301\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.328489 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7ae4f62-184e-4093-9a0a-12d6188ee301-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4bmds\" (UID: \"a7ae4f62-184e-4093-9a0a-12d6188ee301\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.430136 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5sp9\" (UniqueName: \"kubernetes.io/projected/a7ae4f62-184e-4093-9a0a-12d6188ee301-kube-api-access-r5sp9\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4bmds\" (UID: \"a7ae4f62-184e-4093-9a0a-12d6188ee301\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.430234 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a7ae4f62-184e-4093-9a0a-12d6188ee301-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4bmds\" (UID: \"a7ae4f62-184e-4093-9a0a-12d6188ee301\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.430282 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7ae4f62-184e-4093-9a0a-12d6188ee301-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4bmds\" (UID: \"a7ae4f62-184e-4093-9a0a-12d6188ee301\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.430341 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7ae4f62-184e-4093-9a0a-12d6188ee301-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4bmds\" (UID: \"a7ae4f62-184e-4093-9a0a-12d6188ee301\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.434624 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7ae4f62-184e-4093-9a0a-12d6188ee301-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4bmds\" (UID: \"a7ae4f62-184e-4093-9a0a-12d6188ee301\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.435021 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7ae4f62-184e-4093-9a0a-12d6188ee301-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4bmds\" (UID: \"a7ae4f62-184e-4093-9a0a-12d6188ee301\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.435124 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a7ae4f62-184e-4093-9a0a-12d6188ee301-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4bmds\" (UID: \"a7ae4f62-184e-4093-9a0a-12d6188ee301\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.447418 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5sp9\" (UniqueName: \"kubernetes.io/projected/a7ae4f62-184e-4093-9a0a-12d6188ee301-kube-api-access-r5sp9\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4bmds\" (UID: \"a7ae4f62-184e-4093-9a0a-12d6188ee301\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds" Nov 26 01:20:08 crc kubenswrapper[4766]: I1126 01:20:08.559802 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds" Nov 26 01:20:09 crc kubenswrapper[4766]: I1126 01:20:09.175995 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds"] Nov 26 01:20:10 crc kubenswrapper[4766]: I1126 01:20:10.157175 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds" event={"ID":"a7ae4f62-184e-4093-9a0a-12d6188ee301","Type":"ContainerStarted","Data":"a1b20ee3e7ddf7509da4602b3189c8c7322a65df3dbba5833f92cb20092e0586"} Nov 26 01:20:10 crc kubenswrapper[4766]: I1126 01:20:10.158521 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds" event={"ID":"a7ae4f62-184e-4093-9a0a-12d6188ee301","Type":"ContainerStarted","Data":"017b27c2260da39ebdbbee28136e77535d715a70a8c4f5056f0e73df2c38d0d3"} Nov 26 01:20:10 crc kubenswrapper[4766]: I1126 01:20:10.178179 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds" podStartSLOduration=1.652757045 podStartE2EDuration="2.178141019s" podCreationTimestamp="2025-11-26 01:20:08 +0000 UTC" firstStartedPulling="2025-11-26 01:20:09.175006006 +0000 UTC m=+3390.023776436" lastFinishedPulling="2025-11-26 01:20:09.70038998 +0000 UTC m=+3390.549160410" observedRunningTime="2025-11-26 01:20:10.177628477 +0000 UTC m=+3391.026398917" watchObservedRunningTime="2025-11-26 01:20:10.178141019 +0000 UTC m=+3391.026911459" Nov 26 01:20:11 crc kubenswrapper[4766]: I1126 01:20:11.479309 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:20:11 crc kubenswrapper[4766]: I1126 01:20:11.479645 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:20:41 crc kubenswrapper[4766]: I1126 01:20:41.479522 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:20:41 crc kubenswrapper[4766]: I1126 01:20:41.480061 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:20:42 crc kubenswrapper[4766]: I1126 01:20:42.635366 4766 generic.go:334] "Generic (PLEG): container finished" podID="a7ae4f62-184e-4093-9a0a-12d6188ee301" containerID="a1b20ee3e7ddf7509da4602b3189c8c7322a65df3dbba5833f92cb20092e0586" exitCode=0 Nov 26 01:20:42 crc kubenswrapper[4766]: I1126 01:20:42.635494 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds" event={"ID":"a7ae4f62-184e-4093-9a0a-12d6188ee301","Type":"ContainerDied","Data":"a1b20ee3e7ddf7509da4602b3189c8c7322a65df3dbba5833f92cb20092e0586"} Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.198792 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.287222 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7ae4f62-184e-4093-9a0a-12d6188ee301-inventory\") pod \"a7ae4f62-184e-4093-9a0a-12d6188ee301\" (UID: \"a7ae4f62-184e-4093-9a0a-12d6188ee301\") " Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.287430 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7ae4f62-184e-4093-9a0a-12d6188ee301-ssh-key\") pod \"a7ae4f62-184e-4093-9a0a-12d6188ee301\" (UID: \"a7ae4f62-184e-4093-9a0a-12d6188ee301\") " Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.287510 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5sp9\" (UniqueName: \"kubernetes.io/projected/a7ae4f62-184e-4093-9a0a-12d6188ee301-kube-api-access-r5sp9\") pod \"a7ae4f62-184e-4093-9a0a-12d6188ee301\" (UID: \"a7ae4f62-184e-4093-9a0a-12d6188ee301\") " Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.287610 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a7ae4f62-184e-4093-9a0a-12d6188ee301-ceph\") pod \"a7ae4f62-184e-4093-9a0a-12d6188ee301\" (UID: \"a7ae4f62-184e-4093-9a0a-12d6188ee301\") " Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.294891 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7ae4f62-184e-4093-9a0a-12d6188ee301-kube-api-access-r5sp9" (OuterVolumeSpecName: "kube-api-access-r5sp9") pod "a7ae4f62-184e-4093-9a0a-12d6188ee301" (UID: "a7ae4f62-184e-4093-9a0a-12d6188ee301"). InnerVolumeSpecName "kube-api-access-r5sp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.295487 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7ae4f62-184e-4093-9a0a-12d6188ee301-ceph" (OuterVolumeSpecName: "ceph") pod "a7ae4f62-184e-4093-9a0a-12d6188ee301" (UID: "a7ae4f62-184e-4093-9a0a-12d6188ee301"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.321879 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7ae4f62-184e-4093-9a0a-12d6188ee301-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a7ae4f62-184e-4093-9a0a-12d6188ee301" (UID: "a7ae4f62-184e-4093-9a0a-12d6188ee301"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.344530 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7ae4f62-184e-4093-9a0a-12d6188ee301-inventory" (OuterVolumeSpecName: "inventory") pod "a7ae4f62-184e-4093-9a0a-12d6188ee301" (UID: "a7ae4f62-184e-4093-9a0a-12d6188ee301"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.390243 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7ae4f62-184e-4093-9a0a-12d6188ee301-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.390277 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5sp9\" (UniqueName: \"kubernetes.io/projected/a7ae4f62-184e-4093-9a0a-12d6188ee301-kube-api-access-r5sp9\") on node \"crc\" DevicePath \"\"" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.390290 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a7ae4f62-184e-4093-9a0a-12d6188ee301-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.390302 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7ae4f62-184e-4093-9a0a-12d6188ee301-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.484132 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fl8ds"] Nov 26 01:20:44 crc kubenswrapper[4766]: E1126 01:20:44.484731 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7ae4f62-184e-4093-9a0a-12d6188ee301" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.484760 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7ae4f62-184e-4093-9a0a-12d6188ee301" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.485126 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7ae4f62-184e-4093-9a0a-12d6188ee301" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.487201 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fl8ds" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.497008 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fl8ds"] Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.593548 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n47km\" (UniqueName: \"kubernetes.io/projected/5fe23d35-9a77-44f3-a362-7539768d3319-kube-api-access-n47km\") pod \"community-operators-fl8ds\" (UID: \"5fe23d35-9a77-44f3-a362-7539768d3319\") " pod="openshift-marketplace/community-operators-fl8ds" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.593680 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe23d35-9a77-44f3-a362-7539768d3319-utilities\") pod \"community-operators-fl8ds\" (UID: \"5fe23d35-9a77-44f3-a362-7539768d3319\") " pod="openshift-marketplace/community-operators-fl8ds" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.593891 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe23d35-9a77-44f3-a362-7539768d3319-catalog-content\") pod \"community-operators-fl8ds\" (UID: \"5fe23d35-9a77-44f3-a362-7539768d3319\") " pod="openshift-marketplace/community-operators-fl8ds" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.657207 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds" event={"ID":"a7ae4f62-184e-4093-9a0a-12d6188ee301","Type":"ContainerDied","Data":"017b27c2260da39ebdbbee28136e77535d715a70a8c4f5056f0e73df2c38d0d3"} Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.657244 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="017b27c2260da39ebdbbee28136e77535d715a70a8c4f5056f0e73df2c38d0d3" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.657323 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4bmds" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.695752 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe23d35-9a77-44f3-a362-7539768d3319-utilities\") pod \"community-operators-fl8ds\" (UID: \"5fe23d35-9a77-44f3-a362-7539768d3319\") " pod="openshift-marketplace/community-operators-fl8ds" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.695818 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe23d35-9a77-44f3-a362-7539768d3319-catalog-content\") pod \"community-operators-fl8ds\" (UID: \"5fe23d35-9a77-44f3-a362-7539768d3319\") " pod="openshift-marketplace/community-operators-fl8ds" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.695974 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n47km\" (UniqueName: \"kubernetes.io/projected/5fe23d35-9a77-44f3-a362-7539768d3319-kube-api-access-n47km\") pod \"community-operators-fl8ds\" (UID: \"5fe23d35-9a77-44f3-a362-7539768d3319\") " pod="openshift-marketplace/community-operators-fl8ds" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.696226 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe23d35-9a77-44f3-a362-7539768d3319-utilities\") pod \"community-operators-fl8ds\" (UID: \"5fe23d35-9a77-44f3-a362-7539768d3319\") " pod="openshift-marketplace/community-operators-fl8ds" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.696457 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe23d35-9a77-44f3-a362-7539768d3319-catalog-content\") pod \"community-operators-fl8ds\" (UID: \"5fe23d35-9a77-44f3-a362-7539768d3319\") " pod="openshift-marketplace/community-operators-fl8ds" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.719738 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n47km\" (UniqueName: \"kubernetes.io/projected/5fe23d35-9a77-44f3-a362-7539768d3319-kube-api-access-n47km\") pod \"community-operators-fl8ds\" (UID: \"5fe23d35-9a77-44f3-a362-7539768d3319\") " pod="openshift-marketplace/community-operators-fl8ds" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.768472 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs"] Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.770339 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.773042 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.775315 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.775678 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.775865 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.775920 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.789557 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs"] Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.798236 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttzcg\" (UniqueName: \"kubernetes.io/projected/d2efe136-b7e0-475d-9204-f0e36e720e96-kube-api-access-ttzcg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs\" (UID: \"d2efe136-b7e0-475d-9204-f0e36e720e96\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.798492 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d2efe136-b7e0-475d-9204-f0e36e720e96-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs\" (UID: \"d2efe136-b7e0-475d-9204-f0e36e720e96\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.798560 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2efe136-b7e0-475d-9204-f0e36e720e96-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs\" (UID: \"d2efe136-b7e0-475d-9204-f0e36e720e96\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.798584 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2efe136-b7e0-475d-9204-f0e36e720e96-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs\" (UID: \"d2efe136-b7e0-475d-9204-f0e36e720e96\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.829110 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fl8ds" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.900841 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d2efe136-b7e0-475d-9204-f0e36e720e96-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs\" (UID: \"d2efe136-b7e0-475d-9204-f0e36e720e96\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.900897 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2efe136-b7e0-475d-9204-f0e36e720e96-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs\" (UID: \"d2efe136-b7e0-475d-9204-f0e36e720e96\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.900915 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2efe136-b7e0-475d-9204-f0e36e720e96-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs\" (UID: \"d2efe136-b7e0-475d-9204-f0e36e720e96\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.900974 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttzcg\" (UniqueName: \"kubernetes.io/projected/d2efe136-b7e0-475d-9204-f0e36e720e96-kube-api-access-ttzcg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs\" (UID: \"d2efe136-b7e0-475d-9204-f0e36e720e96\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.914347 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d2efe136-b7e0-475d-9204-f0e36e720e96-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs\" (UID: \"d2efe136-b7e0-475d-9204-f0e36e720e96\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.918593 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2efe136-b7e0-475d-9204-f0e36e720e96-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs\" (UID: \"d2efe136-b7e0-475d-9204-f0e36e720e96\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.927212 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttzcg\" (UniqueName: \"kubernetes.io/projected/d2efe136-b7e0-475d-9204-f0e36e720e96-kube-api-access-ttzcg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs\" (UID: \"d2efe136-b7e0-475d-9204-f0e36e720e96\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs" Nov 26 01:20:44 crc kubenswrapper[4766]: I1126 01:20:44.929444 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2efe136-b7e0-475d-9204-f0e36e720e96-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs\" (UID: \"d2efe136-b7e0-475d-9204-f0e36e720e96\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs" Nov 26 01:20:45 crc kubenswrapper[4766]: I1126 01:20:45.100247 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs" Nov 26 01:20:45 crc kubenswrapper[4766]: I1126 01:20:45.411398 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fl8ds"] Nov 26 01:20:45 crc kubenswrapper[4766]: I1126 01:20:45.675977 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fl8ds" event={"ID":"5fe23d35-9a77-44f3-a362-7539768d3319","Type":"ContainerStarted","Data":"a3c7de18dbf66321d3f4780ea895de259080d29a003b918ed0536eddd79d4c7d"} Nov 26 01:20:45 crc kubenswrapper[4766]: I1126 01:20:45.676290 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fl8ds" event={"ID":"5fe23d35-9a77-44f3-a362-7539768d3319","Type":"ContainerStarted","Data":"c9c50109616061da0bfb9466e369ba6d7f42668267030e723bbf726b93d57168"} Nov 26 01:20:45 crc kubenswrapper[4766]: W1126 01:20:45.725604 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2efe136_b7e0_475d_9204_f0e36e720e96.slice/crio-5d41228f1e28439258417aa94df408b8befcf6cf141e0900ad690dc05d65f624 WatchSource:0}: Error finding container 5d41228f1e28439258417aa94df408b8befcf6cf141e0900ad690dc05d65f624: Status 404 returned error can't find the container with id 5d41228f1e28439258417aa94df408b8befcf6cf141e0900ad690dc05d65f624 Nov 26 01:20:45 crc kubenswrapper[4766]: I1126 01:20:45.733332 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs"] Nov 26 01:20:46 crc kubenswrapper[4766]: I1126 01:20:46.687455 4766 generic.go:334] "Generic (PLEG): container finished" podID="5fe23d35-9a77-44f3-a362-7539768d3319" containerID="a3c7de18dbf66321d3f4780ea895de259080d29a003b918ed0536eddd79d4c7d" exitCode=0 Nov 26 01:20:46 crc kubenswrapper[4766]: I1126 01:20:46.687511 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fl8ds" event={"ID":"5fe23d35-9a77-44f3-a362-7539768d3319","Type":"ContainerDied","Data":"a3c7de18dbf66321d3f4780ea895de259080d29a003b918ed0536eddd79d4c7d"} Nov 26 01:20:46 crc kubenswrapper[4766]: I1126 01:20:46.687969 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fl8ds" event={"ID":"5fe23d35-9a77-44f3-a362-7539768d3319","Type":"ContainerStarted","Data":"9ed6059ba0dcaedd4e9d80bd3b8e6e9cf65233b1c8c495a6706b48fa74e462aa"} Nov 26 01:20:46 crc kubenswrapper[4766]: I1126 01:20:46.691082 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs" event={"ID":"d2efe136-b7e0-475d-9204-f0e36e720e96","Type":"ContainerStarted","Data":"39663c088298dc21ccfe305c12aab5ec3d18fd743727f77392b31cc498f7a4cd"} Nov 26 01:20:46 crc kubenswrapper[4766]: I1126 01:20:46.691234 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs" event={"ID":"d2efe136-b7e0-475d-9204-f0e36e720e96","Type":"ContainerStarted","Data":"5d41228f1e28439258417aa94df408b8befcf6cf141e0900ad690dc05d65f624"} Nov 26 01:20:46 crc kubenswrapper[4766]: I1126 01:20:46.747695 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs" podStartSLOduration=2.3314062939999998 podStartE2EDuration="2.747674511s" podCreationTimestamp="2025-11-26 01:20:44 +0000 UTC" firstStartedPulling="2025-11-26 01:20:45.728051987 +0000 UTC m=+3426.576822417" lastFinishedPulling="2025-11-26 01:20:46.144320194 +0000 UTC m=+3426.993090634" observedRunningTime="2025-11-26 01:20:46.725825588 +0000 UTC m=+3427.574596028" watchObservedRunningTime="2025-11-26 01:20:46.747674511 +0000 UTC m=+3427.596444961" Nov 26 01:20:48 crc kubenswrapper[4766]: I1126 01:20:48.714353 4766 generic.go:334] "Generic (PLEG): container finished" podID="5fe23d35-9a77-44f3-a362-7539768d3319" containerID="9ed6059ba0dcaedd4e9d80bd3b8e6e9cf65233b1c8c495a6706b48fa74e462aa" exitCode=0 Nov 26 01:20:48 crc kubenswrapper[4766]: I1126 01:20:48.714480 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fl8ds" event={"ID":"5fe23d35-9a77-44f3-a362-7539768d3319","Type":"ContainerDied","Data":"9ed6059ba0dcaedd4e9d80bd3b8e6e9cf65233b1c8c495a6706b48fa74e462aa"} Nov 26 01:20:49 crc kubenswrapper[4766]: I1126 01:20:49.727059 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fl8ds" event={"ID":"5fe23d35-9a77-44f3-a362-7539768d3319","Type":"ContainerStarted","Data":"13d1d1f326b52f016a9291bab2c91d9716e5b74617403898d8cc2437f7619146"} Nov 26 01:20:49 crc kubenswrapper[4766]: I1126 01:20:49.755727 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fl8ds" podStartSLOduration=2.257317934 podStartE2EDuration="5.755704845s" podCreationTimestamp="2025-11-26 01:20:44 +0000 UTC" firstStartedPulling="2025-11-26 01:20:45.677865791 +0000 UTC m=+3426.526636221" lastFinishedPulling="2025-11-26 01:20:49.176252682 +0000 UTC m=+3430.025023132" observedRunningTime="2025-11-26 01:20:49.751893272 +0000 UTC m=+3430.600663732" watchObservedRunningTime="2025-11-26 01:20:49.755704845 +0000 UTC m=+3430.604475285" Nov 26 01:20:53 crc kubenswrapper[4766]: I1126 01:20:53.769987 4766 generic.go:334] "Generic (PLEG): container finished" podID="d2efe136-b7e0-475d-9204-f0e36e720e96" containerID="39663c088298dc21ccfe305c12aab5ec3d18fd743727f77392b31cc498f7a4cd" exitCode=0 Nov 26 01:20:53 crc kubenswrapper[4766]: I1126 01:20:53.770575 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs" event={"ID":"d2efe136-b7e0-475d-9204-f0e36e720e96","Type":"ContainerDied","Data":"39663c088298dc21ccfe305c12aab5ec3d18fd743727f77392b31cc498f7a4cd"} Nov 26 01:20:54 crc kubenswrapper[4766]: I1126 01:20:54.830613 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fl8ds" Nov 26 01:20:54 crc kubenswrapper[4766]: I1126 01:20:54.830699 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fl8ds" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.387304 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.550160 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2efe136-b7e0-475d-9204-f0e36e720e96-inventory\") pod \"d2efe136-b7e0-475d-9204-f0e36e720e96\" (UID: \"d2efe136-b7e0-475d-9204-f0e36e720e96\") " Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.550549 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d2efe136-b7e0-475d-9204-f0e36e720e96-ceph\") pod \"d2efe136-b7e0-475d-9204-f0e36e720e96\" (UID: \"d2efe136-b7e0-475d-9204-f0e36e720e96\") " Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.550690 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2efe136-b7e0-475d-9204-f0e36e720e96-ssh-key\") pod \"d2efe136-b7e0-475d-9204-f0e36e720e96\" (UID: \"d2efe136-b7e0-475d-9204-f0e36e720e96\") " Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.550743 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttzcg\" (UniqueName: \"kubernetes.io/projected/d2efe136-b7e0-475d-9204-f0e36e720e96-kube-api-access-ttzcg\") pod \"d2efe136-b7e0-475d-9204-f0e36e720e96\" (UID: \"d2efe136-b7e0-475d-9204-f0e36e720e96\") " Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.557861 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2efe136-b7e0-475d-9204-f0e36e720e96-ceph" (OuterVolumeSpecName: "ceph") pod "d2efe136-b7e0-475d-9204-f0e36e720e96" (UID: "d2efe136-b7e0-475d-9204-f0e36e720e96"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.557886 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2efe136-b7e0-475d-9204-f0e36e720e96-kube-api-access-ttzcg" (OuterVolumeSpecName: "kube-api-access-ttzcg") pod "d2efe136-b7e0-475d-9204-f0e36e720e96" (UID: "d2efe136-b7e0-475d-9204-f0e36e720e96"). InnerVolumeSpecName "kube-api-access-ttzcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.587800 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2efe136-b7e0-475d-9204-f0e36e720e96-inventory" (OuterVolumeSpecName: "inventory") pod "d2efe136-b7e0-475d-9204-f0e36e720e96" (UID: "d2efe136-b7e0-475d-9204-f0e36e720e96"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.587818 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2efe136-b7e0-475d-9204-f0e36e720e96-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d2efe136-b7e0-475d-9204-f0e36e720e96" (UID: "d2efe136-b7e0-475d-9204-f0e36e720e96"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.653083 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d2efe136-b7e0-475d-9204-f0e36e720e96-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.653116 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2efe136-b7e0-475d-9204-f0e36e720e96-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.653126 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttzcg\" (UniqueName: \"kubernetes.io/projected/d2efe136-b7e0-475d-9204-f0e36e720e96-kube-api-access-ttzcg\") on node \"crc\" DevicePath \"\"" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.653137 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2efe136-b7e0-475d-9204-f0e36e720e96-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.803283 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.803281 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs" event={"ID":"d2efe136-b7e0-475d-9204-f0e36e720e96","Type":"ContainerDied","Data":"5d41228f1e28439258417aa94df408b8befcf6cf141e0900ad690dc05d65f624"} Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.803369 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d41228f1e28439258417aa94df408b8befcf6cf141e0900ad690dc05d65f624" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.880230 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg"] Nov 26 01:20:55 crc kubenswrapper[4766]: E1126 01:20:55.880668 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2efe136-b7e0-475d-9204-f0e36e720e96" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.880679 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2efe136-b7e0-475d-9204-f0e36e720e96" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.880860 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2efe136-b7e0-475d-9204-f0e36e720e96" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.881523 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.884777 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.884931 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.885054 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.885234 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.885330 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.894225 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg"] Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.911539 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-fl8ds" podUID="5fe23d35-9a77-44f3-a362-7539768d3319" containerName="registry-server" probeResult="failure" output=< Nov 26 01:20:55 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 01:20:55 crc kubenswrapper[4766]: > Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.961009 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d516c45-cb2e-4920-8e14-3a47063212a1-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hvdtg\" (UID: \"5d516c45-cb2e-4920-8e14-3a47063212a1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.961158 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d516c45-cb2e-4920-8e14-3a47063212a1-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hvdtg\" (UID: \"5d516c45-cb2e-4920-8e14-3a47063212a1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.961182 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t6sm\" (UniqueName: \"kubernetes.io/projected/5d516c45-cb2e-4920-8e14-3a47063212a1-kube-api-access-9t6sm\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hvdtg\" (UID: \"5d516c45-cb2e-4920-8e14-3a47063212a1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg" Nov 26 01:20:55 crc kubenswrapper[4766]: I1126 01:20:55.961266 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d516c45-cb2e-4920-8e14-3a47063212a1-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hvdtg\" (UID: \"5d516c45-cb2e-4920-8e14-3a47063212a1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg" Nov 26 01:20:56 crc kubenswrapper[4766]: I1126 01:20:56.062765 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d516c45-cb2e-4920-8e14-3a47063212a1-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hvdtg\" (UID: \"5d516c45-cb2e-4920-8e14-3a47063212a1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg" Nov 26 01:20:56 crc kubenswrapper[4766]: I1126 01:20:56.062807 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t6sm\" (UniqueName: \"kubernetes.io/projected/5d516c45-cb2e-4920-8e14-3a47063212a1-kube-api-access-9t6sm\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hvdtg\" (UID: \"5d516c45-cb2e-4920-8e14-3a47063212a1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg" Nov 26 01:20:56 crc kubenswrapper[4766]: I1126 01:20:56.062872 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d516c45-cb2e-4920-8e14-3a47063212a1-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hvdtg\" (UID: \"5d516c45-cb2e-4920-8e14-3a47063212a1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg" Nov 26 01:20:56 crc kubenswrapper[4766]: I1126 01:20:56.062918 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d516c45-cb2e-4920-8e14-3a47063212a1-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hvdtg\" (UID: \"5d516c45-cb2e-4920-8e14-3a47063212a1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg" Nov 26 01:20:56 crc kubenswrapper[4766]: I1126 01:20:56.067130 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d516c45-cb2e-4920-8e14-3a47063212a1-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hvdtg\" (UID: \"5d516c45-cb2e-4920-8e14-3a47063212a1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg" Nov 26 01:20:56 crc kubenswrapper[4766]: I1126 01:20:56.070036 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d516c45-cb2e-4920-8e14-3a47063212a1-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hvdtg\" (UID: \"5d516c45-cb2e-4920-8e14-3a47063212a1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg" Nov 26 01:20:56 crc kubenswrapper[4766]: I1126 01:20:56.076477 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d516c45-cb2e-4920-8e14-3a47063212a1-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hvdtg\" (UID: \"5d516c45-cb2e-4920-8e14-3a47063212a1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg" Nov 26 01:20:56 crc kubenswrapper[4766]: I1126 01:20:56.082007 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t6sm\" (UniqueName: \"kubernetes.io/projected/5d516c45-cb2e-4920-8e14-3a47063212a1-kube-api-access-9t6sm\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hvdtg\" (UID: \"5d516c45-cb2e-4920-8e14-3a47063212a1\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg" Nov 26 01:20:56 crc kubenswrapper[4766]: I1126 01:20:56.203367 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg" Nov 26 01:20:56 crc kubenswrapper[4766]: I1126 01:20:56.813318 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg"] Nov 26 01:20:57 crc kubenswrapper[4766]: I1126 01:20:57.841404 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg" event={"ID":"5d516c45-cb2e-4920-8e14-3a47063212a1","Type":"ContainerStarted","Data":"32e77e1133a5a89c7aca2a93fad3b22532dc1354537ccff84d8ac22f4d067436"} Nov 26 01:20:57 crc kubenswrapper[4766]: I1126 01:20:57.842306 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg" event={"ID":"5d516c45-cb2e-4920-8e14-3a47063212a1","Type":"ContainerStarted","Data":"c70a9c38bf2f5de5bfb7301f0052ee3705f82ab901077ef4bb10a138e21133d1"} Nov 26 01:20:57 crc kubenswrapper[4766]: I1126 01:20:57.861619 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg" podStartSLOduration=2.411387672 podStartE2EDuration="2.861598539s" podCreationTimestamp="2025-11-26 01:20:55 +0000 UTC" firstStartedPulling="2025-11-26 01:20:56.823545063 +0000 UTC m=+3437.672315503" lastFinishedPulling="2025-11-26 01:20:57.27375593 +0000 UTC m=+3438.122526370" observedRunningTime="2025-11-26 01:20:57.851568754 +0000 UTC m=+3438.700339184" watchObservedRunningTime="2025-11-26 01:20:57.861598539 +0000 UTC m=+3438.710368979" Nov 26 01:21:04 crc kubenswrapper[4766]: I1126 01:21:04.924409 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fl8ds" Nov 26 01:21:04 crc kubenswrapper[4766]: I1126 01:21:04.992924 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fl8ds" Nov 26 01:21:05 crc kubenswrapper[4766]: I1126 01:21:05.172108 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fl8ds"] Nov 26 01:21:06 crc kubenswrapper[4766]: I1126 01:21:06.931640 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fl8ds" podUID="5fe23d35-9a77-44f3-a362-7539768d3319" containerName="registry-server" containerID="cri-o://13d1d1f326b52f016a9291bab2c91d9716e5b74617403898d8cc2437f7619146" gracePeriod=2 Nov 26 01:21:07 crc kubenswrapper[4766]: I1126 01:21:07.383201 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fl8ds" Nov 26 01:21:07 crc kubenswrapper[4766]: I1126 01:21:07.492174 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n47km\" (UniqueName: \"kubernetes.io/projected/5fe23d35-9a77-44f3-a362-7539768d3319-kube-api-access-n47km\") pod \"5fe23d35-9a77-44f3-a362-7539768d3319\" (UID: \"5fe23d35-9a77-44f3-a362-7539768d3319\") " Nov 26 01:21:07 crc kubenswrapper[4766]: I1126 01:21:07.492430 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe23d35-9a77-44f3-a362-7539768d3319-utilities\") pod \"5fe23d35-9a77-44f3-a362-7539768d3319\" (UID: \"5fe23d35-9a77-44f3-a362-7539768d3319\") " Nov 26 01:21:07 crc kubenswrapper[4766]: I1126 01:21:07.492511 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe23d35-9a77-44f3-a362-7539768d3319-catalog-content\") pod \"5fe23d35-9a77-44f3-a362-7539768d3319\" (UID: \"5fe23d35-9a77-44f3-a362-7539768d3319\") " Nov 26 01:21:07 crc kubenswrapper[4766]: I1126 01:21:07.493213 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fe23d35-9a77-44f3-a362-7539768d3319-utilities" (OuterVolumeSpecName: "utilities") pod "5fe23d35-9a77-44f3-a362-7539768d3319" (UID: "5fe23d35-9a77-44f3-a362-7539768d3319"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:21:07 crc kubenswrapper[4766]: I1126 01:21:07.518854 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe23d35-9a77-44f3-a362-7539768d3319-kube-api-access-n47km" (OuterVolumeSpecName: "kube-api-access-n47km") pod "5fe23d35-9a77-44f3-a362-7539768d3319" (UID: "5fe23d35-9a77-44f3-a362-7539768d3319"). InnerVolumeSpecName "kube-api-access-n47km". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:21:07 crc kubenswrapper[4766]: I1126 01:21:07.547103 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fe23d35-9a77-44f3-a362-7539768d3319-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5fe23d35-9a77-44f3-a362-7539768d3319" (UID: "5fe23d35-9a77-44f3-a362-7539768d3319"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:21:07 crc kubenswrapper[4766]: I1126 01:21:07.595162 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe23d35-9a77-44f3-a362-7539768d3319-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 01:21:07 crc kubenswrapper[4766]: I1126 01:21:07.595189 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe23d35-9a77-44f3-a362-7539768d3319-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 01:21:07 crc kubenswrapper[4766]: I1126 01:21:07.595200 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n47km\" (UniqueName: \"kubernetes.io/projected/5fe23d35-9a77-44f3-a362-7539768d3319-kube-api-access-n47km\") on node \"crc\" DevicePath \"\"" Nov 26 01:21:07 crc kubenswrapper[4766]: I1126 01:21:07.943584 4766 generic.go:334] "Generic (PLEG): container finished" podID="5fe23d35-9a77-44f3-a362-7539768d3319" containerID="13d1d1f326b52f016a9291bab2c91d9716e5b74617403898d8cc2437f7619146" exitCode=0 Nov 26 01:21:07 crc kubenswrapper[4766]: I1126 01:21:07.943636 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fl8ds" event={"ID":"5fe23d35-9a77-44f3-a362-7539768d3319","Type":"ContainerDied","Data":"13d1d1f326b52f016a9291bab2c91d9716e5b74617403898d8cc2437f7619146"} Nov 26 01:21:07 crc kubenswrapper[4766]: I1126 01:21:07.943681 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fl8ds" event={"ID":"5fe23d35-9a77-44f3-a362-7539768d3319","Type":"ContainerDied","Data":"c9c50109616061da0bfb9466e369ba6d7f42668267030e723bbf726b93d57168"} Nov 26 01:21:07 crc kubenswrapper[4766]: I1126 01:21:07.943703 4766 scope.go:117] "RemoveContainer" containerID="13d1d1f326b52f016a9291bab2c91d9716e5b74617403898d8cc2437f7619146" Nov 26 01:21:07 crc kubenswrapper[4766]: I1126 01:21:07.943845 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fl8ds" Nov 26 01:21:07 crc kubenswrapper[4766]: I1126 01:21:07.971810 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fl8ds"] Nov 26 01:21:07 crc kubenswrapper[4766]: I1126 01:21:07.972553 4766 scope.go:117] "RemoveContainer" containerID="9ed6059ba0dcaedd4e9d80bd3b8e6e9cf65233b1c8c495a6706b48fa74e462aa" Nov 26 01:21:08 crc kubenswrapper[4766]: I1126 01:21:08.004207 4766 scope.go:117] "RemoveContainer" containerID="a3c7de18dbf66321d3f4780ea895de259080d29a003b918ed0536eddd79d4c7d" Nov 26 01:21:08 crc kubenswrapper[4766]: I1126 01:21:08.005543 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fl8ds"] Nov 26 01:21:08 crc kubenswrapper[4766]: I1126 01:21:08.050288 4766 scope.go:117] "RemoveContainer" containerID="13d1d1f326b52f016a9291bab2c91d9716e5b74617403898d8cc2437f7619146" Nov 26 01:21:08 crc kubenswrapper[4766]: E1126 01:21:08.050815 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13d1d1f326b52f016a9291bab2c91d9716e5b74617403898d8cc2437f7619146\": container with ID starting with 13d1d1f326b52f016a9291bab2c91d9716e5b74617403898d8cc2437f7619146 not found: ID does not exist" containerID="13d1d1f326b52f016a9291bab2c91d9716e5b74617403898d8cc2437f7619146" Nov 26 01:21:08 crc kubenswrapper[4766]: I1126 01:21:08.050850 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13d1d1f326b52f016a9291bab2c91d9716e5b74617403898d8cc2437f7619146"} err="failed to get container status \"13d1d1f326b52f016a9291bab2c91d9716e5b74617403898d8cc2437f7619146\": rpc error: code = NotFound desc = could not find container \"13d1d1f326b52f016a9291bab2c91d9716e5b74617403898d8cc2437f7619146\": container with ID starting with 13d1d1f326b52f016a9291bab2c91d9716e5b74617403898d8cc2437f7619146 not found: ID does not exist" Nov 26 01:21:08 crc kubenswrapper[4766]: I1126 01:21:08.050873 4766 scope.go:117] "RemoveContainer" containerID="9ed6059ba0dcaedd4e9d80bd3b8e6e9cf65233b1c8c495a6706b48fa74e462aa" Nov 26 01:21:08 crc kubenswrapper[4766]: E1126 01:21:08.051152 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ed6059ba0dcaedd4e9d80bd3b8e6e9cf65233b1c8c495a6706b48fa74e462aa\": container with ID starting with 9ed6059ba0dcaedd4e9d80bd3b8e6e9cf65233b1c8c495a6706b48fa74e462aa not found: ID does not exist" containerID="9ed6059ba0dcaedd4e9d80bd3b8e6e9cf65233b1c8c495a6706b48fa74e462aa" Nov 26 01:21:08 crc kubenswrapper[4766]: I1126 01:21:08.051179 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ed6059ba0dcaedd4e9d80bd3b8e6e9cf65233b1c8c495a6706b48fa74e462aa"} err="failed to get container status \"9ed6059ba0dcaedd4e9d80bd3b8e6e9cf65233b1c8c495a6706b48fa74e462aa\": rpc error: code = NotFound desc = could not find container \"9ed6059ba0dcaedd4e9d80bd3b8e6e9cf65233b1c8c495a6706b48fa74e462aa\": container with ID starting with 9ed6059ba0dcaedd4e9d80bd3b8e6e9cf65233b1c8c495a6706b48fa74e462aa not found: ID does not exist" Nov 26 01:21:08 crc kubenswrapper[4766]: I1126 01:21:08.051196 4766 scope.go:117] "RemoveContainer" containerID="a3c7de18dbf66321d3f4780ea895de259080d29a003b918ed0536eddd79d4c7d" Nov 26 01:21:08 crc kubenswrapper[4766]: E1126 01:21:08.051496 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3c7de18dbf66321d3f4780ea895de259080d29a003b918ed0536eddd79d4c7d\": container with ID starting with a3c7de18dbf66321d3f4780ea895de259080d29a003b918ed0536eddd79d4c7d not found: ID does not exist" containerID="a3c7de18dbf66321d3f4780ea895de259080d29a003b918ed0536eddd79d4c7d" Nov 26 01:21:08 crc kubenswrapper[4766]: I1126 01:21:08.051534 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3c7de18dbf66321d3f4780ea895de259080d29a003b918ed0536eddd79d4c7d"} err="failed to get container status \"a3c7de18dbf66321d3f4780ea895de259080d29a003b918ed0536eddd79d4c7d\": rpc error: code = NotFound desc = could not find container \"a3c7de18dbf66321d3f4780ea895de259080d29a003b918ed0536eddd79d4c7d\": container with ID starting with a3c7de18dbf66321d3f4780ea895de259080d29a003b918ed0536eddd79d4c7d not found: ID does not exist" Nov 26 01:21:09 crc kubenswrapper[4766]: I1126 01:21:09.842851 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe23d35-9a77-44f3-a362-7539768d3319" path="/var/lib/kubelet/pods/5fe23d35-9a77-44f3-a362-7539768d3319/volumes" Nov 26 01:21:11 crc kubenswrapper[4766]: I1126 01:21:11.480244 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:21:11 crc kubenswrapper[4766]: I1126 01:21:11.480632 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:21:11 crc kubenswrapper[4766]: I1126 01:21:11.480727 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 01:21:11 crc kubenswrapper[4766]: I1126 01:21:11.481875 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"207a49c35f1945eec7efffff2aa31a2a80326afc58102d8b59af0cccdc0adb5e"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 01:21:11 crc kubenswrapper[4766]: I1126 01:21:11.481958 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://207a49c35f1945eec7efffff2aa31a2a80326afc58102d8b59af0cccdc0adb5e" gracePeriod=600 Nov 26 01:21:11 crc kubenswrapper[4766]: I1126 01:21:11.994371 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="207a49c35f1945eec7efffff2aa31a2a80326afc58102d8b59af0cccdc0adb5e" exitCode=0 Nov 26 01:21:11 crc kubenswrapper[4766]: I1126 01:21:11.994679 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"207a49c35f1945eec7efffff2aa31a2a80326afc58102d8b59af0cccdc0adb5e"} Nov 26 01:21:11 crc kubenswrapper[4766]: I1126 01:21:11.994713 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24"} Nov 26 01:21:11 crc kubenswrapper[4766]: I1126 01:21:11.994737 4766 scope.go:117] "RemoveContainer" containerID="bff0eb03643a0d67e585034ff4c5e3b953d57d5023f35cbe804d99a58f425e50" Nov 26 01:21:49 crc kubenswrapper[4766]: I1126 01:21:49.475725 4766 generic.go:334] "Generic (PLEG): container finished" podID="5d516c45-cb2e-4920-8e14-3a47063212a1" containerID="32e77e1133a5a89c7aca2a93fad3b22532dc1354537ccff84d8ac22f4d067436" exitCode=0 Nov 26 01:21:49 crc kubenswrapper[4766]: I1126 01:21:49.475860 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg" event={"ID":"5d516c45-cb2e-4920-8e14-3a47063212a1","Type":"ContainerDied","Data":"32e77e1133a5a89c7aca2a93fad3b22532dc1354537ccff84d8ac22f4d067436"} Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.021390 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.092957 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d516c45-cb2e-4920-8e14-3a47063212a1-ssh-key\") pod \"5d516c45-cb2e-4920-8e14-3a47063212a1\" (UID: \"5d516c45-cb2e-4920-8e14-3a47063212a1\") " Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.094031 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9t6sm\" (UniqueName: \"kubernetes.io/projected/5d516c45-cb2e-4920-8e14-3a47063212a1-kube-api-access-9t6sm\") pod \"5d516c45-cb2e-4920-8e14-3a47063212a1\" (UID: \"5d516c45-cb2e-4920-8e14-3a47063212a1\") " Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.094177 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d516c45-cb2e-4920-8e14-3a47063212a1-ceph\") pod \"5d516c45-cb2e-4920-8e14-3a47063212a1\" (UID: \"5d516c45-cb2e-4920-8e14-3a47063212a1\") " Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.094525 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d516c45-cb2e-4920-8e14-3a47063212a1-inventory\") pod \"5d516c45-cb2e-4920-8e14-3a47063212a1\" (UID: \"5d516c45-cb2e-4920-8e14-3a47063212a1\") " Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.098564 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d516c45-cb2e-4920-8e14-3a47063212a1-ceph" (OuterVolumeSpecName: "ceph") pod "5d516c45-cb2e-4920-8e14-3a47063212a1" (UID: "5d516c45-cb2e-4920-8e14-3a47063212a1"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.115369 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d516c45-cb2e-4920-8e14-3a47063212a1-kube-api-access-9t6sm" (OuterVolumeSpecName: "kube-api-access-9t6sm") pod "5d516c45-cb2e-4920-8e14-3a47063212a1" (UID: "5d516c45-cb2e-4920-8e14-3a47063212a1"). InnerVolumeSpecName "kube-api-access-9t6sm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.130398 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d516c45-cb2e-4920-8e14-3a47063212a1-inventory" (OuterVolumeSpecName: "inventory") pod "5d516c45-cb2e-4920-8e14-3a47063212a1" (UID: "5d516c45-cb2e-4920-8e14-3a47063212a1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.147836 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d516c45-cb2e-4920-8e14-3a47063212a1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5d516c45-cb2e-4920-8e14-3a47063212a1" (UID: "5d516c45-cb2e-4920-8e14-3a47063212a1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.197300 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d516c45-cb2e-4920-8e14-3a47063212a1-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.197328 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9t6sm\" (UniqueName: \"kubernetes.io/projected/5d516c45-cb2e-4920-8e14-3a47063212a1-kube-api-access-9t6sm\") on node \"crc\" DevicePath \"\"" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.197337 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5d516c45-cb2e-4920-8e14-3a47063212a1-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.197345 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d516c45-cb2e-4920-8e14-3a47063212a1-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.512909 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg" event={"ID":"5d516c45-cb2e-4920-8e14-3a47063212a1","Type":"ContainerDied","Data":"c70a9c38bf2f5de5bfb7301f0052ee3705f82ab901077ef4bb10a138e21133d1"} Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.513233 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c70a9c38bf2f5de5bfb7301f0052ee3705f82ab901077ef4bb10a138e21133d1" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.512971 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvdtg" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.609834 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5"] Nov 26 01:21:51 crc kubenswrapper[4766]: E1126 01:21:51.610251 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe23d35-9a77-44f3-a362-7539768d3319" containerName="extract-content" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.610274 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe23d35-9a77-44f3-a362-7539768d3319" containerName="extract-content" Nov 26 01:21:51 crc kubenswrapper[4766]: E1126 01:21:51.610287 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe23d35-9a77-44f3-a362-7539768d3319" containerName="extract-utilities" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.610293 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe23d35-9a77-44f3-a362-7539768d3319" containerName="extract-utilities" Nov 26 01:21:51 crc kubenswrapper[4766]: E1126 01:21:51.610328 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe23d35-9a77-44f3-a362-7539768d3319" containerName="registry-server" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.610334 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe23d35-9a77-44f3-a362-7539768d3319" containerName="registry-server" Nov 26 01:21:51 crc kubenswrapper[4766]: E1126 01:21:51.610342 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d516c45-cb2e-4920-8e14-3a47063212a1" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.610349 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d516c45-cb2e-4920-8e14-3a47063212a1" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.610526 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fe23d35-9a77-44f3-a362-7539768d3319" containerName="registry-server" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.610586 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d516c45-cb2e-4920-8e14-3a47063212a1" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.611337 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.614058 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.615828 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.615968 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.615989 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.617112 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.631134 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5"] Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.707059 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzsdx\" (UniqueName: \"kubernetes.io/projected/9f63fa32-bcb4-490f-947b-ec7ab639505c-kube-api-access-hzsdx\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5\" (UID: \"9f63fa32-bcb4-490f-947b-ec7ab639505c\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.707170 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9f63fa32-bcb4-490f-947b-ec7ab639505c-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5\" (UID: \"9f63fa32-bcb4-490f-947b-ec7ab639505c\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.707545 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f63fa32-bcb4-490f-947b-ec7ab639505c-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5\" (UID: \"9f63fa32-bcb4-490f-947b-ec7ab639505c\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.707803 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f63fa32-bcb4-490f-947b-ec7ab639505c-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5\" (UID: \"9f63fa32-bcb4-490f-947b-ec7ab639505c\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.809702 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f63fa32-bcb4-490f-947b-ec7ab639505c-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5\" (UID: \"9f63fa32-bcb4-490f-947b-ec7ab639505c\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.809828 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f63fa32-bcb4-490f-947b-ec7ab639505c-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5\" (UID: \"9f63fa32-bcb4-490f-947b-ec7ab639505c\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.809878 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzsdx\" (UniqueName: \"kubernetes.io/projected/9f63fa32-bcb4-490f-947b-ec7ab639505c-kube-api-access-hzsdx\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5\" (UID: \"9f63fa32-bcb4-490f-947b-ec7ab639505c\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.809941 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9f63fa32-bcb4-490f-947b-ec7ab639505c-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5\" (UID: \"9f63fa32-bcb4-490f-947b-ec7ab639505c\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.814305 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f63fa32-bcb4-490f-947b-ec7ab639505c-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5\" (UID: \"9f63fa32-bcb4-490f-947b-ec7ab639505c\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.815155 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f63fa32-bcb4-490f-947b-ec7ab639505c-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5\" (UID: \"9f63fa32-bcb4-490f-947b-ec7ab639505c\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.815609 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9f63fa32-bcb4-490f-947b-ec7ab639505c-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5\" (UID: \"9f63fa32-bcb4-490f-947b-ec7ab639505c\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.842843 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzsdx\" (UniqueName: \"kubernetes.io/projected/9f63fa32-bcb4-490f-947b-ec7ab639505c-kube-api-access-hzsdx\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5\" (UID: \"9f63fa32-bcb4-490f-947b-ec7ab639505c\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5" Nov 26 01:21:51 crc kubenswrapper[4766]: I1126 01:21:51.935749 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5" Nov 26 01:21:52 crc kubenswrapper[4766]: I1126 01:21:52.624449 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5"] Nov 26 01:21:53 crc kubenswrapper[4766]: I1126 01:21:53.546508 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5" event={"ID":"9f63fa32-bcb4-490f-947b-ec7ab639505c","Type":"ContainerStarted","Data":"d44b44c3ff1df4935a92840b283677c3f3c874acb32f4228d2daa794bb2e0202"} Nov 26 01:21:53 crc kubenswrapper[4766]: I1126 01:21:53.547866 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5" event={"ID":"9f63fa32-bcb4-490f-947b-ec7ab639505c","Type":"ContainerStarted","Data":"50be77714e94a4f026c37446f4eeaaafd69bd1be7dbe5a7d60cfa34072174ad5"} Nov 26 01:21:53 crc kubenswrapper[4766]: I1126 01:21:53.593352 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5" podStartSLOduration=2.130341268 podStartE2EDuration="2.593332207s" podCreationTimestamp="2025-11-26 01:21:51 +0000 UTC" firstStartedPulling="2025-11-26 01:21:52.630028127 +0000 UTC m=+3493.478798587" lastFinishedPulling="2025-11-26 01:21:53.093019086 +0000 UTC m=+3493.941789526" observedRunningTime="2025-11-26 01:21:53.574405135 +0000 UTC m=+3494.423175575" watchObservedRunningTime="2025-11-26 01:21:53.593332207 +0000 UTC m=+3494.442102647" Nov 26 01:21:58 crc kubenswrapper[4766]: I1126 01:21:58.610034 4766 generic.go:334] "Generic (PLEG): container finished" podID="9f63fa32-bcb4-490f-947b-ec7ab639505c" containerID="d44b44c3ff1df4935a92840b283677c3f3c874acb32f4228d2daa794bb2e0202" exitCode=0 Nov 26 01:21:58 crc kubenswrapper[4766]: I1126 01:21:58.610090 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5" event={"ID":"9f63fa32-bcb4-490f-947b-ec7ab639505c","Type":"ContainerDied","Data":"d44b44c3ff1df4935a92840b283677c3f3c874acb32f4228d2daa794bb2e0202"} Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.086707 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.228803 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f63fa32-bcb4-490f-947b-ec7ab639505c-ssh-key\") pod \"9f63fa32-bcb4-490f-947b-ec7ab639505c\" (UID: \"9f63fa32-bcb4-490f-947b-ec7ab639505c\") " Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.228876 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9f63fa32-bcb4-490f-947b-ec7ab639505c-ceph\") pod \"9f63fa32-bcb4-490f-947b-ec7ab639505c\" (UID: \"9f63fa32-bcb4-490f-947b-ec7ab639505c\") " Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.228953 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f63fa32-bcb4-490f-947b-ec7ab639505c-inventory\") pod \"9f63fa32-bcb4-490f-947b-ec7ab639505c\" (UID: \"9f63fa32-bcb4-490f-947b-ec7ab639505c\") " Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.229055 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzsdx\" (UniqueName: \"kubernetes.io/projected/9f63fa32-bcb4-490f-947b-ec7ab639505c-kube-api-access-hzsdx\") pod \"9f63fa32-bcb4-490f-947b-ec7ab639505c\" (UID: \"9f63fa32-bcb4-490f-947b-ec7ab639505c\") " Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.234927 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f63fa32-bcb4-490f-947b-ec7ab639505c-kube-api-access-hzsdx" (OuterVolumeSpecName: "kube-api-access-hzsdx") pod "9f63fa32-bcb4-490f-947b-ec7ab639505c" (UID: "9f63fa32-bcb4-490f-947b-ec7ab639505c"). InnerVolumeSpecName "kube-api-access-hzsdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.236777 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f63fa32-bcb4-490f-947b-ec7ab639505c-ceph" (OuterVolumeSpecName: "ceph") pod "9f63fa32-bcb4-490f-947b-ec7ab639505c" (UID: "9f63fa32-bcb4-490f-947b-ec7ab639505c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.268800 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f63fa32-bcb4-490f-947b-ec7ab639505c-inventory" (OuterVolumeSpecName: "inventory") pod "9f63fa32-bcb4-490f-947b-ec7ab639505c" (UID: "9f63fa32-bcb4-490f-947b-ec7ab639505c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.275860 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f63fa32-bcb4-490f-947b-ec7ab639505c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9f63fa32-bcb4-490f-947b-ec7ab639505c" (UID: "9f63fa32-bcb4-490f-947b-ec7ab639505c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.332129 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzsdx\" (UniqueName: \"kubernetes.io/projected/9f63fa32-bcb4-490f-947b-ec7ab639505c-kube-api-access-hzsdx\") on node \"crc\" DevicePath \"\"" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.332169 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f63fa32-bcb4-490f-947b-ec7ab639505c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.332183 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9f63fa32-bcb4-490f-947b-ec7ab639505c-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.332198 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f63fa32-bcb4-490f-947b-ec7ab639505c-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.634984 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5" event={"ID":"9f63fa32-bcb4-490f-947b-ec7ab639505c","Type":"ContainerDied","Data":"50be77714e94a4f026c37446f4eeaaafd69bd1be7dbe5a7d60cfa34072174ad5"} Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.635035 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50be77714e94a4f026c37446f4eeaaafd69bd1be7dbe5a7d60cfa34072174ad5" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.635423 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.717697 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g"] Nov 26 01:22:00 crc kubenswrapper[4766]: E1126 01:22:00.718448 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f63fa32-bcb4-490f-947b-ec7ab639505c" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.718518 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f63fa32-bcb4-490f-947b-ec7ab639505c" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.718839 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f63fa32-bcb4-490f-947b-ec7ab639505c" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.719599 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.721307 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.721572 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.721907 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.722829 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.725298 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.748938 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g"] Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.842029 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/511ab0e6-b14c-4249-92e9-184b57a03147-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9t59g\" (UID: \"511ab0e6-b14c-4249-92e9-184b57a03147\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.842083 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qs2sm\" (UniqueName: \"kubernetes.io/projected/511ab0e6-b14c-4249-92e9-184b57a03147-kube-api-access-qs2sm\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9t59g\" (UID: \"511ab0e6-b14c-4249-92e9-184b57a03147\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.842111 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/511ab0e6-b14c-4249-92e9-184b57a03147-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9t59g\" (UID: \"511ab0e6-b14c-4249-92e9-184b57a03147\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.842207 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/511ab0e6-b14c-4249-92e9-184b57a03147-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9t59g\" (UID: \"511ab0e6-b14c-4249-92e9-184b57a03147\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.944551 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/511ab0e6-b14c-4249-92e9-184b57a03147-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9t59g\" (UID: \"511ab0e6-b14c-4249-92e9-184b57a03147\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.944928 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qs2sm\" (UniqueName: \"kubernetes.io/projected/511ab0e6-b14c-4249-92e9-184b57a03147-kube-api-access-qs2sm\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9t59g\" (UID: \"511ab0e6-b14c-4249-92e9-184b57a03147\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.945129 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/511ab0e6-b14c-4249-92e9-184b57a03147-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9t59g\" (UID: \"511ab0e6-b14c-4249-92e9-184b57a03147\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.947037 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/511ab0e6-b14c-4249-92e9-184b57a03147-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9t59g\" (UID: \"511ab0e6-b14c-4249-92e9-184b57a03147\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.950561 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/511ab0e6-b14c-4249-92e9-184b57a03147-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9t59g\" (UID: \"511ab0e6-b14c-4249-92e9-184b57a03147\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.951277 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/511ab0e6-b14c-4249-92e9-184b57a03147-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9t59g\" (UID: \"511ab0e6-b14c-4249-92e9-184b57a03147\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.952878 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/511ab0e6-b14c-4249-92e9-184b57a03147-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9t59g\" (UID: \"511ab0e6-b14c-4249-92e9-184b57a03147\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g" Nov 26 01:22:00 crc kubenswrapper[4766]: I1126 01:22:00.960865 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qs2sm\" (UniqueName: \"kubernetes.io/projected/511ab0e6-b14c-4249-92e9-184b57a03147-kube-api-access-qs2sm\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9t59g\" (UID: \"511ab0e6-b14c-4249-92e9-184b57a03147\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g" Nov 26 01:22:01 crc kubenswrapper[4766]: I1126 01:22:01.043323 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g" Nov 26 01:22:01 crc kubenswrapper[4766]: I1126 01:22:01.648903 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g"] Nov 26 01:22:02 crc kubenswrapper[4766]: I1126 01:22:02.657694 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g" event={"ID":"511ab0e6-b14c-4249-92e9-184b57a03147","Type":"ContainerStarted","Data":"0c062e05c554d9d71c91e471d94f060ad6fd24810d907eac49b3547f14ba66cb"} Nov 26 01:22:02 crc kubenswrapper[4766]: I1126 01:22:02.657929 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g" event={"ID":"511ab0e6-b14c-4249-92e9-184b57a03147","Type":"ContainerStarted","Data":"8296b2d377a12b6043c4a01d91367d38ed9152623ade38e6d0005a173768a133"} Nov 26 01:22:02 crc kubenswrapper[4766]: I1126 01:22:02.692594 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g" podStartSLOduration=2.273585979 podStartE2EDuration="2.692567673s" podCreationTimestamp="2025-11-26 01:22:00 +0000 UTC" firstStartedPulling="2025-11-26 01:22:01.653958734 +0000 UTC m=+3502.502729164" lastFinishedPulling="2025-11-26 01:22:02.072940398 +0000 UTC m=+3502.921710858" observedRunningTime="2025-11-26 01:22:02.682678212 +0000 UTC m=+3503.531448672" watchObservedRunningTime="2025-11-26 01:22:02.692567673 +0000 UTC m=+3503.541338143" Nov 26 01:23:07 crc kubenswrapper[4766]: I1126 01:23:07.511999 4766 generic.go:334] "Generic (PLEG): container finished" podID="511ab0e6-b14c-4249-92e9-184b57a03147" containerID="0c062e05c554d9d71c91e471d94f060ad6fd24810d907eac49b3547f14ba66cb" exitCode=0 Nov 26 01:23:07 crc kubenswrapper[4766]: I1126 01:23:07.512104 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g" event={"ID":"511ab0e6-b14c-4249-92e9-184b57a03147","Type":"ContainerDied","Data":"0c062e05c554d9d71c91e471d94f060ad6fd24810d907eac49b3547f14ba66cb"} Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.031840 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.148473 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/511ab0e6-b14c-4249-92e9-184b57a03147-inventory\") pod \"511ab0e6-b14c-4249-92e9-184b57a03147\" (UID: \"511ab0e6-b14c-4249-92e9-184b57a03147\") " Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.148814 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/511ab0e6-b14c-4249-92e9-184b57a03147-ceph\") pod \"511ab0e6-b14c-4249-92e9-184b57a03147\" (UID: \"511ab0e6-b14c-4249-92e9-184b57a03147\") " Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.148865 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs2sm\" (UniqueName: \"kubernetes.io/projected/511ab0e6-b14c-4249-92e9-184b57a03147-kube-api-access-qs2sm\") pod \"511ab0e6-b14c-4249-92e9-184b57a03147\" (UID: \"511ab0e6-b14c-4249-92e9-184b57a03147\") " Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.149082 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/511ab0e6-b14c-4249-92e9-184b57a03147-ssh-key\") pod \"511ab0e6-b14c-4249-92e9-184b57a03147\" (UID: \"511ab0e6-b14c-4249-92e9-184b57a03147\") " Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.154846 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/511ab0e6-b14c-4249-92e9-184b57a03147-ceph" (OuterVolumeSpecName: "ceph") pod "511ab0e6-b14c-4249-92e9-184b57a03147" (UID: "511ab0e6-b14c-4249-92e9-184b57a03147"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.155369 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/511ab0e6-b14c-4249-92e9-184b57a03147-kube-api-access-qs2sm" (OuterVolumeSpecName: "kube-api-access-qs2sm") pod "511ab0e6-b14c-4249-92e9-184b57a03147" (UID: "511ab0e6-b14c-4249-92e9-184b57a03147"). InnerVolumeSpecName "kube-api-access-qs2sm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.193311 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/511ab0e6-b14c-4249-92e9-184b57a03147-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "511ab0e6-b14c-4249-92e9-184b57a03147" (UID: "511ab0e6-b14c-4249-92e9-184b57a03147"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.200030 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/511ab0e6-b14c-4249-92e9-184b57a03147-inventory" (OuterVolumeSpecName: "inventory") pod "511ab0e6-b14c-4249-92e9-184b57a03147" (UID: "511ab0e6-b14c-4249-92e9-184b57a03147"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.252036 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/511ab0e6-b14c-4249-92e9-184b57a03147-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.252069 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/511ab0e6-b14c-4249-92e9-184b57a03147-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.252083 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/511ab0e6-b14c-4249-92e9-184b57a03147-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.252094 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs2sm\" (UniqueName: \"kubernetes.io/projected/511ab0e6-b14c-4249-92e9-184b57a03147-kube-api-access-qs2sm\") on node \"crc\" DevicePath \"\"" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.548277 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g" event={"ID":"511ab0e6-b14c-4249-92e9-184b57a03147","Type":"ContainerDied","Data":"8296b2d377a12b6043c4a01d91367d38ed9152623ade38e6d0005a173768a133"} Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.548332 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8296b2d377a12b6043c4a01d91367d38ed9152623ade38e6d0005a173768a133" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.548420 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9t59g" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.668915 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-qlskl"] Nov 26 01:23:09 crc kubenswrapper[4766]: E1126 01:23:09.669476 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="511ab0e6-b14c-4249-92e9-184b57a03147" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.669498 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="511ab0e6-b14c-4249-92e9-184b57a03147" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.669772 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="511ab0e6-b14c-4249-92e9-184b57a03147" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.670614 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-qlskl" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.677195 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.680318 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.680318 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.680331 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.681298 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.689880 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-qlskl"] Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.763017 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-ceph\") pod \"ssh-known-hosts-edpm-deployment-qlskl\" (UID: \"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa\") " pod="openstack/ssh-known-hosts-edpm-deployment-qlskl" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.763153 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-qlskl\" (UID: \"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa\") " pod="openstack/ssh-known-hosts-edpm-deployment-qlskl" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.763196 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-qlskl\" (UID: \"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa\") " pod="openstack/ssh-known-hosts-edpm-deployment-qlskl" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.763258 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hs44k\" (UniqueName: \"kubernetes.io/projected/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-kube-api-access-hs44k\") pod \"ssh-known-hosts-edpm-deployment-qlskl\" (UID: \"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa\") " pod="openstack/ssh-known-hosts-edpm-deployment-qlskl" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.865335 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-ceph\") pod \"ssh-known-hosts-edpm-deployment-qlskl\" (UID: \"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa\") " pod="openstack/ssh-known-hosts-edpm-deployment-qlskl" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.865738 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-qlskl\" (UID: \"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa\") " pod="openstack/ssh-known-hosts-edpm-deployment-qlskl" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.865961 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-qlskl\" (UID: \"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa\") " pod="openstack/ssh-known-hosts-edpm-deployment-qlskl" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.866126 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hs44k\" (UniqueName: \"kubernetes.io/projected/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-kube-api-access-hs44k\") pod \"ssh-known-hosts-edpm-deployment-qlskl\" (UID: \"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa\") " pod="openstack/ssh-known-hosts-edpm-deployment-qlskl" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.872977 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-qlskl\" (UID: \"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa\") " pod="openstack/ssh-known-hosts-edpm-deployment-qlskl" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.877122 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-ceph\") pod \"ssh-known-hosts-edpm-deployment-qlskl\" (UID: \"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa\") " pod="openstack/ssh-known-hosts-edpm-deployment-qlskl" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.882767 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-qlskl\" (UID: \"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa\") " pod="openstack/ssh-known-hosts-edpm-deployment-qlskl" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.899300 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hs44k\" (UniqueName: \"kubernetes.io/projected/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-kube-api-access-hs44k\") pod \"ssh-known-hosts-edpm-deployment-qlskl\" (UID: \"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa\") " pod="openstack/ssh-known-hosts-edpm-deployment-qlskl" Nov 26 01:23:09 crc kubenswrapper[4766]: I1126 01:23:09.999693 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-qlskl" Nov 26 01:23:10 crc kubenswrapper[4766]: I1126 01:23:10.589277 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-qlskl"] Nov 26 01:23:10 crc kubenswrapper[4766]: W1126 01:23:10.595492 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57e4c16c_215b_4a8f_8c6c_a8d12e0b04fa.slice/crio-d93984db28a7465c656c38951e18822ec547b6278c5a8c407215eaeb9698f4d2 WatchSource:0}: Error finding container d93984db28a7465c656c38951e18822ec547b6278c5a8c407215eaeb9698f4d2: Status 404 returned error can't find the container with id d93984db28a7465c656c38951e18822ec547b6278c5a8c407215eaeb9698f4d2 Nov 26 01:23:10 crc kubenswrapper[4766]: I1126 01:23:10.600671 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 01:23:11 crc kubenswrapper[4766]: I1126 01:23:11.479710 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:23:11 crc kubenswrapper[4766]: I1126 01:23:11.480072 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:23:11 crc kubenswrapper[4766]: I1126 01:23:11.573153 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-qlskl" event={"ID":"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa","Type":"ContainerStarted","Data":"d93984db28a7465c656c38951e18822ec547b6278c5a8c407215eaeb9698f4d2"} Nov 26 01:23:12 crc kubenswrapper[4766]: I1126 01:23:12.589148 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-qlskl" event={"ID":"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa","Type":"ContainerStarted","Data":"c3e701c5e7d9102e9cbb98eaa4be2ca03f16a4ba327b9eddb01fe8f7ae1f9517"} Nov 26 01:23:12 crc kubenswrapper[4766]: I1126 01:23:12.608997 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-qlskl" podStartSLOduration=3.161446302 podStartE2EDuration="3.608977554s" podCreationTimestamp="2025-11-26 01:23:09 +0000 UTC" firstStartedPulling="2025-11-26 01:23:10.599038848 +0000 UTC m=+3571.447809288" lastFinishedPulling="2025-11-26 01:23:11.04657007 +0000 UTC m=+3571.895340540" observedRunningTime="2025-11-26 01:23:12.607422106 +0000 UTC m=+3573.456192586" watchObservedRunningTime="2025-11-26 01:23:12.608977554 +0000 UTC m=+3573.457747994" Nov 26 01:23:24 crc kubenswrapper[4766]: I1126 01:23:24.756009 4766 generic.go:334] "Generic (PLEG): container finished" podID="57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa" containerID="c3e701c5e7d9102e9cbb98eaa4be2ca03f16a4ba327b9eddb01fe8f7ae1f9517" exitCode=0 Nov 26 01:23:24 crc kubenswrapper[4766]: I1126 01:23:24.756095 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-qlskl" event={"ID":"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa","Type":"ContainerDied","Data":"c3e701c5e7d9102e9cbb98eaa4be2ca03f16a4ba327b9eddb01fe8f7ae1f9517"} Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.291873 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-qlskl" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.488179 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-ceph\") pod \"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa\" (UID: \"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa\") " Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.488457 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-ssh-key-openstack-edpm-ipam\") pod \"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa\" (UID: \"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa\") " Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.488484 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hs44k\" (UniqueName: \"kubernetes.io/projected/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-kube-api-access-hs44k\") pod \"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa\" (UID: \"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa\") " Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.488636 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-inventory-0\") pod \"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa\" (UID: \"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa\") " Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.498718 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-ceph" (OuterVolumeSpecName: "ceph") pod "57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa" (UID: "57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.498839 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-kube-api-access-hs44k" (OuterVolumeSpecName: "kube-api-access-hs44k") pod "57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa" (UID: "57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa"). InnerVolumeSpecName "kube-api-access-hs44k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.525366 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa" (UID: "57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.529167 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa" (UID: "57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.590886 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.590927 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.590943 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hs44k\" (UniqueName: \"kubernetes.io/projected/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-kube-api-access-hs44k\") on node \"crc\" DevicePath \"\"" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.590955 4766 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.777895 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-qlskl" event={"ID":"57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa","Type":"ContainerDied","Data":"d93984db28a7465c656c38951e18822ec547b6278c5a8c407215eaeb9698f4d2"} Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.777950 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d93984db28a7465c656c38951e18822ec547b6278c5a8c407215eaeb9698f4d2" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.778004 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-qlskl" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.891412 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58"] Nov 26 01:23:26 crc kubenswrapper[4766]: E1126 01:23:26.891862 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa" containerName="ssh-known-hosts-edpm-deployment" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.891892 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa" containerName="ssh-known-hosts-edpm-deployment" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.892210 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa" containerName="ssh-known-hosts-edpm-deployment" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.893150 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.895048 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.896200 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.897116 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.897281 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.897573 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.910925 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58"] Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.998553 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwlfn\" (UniqueName: \"kubernetes.io/projected/14ea5f7a-f044-426a-919e-a05cdcfe900c-kube-api-access-lwlfn\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tzd58\" (UID: \"14ea5f7a-f044-426a-919e-a05cdcfe900c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.998767 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/14ea5f7a-f044-426a-919e-a05cdcfe900c-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tzd58\" (UID: \"14ea5f7a-f044-426a-919e-a05cdcfe900c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.999071 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/14ea5f7a-f044-426a-919e-a05cdcfe900c-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tzd58\" (UID: \"14ea5f7a-f044-426a-919e-a05cdcfe900c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58" Nov 26 01:23:26 crc kubenswrapper[4766]: I1126 01:23:26.999134 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14ea5f7a-f044-426a-919e-a05cdcfe900c-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tzd58\" (UID: \"14ea5f7a-f044-426a-919e-a05cdcfe900c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58" Nov 26 01:23:27 crc kubenswrapper[4766]: I1126 01:23:27.102050 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/14ea5f7a-f044-426a-919e-a05cdcfe900c-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tzd58\" (UID: \"14ea5f7a-f044-426a-919e-a05cdcfe900c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58" Nov 26 01:23:27 crc kubenswrapper[4766]: I1126 01:23:27.102123 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14ea5f7a-f044-426a-919e-a05cdcfe900c-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tzd58\" (UID: \"14ea5f7a-f044-426a-919e-a05cdcfe900c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58" Nov 26 01:23:27 crc kubenswrapper[4766]: I1126 01:23:27.102288 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwlfn\" (UniqueName: \"kubernetes.io/projected/14ea5f7a-f044-426a-919e-a05cdcfe900c-kube-api-access-lwlfn\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tzd58\" (UID: \"14ea5f7a-f044-426a-919e-a05cdcfe900c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58" Nov 26 01:23:27 crc kubenswrapper[4766]: I1126 01:23:27.102396 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/14ea5f7a-f044-426a-919e-a05cdcfe900c-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tzd58\" (UID: \"14ea5f7a-f044-426a-919e-a05cdcfe900c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58" Nov 26 01:23:27 crc kubenswrapper[4766]: I1126 01:23:27.108144 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/14ea5f7a-f044-426a-919e-a05cdcfe900c-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tzd58\" (UID: \"14ea5f7a-f044-426a-919e-a05cdcfe900c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58" Nov 26 01:23:27 crc kubenswrapper[4766]: I1126 01:23:27.108291 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/14ea5f7a-f044-426a-919e-a05cdcfe900c-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tzd58\" (UID: \"14ea5f7a-f044-426a-919e-a05cdcfe900c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58" Nov 26 01:23:27 crc kubenswrapper[4766]: I1126 01:23:27.108363 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14ea5f7a-f044-426a-919e-a05cdcfe900c-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tzd58\" (UID: \"14ea5f7a-f044-426a-919e-a05cdcfe900c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58" Nov 26 01:23:27 crc kubenswrapper[4766]: I1126 01:23:27.134337 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwlfn\" (UniqueName: \"kubernetes.io/projected/14ea5f7a-f044-426a-919e-a05cdcfe900c-kube-api-access-lwlfn\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-tzd58\" (UID: \"14ea5f7a-f044-426a-919e-a05cdcfe900c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58" Nov 26 01:23:27 crc kubenswrapper[4766]: I1126 01:23:27.219000 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58" Nov 26 01:23:27 crc kubenswrapper[4766]: I1126 01:23:27.818941 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58"] Nov 26 01:23:28 crc kubenswrapper[4766]: I1126 01:23:28.807364 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58" event={"ID":"14ea5f7a-f044-426a-919e-a05cdcfe900c","Type":"ContainerStarted","Data":"153b6ddd64cbbb4fcdc580a7baf27e5b142eecb43a6428f09779dac89da60581"} Nov 26 01:23:28 crc kubenswrapper[4766]: I1126 01:23:28.808048 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58" event={"ID":"14ea5f7a-f044-426a-919e-a05cdcfe900c","Type":"ContainerStarted","Data":"603eb79b390db58d87d793f1e34712e81192c061b56332923bcb9130339e4bdc"} Nov 26 01:23:28 crc kubenswrapper[4766]: I1126 01:23:28.831767 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58" podStartSLOduration=2.381871419 podStartE2EDuration="2.831737948s" podCreationTimestamp="2025-11-26 01:23:26 +0000 UTC" firstStartedPulling="2025-11-26 01:23:27.822885025 +0000 UTC m=+3588.671655495" lastFinishedPulling="2025-11-26 01:23:28.272751554 +0000 UTC m=+3589.121522024" observedRunningTime="2025-11-26 01:23:28.827925605 +0000 UTC m=+3589.676696075" watchObservedRunningTime="2025-11-26 01:23:28.831737948 +0000 UTC m=+3589.680508418" Nov 26 01:23:38 crc kubenswrapper[4766]: I1126 01:23:38.916498 4766 generic.go:334] "Generic (PLEG): container finished" podID="14ea5f7a-f044-426a-919e-a05cdcfe900c" containerID="153b6ddd64cbbb4fcdc580a7baf27e5b142eecb43a6428f09779dac89da60581" exitCode=0 Nov 26 01:23:38 crc kubenswrapper[4766]: I1126 01:23:38.916704 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58" event={"ID":"14ea5f7a-f044-426a-919e-a05cdcfe900c","Type":"ContainerDied","Data":"153b6ddd64cbbb4fcdc580a7baf27e5b142eecb43a6428f09779dac89da60581"} Nov 26 01:23:40 crc kubenswrapper[4766]: I1126 01:23:40.531227 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58" Nov 26 01:23:40 crc kubenswrapper[4766]: I1126 01:23:40.616877 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/14ea5f7a-f044-426a-919e-a05cdcfe900c-ssh-key\") pod \"14ea5f7a-f044-426a-919e-a05cdcfe900c\" (UID: \"14ea5f7a-f044-426a-919e-a05cdcfe900c\") " Nov 26 01:23:40 crc kubenswrapper[4766]: I1126 01:23:40.617020 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwlfn\" (UniqueName: \"kubernetes.io/projected/14ea5f7a-f044-426a-919e-a05cdcfe900c-kube-api-access-lwlfn\") pod \"14ea5f7a-f044-426a-919e-a05cdcfe900c\" (UID: \"14ea5f7a-f044-426a-919e-a05cdcfe900c\") " Nov 26 01:23:40 crc kubenswrapper[4766]: I1126 01:23:40.617139 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14ea5f7a-f044-426a-919e-a05cdcfe900c-inventory\") pod \"14ea5f7a-f044-426a-919e-a05cdcfe900c\" (UID: \"14ea5f7a-f044-426a-919e-a05cdcfe900c\") " Nov 26 01:23:40 crc kubenswrapper[4766]: I1126 01:23:40.617196 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/14ea5f7a-f044-426a-919e-a05cdcfe900c-ceph\") pod \"14ea5f7a-f044-426a-919e-a05cdcfe900c\" (UID: \"14ea5f7a-f044-426a-919e-a05cdcfe900c\") " Nov 26 01:23:40 crc kubenswrapper[4766]: I1126 01:23:40.623571 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ea5f7a-f044-426a-919e-a05cdcfe900c-ceph" (OuterVolumeSpecName: "ceph") pod "14ea5f7a-f044-426a-919e-a05cdcfe900c" (UID: "14ea5f7a-f044-426a-919e-a05cdcfe900c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:23:40 crc kubenswrapper[4766]: I1126 01:23:40.625222 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14ea5f7a-f044-426a-919e-a05cdcfe900c-kube-api-access-lwlfn" (OuterVolumeSpecName: "kube-api-access-lwlfn") pod "14ea5f7a-f044-426a-919e-a05cdcfe900c" (UID: "14ea5f7a-f044-426a-919e-a05cdcfe900c"). InnerVolumeSpecName "kube-api-access-lwlfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:23:40 crc kubenswrapper[4766]: I1126 01:23:40.662956 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ea5f7a-f044-426a-919e-a05cdcfe900c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "14ea5f7a-f044-426a-919e-a05cdcfe900c" (UID: "14ea5f7a-f044-426a-919e-a05cdcfe900c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:23:40 crc kubenswrapper[4766]: I1126 01:23:40.672955 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14ea5f7a-f044-426a-919e-a05cdcfe900c-inventory" (OuterVolumeSpecName: "inventory") pod "14ea5f7a-f044-426a-919e-a05cdcfe900c" (UID: "14ea5f7a-f044-426a-919e-a05cdcfe900c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:23:40 crc kubenswrapper[4766]: I1126 01:23:40.719473 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwlfn\" (UniqueName: \"kubernetes.io/projected/14ea5f7a-f044-426a-919e-a05cdcfe900c-kube-api-access-lwlfn\") on node \"crc\" DevicePath \"\"" Nov 26 01:23:40 crc kubenswrapper[4766]: I1126 01:23:40.719503 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14ea5f7a-f044-426a-919e-a05cdcfe900c-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:23:40 crc kubenswrapper[4766]: I1126 01:23:40.719513 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/14ea5f7a-f044-426a-919e-a05cdcfe900c-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:23:40 crc kubenswrapper[4766]: I1126 01:23:40.719521 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/14ea5f7a-f044-426a-919e-a05cdcfe900c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:23:40 crc kubenswrapper[4766]: I1126 01:23:40.944742 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58" event={"ID":"14ea5f7a-f044-426a-919e-a05cdcfe900c","Type":"ContainerDied","Data":"603eb79b390db58d87d793f1e34712e81192c061b56332923bcb9130339e4bdc"} Nov 26 01:23:40 crc kubenswrapper[4766]: I1126 01:23:40.944782 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="603eb79b390db58d87d793f1e34712e81192c061b56332923bcb9130339e4bdc" Nov 26 01:23:40 crc kubenswrapper[4766]: I1126 01:23:40.944838 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-tzd58" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.070842 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj"] Nov 26 01:23:41 crc kubenswrapper[4766]: E1126 01:23:41.071247 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ea5f7a-f044-426a-919e-a05cdcfe900c" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.071264 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ea5f7a-f044-426a-919e-a05cdcfe900c" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.071517 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="14ea5f7a-f044-426a-919e-a05cdcfe900c" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.072387 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.074812 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.075152 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.075186 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.075312 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.079614 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.088717 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj"] Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.127090 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj\" (UID: \"b59fb3c0-2626-47c2-b2f4-f13c13b011a5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.127165 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj\" (UID: \"b59fb3c0-2626-47c2-b2f4-f13c13b011a5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.127208 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj\" (UID: \"b59fb3c0-2626-47c2-b2f4-f13c13b011a5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.127276 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn6fz\" (UniqueName: \"kubernetes.io/projected/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-kube-api-access-vn6fz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj\" (UID: \"b59fb3c0-2626-47c2-b2f4-f13c13b011a5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.228545 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj\" (UID: \"b59fb3c0-2626-47c2-b2f4-f13c13b011a5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.228873 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj\" (UID: \"b59fb3c0-2626-47c2-b2f4-f13c13b011a5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.228909 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj\" (UID: \"b59fb3c0-2626-47c2-b2f4-f13c13b011a5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.228978 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn6fz\" (UniqueName: \"kubernetes.io/projected/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-kube-api-access-vn6fz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj\" (UID: \"b59fb3c0-2626-47c2-b2f4-f13c13b011a5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.233705 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj\" (UID: \"b59fb3c0-2626-47c2-b2f4-f13c13b011a5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.233920 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj\" (UID: \"b59fb3c0-2626-47c2-b2f4-f13c13b011a5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.236935 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj\" (UID: \"b59fb3c0-2626-47c2-b2f4-f13c13b011a5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.246717 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn6fz\" (UniqueName: \"kubernetes.io/projected/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-kube-api-access-vn6fz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj\" (UID: \"b59fb3c0-2626-47c2-b2f4-f13c13b011a5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.395353 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj" Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.479248 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:23:41 crc kubenswrapper[4766]: I1126 01:23:41.479313 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:23:42 crc kubenswrapper[4766]: W1126 01:23:42.030852 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb59fb3c0_2626_47c2_b2f4_f13c13b011a5.slice/crio-acc987b91d7e88c0c63b87e88d44ad040cc375e4b2c9d76427889dae9435da35 WatchSource:0}: Error finding container acc987b91d7e88c0c63b87e88d44ad040cc375e4b2c9d76427889dae9435da35: Status 404 returned error can't find the container with id acc987b91d7e88c0c63b87e88d44ad040cc375e4b2c9d76427889dae9435da35 Nov 26 01:23:42 crc kubenswrapper[4766]: I1126 01:23:42.040601 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj"] Nov 26 01:23:42 crc kubenswrapper[4766]: I1126 01:23:42.969353 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj" event={"ID":"b59fb3c0-2626-47c2-b2f4-f13c13b011a5","Type":"ContainerStarted","Data":"acc987b91d7e88c0c63b87e88d44ad040cc375e4b2c9d76427889dae9435da35"} Nov 26 01:23:43 crc kubenswrapper[4766]: I1126 01:23:43.978363 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj" event={"ID":"b59fb3c0-2626-47c2-b2f4-f13c13b011a5","Type":"ContainerStarted","Data":"a902bf19e4aa417de19290ebbd6206821d02c2cb6d1cc60e694c6c7a9d6b1668"} Nov 26 01:23:44 crc kubenswrapper[4766]: I1126 01:23:44.010623 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj" podStartSLOduration=2.337917995 podStartE2EDuration="3.010606286s" podCreationTimestamp="2025-11-26 01:23:41 +0000 UTC" firstStartedPulling="2025-11-26 01:23:42.034051676 +0000 UTC m=+3602.882822126" lastFinishedPulling="2025-11-26 01:23:42.706739947 +0000 UTC m=+3603.555510417" observedRunningTime="2025-11-26 01:23:43.994286717 +0000 UTC m=+3604.843057147" watchObservedRunningTime="2025-11-26 01:23:44.010606286 +0000 UTC m=+3604.859376706" Nov 26 01:23:56 crc kubenswrapper[4766]: I1126 01:23:56.147259 4766 generic.go:334] "Generic (PLEG): container finished" podID="b59fb3c0-2626-47c2-b2f4-f13c13b011a5" containerID="a902bf19e4aa417de19290ebbd6206821d02c2cb6d1cc60e694c6c7a9d6b1668" exitCode=0 Nov 26 01:23:56 crc kubenswrapper[4766]: I1126 01:23:56.147387 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj" event={"ID":"b59fb3c0-2626-47c2-b2f4-f13c13b011a5","Type":"ContainerDied","Data":"a902bf19e4aa417de19290ebbd6206821d02c2cb6d1cc60e694c6c7a9d6b1668"} Nov 26 01:23:57 crc kubenswrapper[4766]: I1126 01:23:57.785047 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj" Nov 26 01:23:57 crc kubenswrapper[4766]: I1126 01:23:57.830839 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-ceph\") pod \"b59fb3c0-2626-47c2-b2f4-f13c13b011a5\" (UID: \"b59fb3c0-2626-47c2-b2f4-f13c13b011a5\") " Nov 26 01:23:57 crc kubenswrapper[4766]: I1126 01:23:57.830921 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-ssh-key\") pod \"b59fb3c0-2626-47c2-b2f4-f13c13b011a5\" (UID: \"b59fb3c0-2626-47c2-b2f4-f13c13b011a5\") " Nov 26 01:23:57 crc kubenswrapper[4766]: I1126 01:23:57.831019 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vn6fz\" (UniqueName: \"kubernetes.io/projected/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-kube-api-access-vn6fz\") pod \"b59fb3c0-2626-47c2-b2f4-f13c13b011a5\" (UID: \"b59fb3c0-2626-47c2-b2f4-f13c13b011a5\") " Nov 26 01:23:57 crc kubenswrapper[4766]: I1126 01:23:57.831106 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-inventory\") pod \"b59fb3c0-2626-47c2-b2f4-f13c13b011a5\" (UID: \"b59fb3c0-2626-47c2-b2f4-f13c13b011a5\") " Nov 26 01:23:57 crc kubenswrapper[4766]: I1126 01:23:57.843219 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-kube-api-access-vn6fz" (OuterVolumeSpecName: "kube-api-access-vn6fz") pod "b59fb3c0-2626-47c2-b2f4-f13c13b011a5" (UID: "b59fb3c0-2626-47c2-b2f4-f13c13b011a5"). InnerVolumeSpecName "kube-api-access-vn6fz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:23:57 crc kubenswrapper[4766]: I1126 01:23:57.858034 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-ceph" (OuterVolumeSpecName: "ceph") pod "b59fb3c0-2626-47c2-b2f4-f13c13b011a5" (UID: "b59fb3c0-2626-47c2-b2f4-f13c13b011a5"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:23:57 crc kubenswrapper[4766]: I1126 01:23:57.878894 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b59fb3c0-2626-47c2-b2f4-f13c13b011a5" (UID: "b59fb3c0-2626-47c2-b2f4-f13c13b011a5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:23:57 crc kubenswrapper[4766]: I1126 01:23:57.889914 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-inventory" (OuterVolumeSpecName: "inventory") pod "b59fb3c0-2626-47c2-b2f4-f13c13b011a5" (UID: "b59fb3c0-2626-47c2-b2f4-f13c13b011a5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:23:57 crc kubenswrapper[4766]: I1126 01:23:57.933484 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:23:57 crc kubenswrapper[4766]: I1126 01:23:57.933515 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:23:57 crc kubenswrapper[4766]: I1126 01:23:57.933532 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vn6fz\" (UniqueName: \"kubernetes.io/projected/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-kube-api-access-vn6fz\") on node \"crc\" DevicePath \"\"" Nov 26 01:23:57 crc kubenswrapper[4766]: I1126 01:23:57.933545 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b59fb3c0-2626-47c2-b2f4-f13c13b011a5-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.174025 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj" event={"ID":"b59fb3c0-2626-47c2-b2f4-f13c13b011a5","Type":"ContainerDied","Data":"acc987b91d7e88c0c63b87e88d44ad040cc375e4b2c9d76427889dae9435da35"} Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.174074 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="acc987b91d7e88c0c63b87e88d44ad040cc375e4b2c9d76427889dae9435da35" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.174151 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.314587 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n"] Nov 26 01:23:58 crc kubenswrapper[4766]: E1126 01:23:58.315763 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b59fb3c0-2626-47c2-b2f4-f13c13b011a5" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.315783 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59fb3c0-2626-47c2-b2f4-f13c13b011a5" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.316434 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="b59fb3c0-2626-47c2-b2f4-f13c13b011a5" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.319466 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.345007 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.345494 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.345982 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n"] Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.346121 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.346256 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.346670 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.347836 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.348799 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.348978 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.349302 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.354312 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.355818 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.355864 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.355889 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.355919 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.355958 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.355987 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.356110 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.356192 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.356224 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.356283 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.356320 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.356345 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.356375 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdnt6\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-kube-api-access-qdnt6\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.356434 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.356471 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.356571 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.356631 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.461917 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.461968 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.462036 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.462089 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.462120 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.462142 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.462160 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.462182 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.462204 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.462222 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.462260 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.462299 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.462317 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.462352 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.462448 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.462704 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.462731 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdnt6\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-kube-api-access-qdnt6\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.465392 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.465734 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.466955 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.467298 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.468135 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.468730 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.468936 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.469718 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.469948 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.470007 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.470344 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.470440 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.471514 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.471837 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.471972 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.476723 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.479554 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdnt6\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-kube-api-access-qdnt6\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-l279n\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:58 crc kubenswrapper[4766]: I1126 01:23:58.672885 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:23:59 crc kubenswrapper[4766]: W1126 01:23:59.263138 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7509ec1a_b269_4bbe_bfca_0dcb052360ea.slice/crio-1dae38851335968cdf8dfa100555d192f052400bb4382d4cb2e732ae2a9504e4 WatchSource:0}: Error finding container 1dae38851335968cdf8dfa100555d192f052400bb4382d4cb2e732ae2a9504e4: Status 404 returned error can't find the container with id 1dae38851335968cdf8dfa100555d192f052400bb4382d4cb2e732ae2a9504e4 Nov 26 01:23:59 crc kubenswrapper[4766]: I1126 01:23:59.265839 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n"] Nov 26 01:24:00 crc kubenswrapper[4766]: I1126 01:24:00.220887 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" event={"ID":"7509ec1a-b269-4bbe-bfca-0dcb052360ea","Type":"ContainerStarted","Data":"e675dd4dd6095979272355d89c719f06b37233203371017c8360c0a378657c51"} Nov 26 01:24:00 crc kubenswrapper[4766]: I1126 01:24:00.221350 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" event={"ID":"7509ec1a-b269-4bbe-bfca-0dcb052360ea","Type":"ContainerStarted","Data":"1dae38851335968cdf8dfa100555d192f052400bb4382d4cb2e732ae2a9504e4"} Nov 26 01:24:00 crc kubenswrapper[4766]: I1126 01:24:00.255269 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" podStartSLOduration=1.805967156 podStartE2EDuration="2.255253608s" podCreationTimestamp="2025-11-26 01:23:58 +0000 UTC" firstStartedPulling="2025-11-26 01:23:59.266721629 +0000 UTC m=+3620.115492059" lastFinishedPulling="2025-11-26 01:23:59.716008071 +0000 UTC m=+3620.564778511" observedRunningTime="2025-11-26 01:24:00.246855104 +0000 UTC m=+3621.095625524" watchObservedRunningTime="2025-11-26 01:24:00.255253608 +0000 UTC m=+3621.104024038" Nov 26 01:24:11 crc kubenswrapper[4766]: I1126 01:24:11.480130 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:24:11 crc kubenswrapper[4766]: I1126 01:24:11.480985 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:24:11 crc kubenswrapper[4766]: I1126 01:24:11.481052 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 01:24:11 crc kubenswrapper[4766]: I1126 01:24:11.482243 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 01:24:11 crc kubenswrapper[4766]: I1126 01:24:11.482336 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" gracePeriod=600 Nov 26 01:24:11 crc kubenswrapper[4766]: E1126 01:24:11.942611 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:24:12 crc kubenswrapper[4766]: I1126 01:24:12.889560 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" exitCode=0 Nov 26 01:24:12 crc kubenswrapper[4766]: I1126 01:24:12.889627 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24"} Nov 26 01:24:12 crc kubenswrapper[4766]: I1126 01:24:12.889715 4766 scope.go:117] "RemoveContainer" containerID="207a49c35f1945eec7efffff2aa31a2a80326afc58102d8b59af0cccdc0adb5e" Nov 26 01:24:12 crc kubenswrapper[4766]: I1126 01:24:12.890678 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:24:12 crc kubenswrapper[4766]: E1126 01:24:12.891234 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:24:27 crc kubenswrapper[4766]: I1126 01:24:27.828931 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:24:27 crc kubenswrapper[4766]: E1126 01:24:27.829949 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:24:38 crc kubenswrapper[4766]: I1126 01:24:38.827188 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:24:38 crc kubenswrapper[4766]: E1126 01:24:38.828292 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.129027 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vmzkg"] Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.136996 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vmzkg" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.172945 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vmzkg"] Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.212105 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02c511b4-bb27-45a2-bc48-cb21f30158af-catalog-content\") pod \"redhat-marketplace-vmzkg\" (UID: \"02c511b4-bb27-45a2-bc48-cb21f30158af\") " pod="openshift-marketplace/redhat-marketplace-vmzkg" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.212189 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02c511b4-bb27-45a2-bc48-cb21f30158af-utilities\") pod \"redhat-marketplace-vmzkg\" (UID: \"02c511b4-bb27-45a2-bc48-cb21f30158af\") " pod="openshift-marketplace/redhat-marketplace-vmzkg" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.212249 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm9dw\" (UniqueName: \"kubernetes.io/projected/02c511b4-bb27-45a2-bc48-cb21f30158af-kube-api-access-zm9dw\") pod \"redhat-marketplace-vmzkg\" (UID: \"02c511b4-bb27-45a2-bc48-cb21f30158af\") " pod="openshift-marketplace/redhat-marketplace-vmzkg" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.309970 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gd7g9"] Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.312765 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gd7g9" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.313879 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02c511b4-bb27-45a2-bc48-cb21f30158af-catalog-content\") pod \"redhat-marketplace-vmzkg\" (UID: \"02c511b4-bb27-45a2-bc48-cb21f30158af\") " pod="openshift-marketplace/redhat-marketplace-vmzkg" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.313971 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02c511b4-bb27-45a2-bc48-cb21f30158af-utilities\") pod \"redhat-marketplace-vmzkg\" (UID: \"02c511b4-bb27-45a2-bc48-cb21f30158af\") " pod="openshift-marketplace/redhat-marketplace-vmzkg" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.314012 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm9dw\" (UniqueName: \"kubernetes.io/projected/02c511b4-bb27-45a2-bc48-cb21f30158af-kube-api-access-zm9dw\") pod \"redhat-marketplace-vmzkg\" (UID: \"02c511b4-bb27-45a2-bc48-cb21f30158af\") " pod="openshift-marketplace/redhat-marketplace-vmzkg" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.314735 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02c511b4-bb27-45a2-bc48-cb21f30158af-utilities\") pod \"redhat-marketplace-vmzkg\" (UID: \"02c511b4-bb27-45a2-bc48-cb21f30158af\") " pod="openshift-marketplace/redhat-marketplace-vmzkg" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.314934 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02c511b4-bb27-45a2-bc48-cb21f30158af-catalog-content\") pod \"redhat-marketplace-vmzkg\" (UID: \"02c511b4-bb27-45a2-bc48-cb21f30158af\") " pod="openshift-marketplace/redhat-marketplace-vmzkg" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.335079 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gd7g9"] Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.363172 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm9dw\" (UniqueName: \"kubernetes.io/projected/02c511b4-bb27-45a2-bc48-cb21f30158af-kube-api-access-zm9dw\") pod \"redhat-marketplace-vmzkg\" (UID: \"02c511b4-bb27-45a2-bc48-cb21f30158af\") " pod="openshift-marketplace/redhat-marketplace-vmzkg" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.416501 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj92c\" (UniqueName: \"kubernetes.io/projected/3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf-kube-api-access-hj92c\") pod \"redhat-operators-gd7g9\" (UID: \"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf\") " pod="openshift-marketplace/redhat-operators-gd7g9" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.416861 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf-catalog-content\") pod \"redhat-operators-gd7g9\" (UID: \"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf\") " pod="openshift-marketplace/redhat-operators-gd7g9" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.416916 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf-utilities\") pod \"redhat-operators-gd7g9\" (UID: \"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf\") " pod="openshift-marketplace/redhat-operators-gd7g9" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.477311 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vmzkg" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.519804 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj92c\" (UniqueName: \"kubernetes.io/projected/3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf-kube-api-access-hj92c\") pod \"redhat-operators-gd7g9\" (UID: \"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf\") " pod="openshift-marketplace/redhat-operators-gd7g9" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.519881 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf-catalog-content\") pod \"redhat-operators-gd7g9\" (UID: \"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf\") " pod="openshift-marketplace/redhat-operators-gd7g9" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.519942 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf-utilities\") pod \"redhat-operators-gd7g9\" (UID: \"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf\") " pod="openshift-marketplace/redhat-operators-gd7g9" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.520511 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf-catalog-content\") pod \"redhat-operators-gd7g9\" (UID: \"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf\") " pod="openshift-marketplace/redhat-operators-gd7g9" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.523372 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf-utilities\") pod \"redhat-operators-gd7g9\" (UID: \"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf\") " pod="openshift-marketplace/redhat-operators-gd7g9" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.545441 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj92c\" (UniqueName: \"kubernetes.io/projected/3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf-kube-api-access-hj92c\") pod \"redhat-operators-gd7g9\" (UID: \"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf\") " pod="openshift-marketplace/redhat-operators-gd7g9" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.644869 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gd7g9" Nov 26 01:24:51 crc kubenswrapper[4766]: I1126 01:24:51.971271 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vmzkg"] Nov 26 01:24:52 crc kubenswrapper[4766]: I1126 01:24:52.125748 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gd7g9"] Nov 26 01:24:52 crc kubenswrapper[4766]: I1126 01:24:52.397146 4766 generic.go:334] "Generic (PLEG): container finished" podID="02c511b4-bb27-45a2-bc48-cb21f30158af" containerID="00888a4dbda175f5c4892124c37b1f571439cf4e14c4949c2f36ca84654597a4" exitCode=0 Nov 26 01:24:52 crc kubenswrapper[4766]: I1126 01:24:52.397215 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vmzkg" event={"ID":"02c511b4-bb27-45a2-bc48-cb21f30158af","Type":"ContainerDied","Data":"00888a4dbda175f5c4892124c37b1f571439cf4e14c4949c2f36ca84654597a4"} Nov 26 01:24:52 crc kubenswrapper[4766]: I1126 01:24:52.397243 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vmzkg" event={"ID":"02c511b4-bb27-45a2-bc48-cb21f30158af","Type":"ContainerStarted","Data":"91e4634373fea50582aff68d2d94f57a8625c20f4e400eccc7a96ec6c35c7e49"} Nov 26 01:24:52 crc kubenswrapper[4766]: I1126 01:24:52.398986 4766 generic.go:334] "Generic (PLEG): container finished" podID="3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf" containerID="b2af83b85a9cc51edd772ea31a1b8e6149c3f9a8c0b3ad3aa01ff0c586420ade" exitCode=0 Nov 26 01:24:52 crc kubenswrapper[4766]: I1126 01:24:52.399033 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gd7g9" event={"ID":"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf","Type":"ContainerDied","Data":"b2af83b85a9cc51edd772ea31a1b8e6149c3f9a8c0b3ad3aa01ff0c586420ade"} Nov 26 01:24:52 crc kubenswrapper[4766]: I1126 01:24:52.399062 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gd7g9" event={"ID":"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf","Type":"ContainerStarted","Data":"b116cb3bf0f6ebc6b4dc2056949058fe6f8343a7069262cf0b290de3d48f73d1"} Nov 26 01:24:52 crc kubenswrapper[4766]: I1126 01:24:52.827575 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:24:52 crc kubenswrapper[4766]: E1126 01:24:52.827882 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:24:53 crc kubenswrapper[4766]: I1126 01:24:53.419112 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gd7g9" event={"ID":"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf","Type":"ContainerStarted","Data":"ddf93ecfd0a2e8c8fa5d5ecb62d0a03ab918594b140b199b6f4d327dadc92a90"} Nov 26 01:24:53 crc kubenswrapper[4766]: I1126 01:24:53.424172 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vmzkg" event={"ID":"02c511b4-bb27-45a2-bc48-cb21f30158af","Type":"ContainerStarted","Data":"ea970a16ba0ba26b59e65b37d674ff8ed0d4fa8b7aca852222e8f4d71bb8c8dc"} Nov 26 01:24:54 crc kubenswrapper[4766]: I1126 01:24:54.438632 4766 generic.go:334] "Generic (PLEG): container finished" podID="02c511b4-bb27-45a2-bc48-cb21f30158af" containerID="ea970a16ba0ba26b59e65b37d674ff8ed0d4fa8b7aca852222e8f4d71bb8c8dc" exitCode=0 Nov 26 01:24:54 crc kubenswrapper[4766]: I1126 01:24:54.438717 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vmzkg" event={"ID":"02c511b4-bb27-45a2-bc48-cb21f30158af","Type":"ContainerDied","Data":"ea970a16ba0ba26b59e65b37d674ff8ed0d4fa8b7aca852222e8f4d71bb8c8dc"} Nov 26 01:24:54 crc kubenswrapper[4766]: I1126 01:24:54.717594 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jvvwl"] Nov 26 01:24:54 crc kubenswrapper[4766]: I1126 01:24:54.721624 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jvvwl" Nov 26 01:24:54 crc kubenswrapper[4766]: I1126 01:24:54.737779 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jvvwl"] Nov 26 01:24:54 crc kubenswrapper[4766]: I1126 01:24:54.822448 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzdnp\" (UniqueName: \"kubernetes.io/projected/74038ed5-c727-40ae-bc0f-d779aa47613d-kube-api-access-xzdnp\") pod \"certified-operators-jvvwl\" (UID: \"74038ed5-c727-40ae-bc0f-d779aa47613d\") " pod="openshift-marketplace/certified-operators-jvvwl" Nov 26 01:24:54 crc kubenswrapper[4766]: I1126 01:24:54.822743 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74038ed5-c727-40ae-bc0f-d779aa47613d-utilities\") pod \"certified-operators-jvvwl\" (UID: \"74038ed5-c727-40ae-bc0f-d779aa47613d\") " pod="openshift-marketplace/certified-operators-jvvwl" Nov 26 01:24:54 crc kubenswrapper[4766]: I1126 01:24:54.822790 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74038ed5-c727-40ae-bc0f-d779aa47613d-catalog-content\") pod \"certified-operators-jvvwl\" (UID: \"74038ed5-c727-40ae-bc0f-d779aa47613d\") " pod="openshift-marketplace/certified-operators-jvvwl" Nov 26 01:24:54 crc kubenswrapper[4766]: I1126 01:24:54.924610 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74038ed5-c727-40ae-bc0f-d779aa47613d-utilities\") pod \"certified-operators-jvvwl\" (UID: \"74038ed5-c727-40ae-bc0f-d779aa47613d\") " pod="openshift-marketplace/certified-operators-jvvwl" Nov 26 01:24:54 crc kubenswrapper[4766]: I1126 01:24:54.924662 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74038ed5-c727-40ae-bc0f-d779aa47613d-catalog-content\") pod \"certified-operators-jvvwl\" (UID: \"74038ed5-c727-40ae-bc0f-d779aa47613d\") " pod="openshift-marketplace/certified-operators-jvvwl" Nov 26 01:24:54 crc kubenswrapper[4766]: I1126 01:24:54.924805 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzdnp\" (UniqueName: \"kubernetes.io/projected/74038ed5-c727-40ae-bc0f-d779aa47613d-kube-api-access-xzdnp\") pod \"certified-operators-jvvwl\" (UID: \"74038ed5-c727-40ae-bc0f-d779aa47613d\") " pod="openshift-marketplace/certified-operators-jvvwl" Nov 26 01:24:54 crc kubenswrapper[4766]: I1126 01:24:54.925909 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74038ed5-c727-40ae-bc0f-d779aa47613d-utilities\") pod \"certified-operators-jvvwl\" (UID: \"74038ed5-c727-40ae-bc0f-d779aa47613d\") " pod="openshift-marketplace/certified-operators-jvvwl" Nov 26 01:24:54 crc kubenswrapper[4766]: I1126 01:24:54.926006 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74038ed5-c727-40ae-bc0f-d779aa47613d-catalog-content\") pod \"certified-operators-jvvwl\" (UID: \"74038ed5-c727-40ae-bc0f-d779aa47613d\") " pod="openshift-marketplace/certified-operators-jvvwl" Nov 26 01:24:54 crc kubenswrapper[4766]: I1126 01:24:54.953710 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzdnp\" (UniqueName: \"kubernetes.io/projected/74038ed5-c727-40ae-bc0f-d779aa47613d-kube-api-access-xzdnp\") pod \"certified-operators-jvvwl\" (UID: \"74038ed5-c727-40ae-bc0f-d779aa47613d\") " pod="openshift-marketplace/certified-operators-jvvwl" Nov 26 01:24:55 crc kubenswrapper[4766]: I1126 01:24:55.045382 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jvvwl" Nov 26 01:24:55 crc kubenswrapper[4766]: I1126 01:24:55.452717 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vmzkg" event={"ID":"02c511b4-bb27-45a2-bc48-cb21f30158af","Type":"ContainerStarted","Data":"d56b3000f57f26facc5914736d48477f7a815124247f352d023348516eeb625c"} Nov 26 01:24:55 crc kubenswrapper[4766]: I1126 01:24:55.483351 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vmzkg" podStartSLOduration=1.872554954 podStartE2EDuration="4.483334411s" podCreationTimestamp="2025-11-26 01:24:51 +0000 UTC" firstStartedPulling="2025-11-26 01:24:52.398829478 +0000 UTC m=+3673.247599908" lastFinishedPulling="2025-11-26 01:24:55.009608935 +0000 UTC m=+3675.858379365" observedRunningTime="2025-11-26 01:24:55.468901611 +0000 UTC m=+3676.317672041" watchObservedRunningTime="2025-11-26 01:24:55.483334411 +0000 UTC m=+3676.332104841" Nov 26 01:24:55 crc kubenswrapper[4766]: I1126 01:24:55.613054 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jvvwl"] Nov 26 01:24:56 crc kubenswrapper[4766]: I1126 01:24:56.480813 4766 generic.go:334] "Generic (PLEG): container finished" podID="74038ed5-c727-40ae-bc0f-d779aa47613d" containerID="eb77f0605efe22a6812983b1af16499a35d4809b1273a1f6d5794feeb10efec2" exitCode=0 Nov 26 01:24:56 crc kubenswrapper[4766]: I1126 01:24:56.481308 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jvvwl" event={"ID":"74038ed5-c727-40ae-bc0f-d779aa47613d","Type":"ContainerDied","Data":"eb77f0605efe22a6812983b1af16499a35d4809b1273a1f6d5794feeb10efec2"} Nov 26 01:24:56 crc kubenswrapper[4766]: I1126 01:24:56.481350 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jvvwl" event={"ID":"74038ed5-c727-40ae-bc0f-d779aa47613d","Type":"ContainerStarted","Data":"f6f0256073ac2b5d55b719046700c427630ee1ef7009ead5ee0c35afad708ef6"} Nov 26 01:24:58 crc kubenswrapper[4766]: I1126 01:24:58.505021 4766 generic.go:334] "Generic (PLEG): container finished" podID="3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf" containerID="ddf93ecfd0a2e8c8fa5d5ecb62d0a03ab918594b140b199b6f4d327dadc92a90" exitCode=0 Nov 26 01:24:58 crc kubenswrapper[4766]: I1126 01:24:58.505136 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gd7g9" event={"ID":"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf","Type":"ContainerDied","Data":"ddf93ecfd0a2e8c8fa5d5ecb62d0a03ab918594b140b199b6f4d327dadc92a90"} Nov 26 01:25:01 crc kubenswrapper[4766]: I1126 01:25:01.478564 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vmzkg" Nov 26 01:25:01 crc kubenswrapper[4766]: I1126 01:25:01.480145 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vmzkg" Nov 26 01:25:01 crc kubenswrapper[4766]: I1126 01:25:01.546358 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jvvwl" event={"ID":"74038ed5-c727-40ae-bc0f-d779aa47613d","Type":"ContainerStarted","Data":"ccd3cac5411b5bfd0f71b698c514508dd1bbac026c04d7099eea3623bd2d2e35"} Nov 26 01:25:01 crc kubenswrapper[4766]: I1126 01:25:01.555201 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gd7g9" event={"ID":"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf","Type":"ContainerStarted","Data":"21a4e1449393042730a65f9b62cfba2481d6dfab6a2dd2367d08bce002f447f7"} Nov 26 01:25:01 crc kubenswrapper[4766]: I1126 01:25:01.559112 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vmzkg" Nov 26 01:25:01 crc kubenswrapper[4766]: I1126 01:25:01.612579 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gd7g9" podStartSLOduration=2.093204359 podStartE2EDuration="10.612560143s" podCreationTimestamp="2025-11-26 01:24:51 +0000 UTC" firstStartedPulling="2025-11-26 01:24:52.400266883 +0000 UTC m=+3673.249037313" lastFinishedPulling="2025-11-26 01:25:00.919622657 +0000 UTC m=+3681.768393097" observedRunningTime="2025-11-26 01:25:01.603909633 +0000 UTC m=+3682.452680093" watchObservedRunningTime="2025-11-26 01:25:01.612560143 +0000 UTC m=+3682.461330573" Nov 26 01:25:01 crc kubenswrapper[4766]: I1126 01:25:01.619097 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vmzkg" Nov 26 01:25:01 crc kubenswrapper[4766]: I1126 01:25:01.646177 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gd7g9" Nov 26 01:25:01 crc kubenswrapper[4766]: I1126 01:25:01.646227 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gd7g9" Nov 26 01:25:02 crc kubenswrapper[4766]: I1126 01:25:02.521600 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vmzkg"] Nov 26 01:25:02 crc kubenswrapper[4766]: I1126 01:25:02.566918 4766 generic.go:334] "Generic (PLEG): container finished" podID="74038ed5-c727-40ae-bc0f-d779aa47613d" containerID="ccd3cac5411b5bfd0f71b698c514508dd1bbac026c04d7099eea3623bd2d2e35" exitCode=0 Nov 26 01:25:02 crc kubenswrapper[4766]: I1126 01:25:02.567249 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jvvwl" event={"ID":"74038ed5-c727-40ae-bc0f-d779aa47613d","Type":"ContainerDied","Data":"ccd3cac5411b5bfd0f71b698c514508dd1bbac026c04d7099eea3623bd2d2e35"} Nov 26 01:25:02 crc kubenswrapper[4766]: I1126 01:25:02.567294 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jvvwl" event={"ID":"74038ed5-c727-40ae-bc0f-d779aa47613d","Type":"ContainerStarted","Data":"7c971439032e7ee1c62ade5e3fa515fdd3c9e47ce1abf8a236a34c7bc3076622"} Nov 26 01:25:02 crc kubenswrapper[4766]: I1126 01:25:02.610596 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jvvwl" podStartSLOduration=3.165397772 podStartE2EDuration="8.610577493s" podCreationTimestamp="2025-11-26 01:24:54 +0000 UTC" firstStartedPulling="2025-11-26 01:24:56.487226424 +0000 UTC m=+3677.335996854" lastFinishedPulling="2025-11-26 01:25:01.932406135 +0000 UTC m=+3682.781176575" observedRunningTime="2025-11-26 01:25:02.604016613 +0000 UTC m=+3683.452787043" watchObservedRunningTime="2025-11-26 01:25:02.610577493 +0000 UTC m=+3683.459347923" Nov 26 01:25:02 crc kubenswrapper[4766]: I1126 01:25:02.694517 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gd7g9" podUID="3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf" containerName="registry-server" probeResult="failure" output=< Nov 26 01:25:02 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 01:25:02 crc kubenswrapper[4766]: > Nov 26 01:25:03 crc kubenswrapper[4766]: I1126 01:25:03.576193 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vmzkg" podUID="02c511b4-bb27-45a2-bc48-cb21f30158af" containerName="registry-server" containerID="cri-o://d56b3000f57f26facc5914736d48477f7a815124247f352d023348516eeb625c" gracePeriod=2 Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.269287 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vmzkg" Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.444108 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02c511b4-bb27-45a2-bc48-cb21f30158af-utilities\") pod \"02c511b4-bb27-45a2-bc48-cb21f30158af\" (UID: \"02c511b4-bb27-45a2-bc48-cb21f30158af\") " Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.444309 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zm9dw\" (UniqueName: \"kubernetes.io/projected/02c511b4-bb27-45a2-bc48-cb21f30158af-kube-api-access-zm9dw\") pod \"02c511b4-bb27-45a2-bc48-cb21f30158af\" (UID: \"02c511b4-bb27-45a2-bc48-cb21f30158af\") " Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.444412 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02c511b4-bb27-45a2-bc48-cb21f30158af-catalog-content\") pod \"02c511b4-bb27-45a2-bc48-cb21f30158af\" (UID: \"02c511b4-bb27-45a2-bc48-cb21f30158af\") " Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.445129 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02c511b4-bb27-45a2-bc48-cb21f30158af-utilities" (OuterVolumeSpecName: "utilities") pod "02c511b4-bb27-45a2-bc48-cb21f30158af" (UID: "02c511b4-bb27-45a2-bc48-cb21f30158af"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.446168 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02c511b4-bb27-45a2-bc48-cb21f30158af-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.454726 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02c511b4-bb27-45a2-bc48-cb21f30158af-kube-api-access-zm9dw" (OuterVolumeSpecName: "kube-api-access-zm9dw") pod "02c511b4-bb27-45a2-bc48-cb21f30158af" (UID: "02c511b4-bb27-45a2-bc48-cb21f30158af"). InnerVolumeSpecName "kube-api-access-zm9dw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.465696 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02c511b4-bb27-45a2-bc48-cb21f30158af-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "02c511b4-bb27-45a2-bc48-cb21f30158af" (UID: "02c511b4-bb27-45a2-bc48-cb21f30158af"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.549019 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02c511b4-bb27-45a2-bc48-cb21f30158af-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.549069 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zm9dw\" (UniqueName: \"kubernetes.io/projected/02c511b4-bb27-45a2-bc48-cb21f30158af-kube-api-access-zm9dw\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.592264 4766 generic.go:334] "Generic (PLEG): container finished" podID="02c511b4-bb27-45a2-bc48-cb21f30158af" containerID="d56b3000f57f26facc5914736d48477f7a815124247f352d023348516eeb625c" exitCode=0 Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.592327 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vmzkg" event={"ID":"02c511b4-bb27-45a2-bc48-cb21f30158af","Type":"ContainerDied","Data":"d56b3000f57f26facc5914736d48477f7a815124247f352d023348516eeb625c"} Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.592369 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vmzkg" event={"ID":"02c511b4-bb27-45a2-bc48-cb21f30158af","Type":"ContainerDied","Data":"91e4634373fea50582aff68d2d94f57a8625c20f4e400eccc7a96ec6c35c7e49"} Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.592393 4766 scope.go:117] "RemoveContainer" containerID="d56b3000f57f26facc5914736d48477f7a815124247f352d023348516eeb625c" Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.592401 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vmzkg" Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.635718 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vmzkg"] Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.641336 4766 scope.go:117] "RemoveContainer" containerID="ea970a16ba0ba26b59e65b37d674ff8ed0d4fa8b7aca852222e8f4d71bb8c8dc" Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.647749 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vmzkg"] Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.676267 4766 scope.go:117] "RemoveContainer" containerID="00888a4dbda175f5c4892124c37b1f571439cf4e14c4949c2f36ca84654597a4" Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.723733 4766 scope.go:117] "RemoveContainer" containerID="d56b3000f57f26facc5914736d48477f7a815124247f352d023348516eeb625c" Nov 26 01:25:04 crc kubenswrapper[4766]: E1126 01:25:04.724923 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d56b3000f57f26facc5914736d48477f7a815124247f352d023348516eeb625c\": container with ID starting with d56b3000f57f26facc5914736d48477f7a815124247f352d023348516eeb625c not found: ID does not exist" containerID="d56b3000f57f26facc5914736d48477f7a815124247f352d023348516eeb625c" Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.724971 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d56b3000f57f26facc5914736d48477f7a815124247f352d023348516eeb625c"} err="failed to get container status \"d56b3000f57f26facc5914736d48477f7a815124247f352d023348516eeb625c\": rpc error: code = NotFound desc = could not find container \"d56b3000f57f26facc5914736d48477f7a815124247f352d023348516eeb625c\": container with ID starting with d56b3000f57f26facc5914736d48477f7a815124247f352d023348516eeb625c not found: ID does not exist" Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.725022 4766 scope.go:117] "RemoveContainer" containerID="ea970a16ba0ba26b59e65b37d674ff8ed0d4fa8b7aca852222e8f4d71bb8c8dc" Nov 26 01:25:04 crc kubenswrapper[4766]: E1126 01:25:04.725869 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea970a16ba0ba26b59e65b37d674ff8ed0d4fa8b7aca852222e8f4d71bb8c8dc\": container with ID starting with ea970a16ba0ba26b59e65b37d674ff8ed0d4fa8b7aca852222e8f4d71bb8c8dc not found: ID does not exist" containerID="ea970a16ba0ba26b59e65b37d674ff8ed0d4fa8b7aca852222e8f4d71bb8c8dc" Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.725921 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea970a16ba0ba26b59e65b37d674ff8ed0d4fa8b7aca852222e8f4d71bb8c8dc"} err="failed to get container status \"ea970a16ba0ba26b59e65b37d674ff8ed0d4fa8b7aca852222e8f4d71bb8c8dc\": rpc error: code = NotFound desc = could not find container \"ea970a16ba0ba26b59e65b37d674ff8ed0d4fa8b7aca852222e8f4d71bb8c8dc\": container with ID starting with ea970a16ba0ba26b59e65b37d674ff8ed0d4fa8b7aca852222e8f4d71bb8c8dc not found: ID does not exist" Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.725935 4766 scope.go:117] "RemoveContainer" containerID="00888a4dbda175f5c4892124c37b1f571439cf4e14c4949c2f36ca84654597a4" Nov 26 01:25:04 crc kubenswrapper[4766]: E1126 01:25:04.726155 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00888a4dbda175f5c4892124c37b1f571439cf4e14c4949c2f36ca84654597a4\": container with ID starting with 00888a4dbda175f5c4892124c37b1f571439cf4e14c4949c2f36ca84654597a4 not found: ID does not exist" containerID="00888a4dbda175f5c4892124c37b1f571439cf4e14c4949c2f36ca84654597a4" Nov 26 01:25:04 crc kubenswrapper[4766]: I1126 01:25:04.726176 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00888a4dbda175f5c4892124c37b1f571439cf4e14c4949c2f36ca84654597a4"} err="failed to get container status \"00888a4dbda175f5c4892124c37b1f571439cf4e14c4949c2f36ca84654597a4\": rpc error: code = NotFound desc = could not find container \"00888a4dbda175f5c4892124c37b1f571439cf4e14c4949c2f36ca84654597a4\": container with ID starting with 00888a4dbda175f5c4892124c37b1f571439cf4e14c4949c2f36ca84654597a4 not found: ID does not exist" Nov 26 01:25:05 crc kubenswrapper[4766]: I1126 01:25:05.046549 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jvvwl" Nov 26 01:25:05 crc kubenswrapper[4766]: I1126 01:25:05.046646 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jvvwl" Nov 26 01:25:05 crc kubenswrapper[4766]: I1126 01:25:05.121440 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jvvwl" Nov 26 01:25:05 crc kubenswrapper[4766]: I1126 01:25:05.827427 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:25:05 crc kubenswrapper[4766]: E1126 01:25:05.828417 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:25:05 crc kubenswrapper[4766]: I1126 01:25:05.851238 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02c511b4-bb27-45a2-bc48-cb21f30158af" path="/var/lib/kubelet/pods/02c511b4-bb27-45a2-bc48-cb21f30158af/volumes" Nov 26 01:25:08 crc kubenswrapper[4766]: I1126 01:25:08.635887 4766 generic.go:334] "Generic (PLEG): container finished" podID="7509ec1a-b269-4bbe-bfca-0dcb052360ea" containerID="e675dd4dd6095979272355d89c719f06b37233203371017c8360c0a378657c51" exitCode=0 Nov 26 01:25:08 crc kubenswrapper[4766]: I1126 01:25:08.636066 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" event={"ID":"7509ec1a-b269-4bbe-bfca-0dcb052360ea","Type":"ContainerDied","Data":"e675dd4dd6095979272355d89c719f06b37233203371017c8360c0a378657c51"} Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.363047 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.472961 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-libvirt-combined-ca-bundle\") pod \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.473024 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-telemetry-combined-ca-bundle\") pod \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.473050 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-nova-combined-ca-bundle\") pod \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.473142 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-ovn-default-certs-0\") pod \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.473187 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.473217 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-bootstrap-combined-ca-bundle\") pod \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.473263 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.473306 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-ovn-combined-ca-bundle\") pod \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.473361 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-repo-setup-combined-ca-bundle\") pod \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.473392 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-ceph\") pod \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.473428 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-telemetry-power-monitoring-combined-ca-bundle\") pod \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.473504 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-inventory\") pod \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.473571 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-ssh-key\") pod \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.473634 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdnt6\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-kube-api-access-qdnt6\") pod \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.473729 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.473786 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.473842 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-neutron-metadata-combined-ca-bundle\") pod \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\" (UID: \"7509ec1a-b269-4bbe-bfca-0dcb052360ea\") " Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.479860 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "7509ec1a-b269-4bbe-bfca-0dcb052360ea" (UID: "7509ec1a-b269-4bbe-bfca-0dcb052360ea"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.482771 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-ceph" (OuterVolumeSpecName: "ceph") pod "7509ec1a-b269-4bbe-bfca-0dcb052360ea" (UID: "7509ec1a-b269-4bbe-bfca-0dcb052360ea"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.483315 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "7509ec1a-b269-4bbe-bfca-0dcb052360ea" (UID: "7509ec1a-b269-4bbe-bfca-0dcb052360ea"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.484965 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "7509ec1a-b269-4bbe-bfca-0dcb052360ea" (UID: "7509ec1a-b269-4bbe-bfca-0dcb052360ea"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.485107 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "7509ec1a-b269-4bbe-bfca-0dcb052360ea" (UID: "7509ec1a-b269-4bbe-bfca-0dcb052360ea"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.485165 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "7509ec1a-b269-4bbe-bfca-0dcb052360ea" (UID: "7509ec1a-b269-4bbe-bfca-0dcb052360ea"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.485209 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "7509ec1a-b269-4bbe-bfca-0dcb052360ea" (UID: "7509ec1a-b269-4bbe-bfca-0dcb052360ea"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.485238 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "7509ec1a-b269-4bbe-bfca-0dcb052360ea" (UID: "7509ec1a-b269-4bbe-bfca-0dcb052360ea"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.485854 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "7509ec1a-b269-4bbe-bfca-0dcb052360ea" (UID: "7509ec1a-b269-4bbe-bfca-0dcb052360ea"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.486492 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-kube-api-access-qdnt6" (OuterVolumeSpecName: "kube-api-access-qdnt6") pod "7509ec1a-b269-4bbe-bfca-0dcb052360ea" (UID: "7509ec1a-b269-4bbe-bfca-0dcb052360ea"). InnerVolumeSpecName "kube-api-access-qdnt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.487345 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "7509ec1a-b269-4bbe-bfca-0dcb052360ea" (UID: "7509ec1a-b269-4bbe-bfca-0dcb052360ea"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.487465 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0") pod "7509ec1a-b269-4bbe-bfca-0dcb052360ea" (UID: "7509ec1a-b269-4bbe-bfca-0dcb052360ea"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.488003 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "7509ec1a-b269-4bbe-bfca-0dcb052360ea" (UID: "7509ec1a-b269-4bbe-bfca-0dcb052360ea"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.489556 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "7509ec1a-b269-4bbe-bfca-0dcb052360ea" (UID: "7509ec1a-b269-4bbe-bfca-0dcb052360ea"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.492711 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "7509ec1a-b269-4bbe-bfca-0dcb052360ea" (UID: "7509ec1a-b269-4bbe-bfca-0dcb052360ea"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.527814 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7509ec1a-b269-4bbe-bfca-0dcb052360ea" (UID: "7509ec1a-b269-4bbe-bfca-0dcb052360ea"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.533820 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-inventory" (OuterVolumeSpecName: "inventory") pod "7509ec1a-b269-4bbe-bfca-0dcb052360ea" (UID: "7509ec1a-b269-4bbe-bfca-0dcb052360ea"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.576736 4766 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.576779 4766 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.576796 4766 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.576810 4766 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.576824 4766 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.576835 4766 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.576847 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.576860 4766 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.576875 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.576888 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.576901 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdnt6\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-kube-api-access-qdnt6\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.576917 4766 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.576933 4766 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7509ec1a-b269-4bbe-bfca-0dcb052360ea-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.576948 4766 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.576960 4766 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.576971 4766 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.576982 4766 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7509ec1a-b269-4bbe-bfca-0dcb052360ea-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.661973 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" event={"ID":"7509ec1a-b269-4bbe-bfca-0dcb052360ea","Type":"ContainerDied","Data":"1dae38851335968cdf8dfa100555d192f052400bb4382d4cb2e732ae2a9504e4"} Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.662018 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1dae38851335968cdf8dfa100555d192f052400bb4382d4cb2e732ae2a9504e4" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.662038 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-l279n" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.776908 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl"] Nov 26 01:25:10 crc kubenswrapper[4766]: E1126 01:25:10.777363 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7509ec1a-b269-4bbe-bfca-0dcb052360ea" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.777383 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="7509ec1a-b269-4bbe-bfca-0dcb052360ea" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 26 01:25:10 crc kubenswrapper[4766]: E1126 01:25:10.777408 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02c511b4-bb27-45a2-bc48-cb21f30158af" containerName="registry-server" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.777418 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="02c511b4-bb27-45a2-bc48-cb21f30158af" containerName="registry-server" Nov 26 01:25:10 crc kubenswrapper[4766]: E1126 01:25:10.777467 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02c511b4-bb27-45a2-bc48-cb21f30158af" containerName="extract-utilities" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.777475 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="02c511b4-bb27-45a2-bc48-cb21f30158af" containerName="extract-utilities" Nov 26 01:25:10 crc kubenswrapper[4766]: E1126 01:25:10.777489 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02c511b4-bb27-45a2-bc48-cb21f30158af" containerName="extract-content" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.777496 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="02c511b4-bb27-45a2-bc48-cb21f30158af" containerName="extract-content" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.777756 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="02c511b4-bb27-45a2-bc48-cb21f30158af" containerName="registry-server" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.777779 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="7509ec1a-b269-4bbe-bfca-0dcb052360ea" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.778584 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.782392 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.784392 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.786224 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.787069 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.789475 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.809878 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl"] Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.882262 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g6p9\" (UniqueName: \"kubernetes.io/projected/91756acc-c3e0-471b-b79d-4eb1cf2a80df-kube-api-access-6g6p9\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl\" (UID: \"91756acc-c3e0-471b-b79d-4eb1cf2a80df\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.882322 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/91756acc-c3e0-471b-b79d-4eb1cf2a80df-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl\" (UID: \"91756acc-c3e0-471b-b79d-4eb1cf2a80df\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.882357 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91756acc-c3e0-471b-b79d-4eb1cf2a80df-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl\" (UID: \"91756acc-c3e0-471b-b79d-4eb1cf2a80df\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.882625 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91756acc-c3e0-471b-b79d-4eb1cf2a80df-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl\" (UID: \"91756acc-c3e0-471b-b79d-4eb1cf2a80df\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.984735 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91756acc-c3e0-471b-b79d-4eb1cf2a80df-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl\" (UID: \"91756acc-c3e0-471b-b79d-4eb1cf2a80df\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.984962 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g6p9\" (UniqueName: \"kubernetes.io/projected/91756acc-c3e0-471b-b79d-4eb1cf2a80df-kube-api-access-6g6p9\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl\" (UID: \"91756acc-c3e0-471b-b79d-4eb1cf2a80df\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.985040 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/91756acc-c3e0-471b-b79d-4eb1cf2a80df-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl\" (UID: \"91756acc-c3e0-471b-b79d-4eb1cf2a80df\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.985095 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91756acc-c3e0-471b-b79d-4eb1cf2a80df-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl\" (UID: \"91756acc-c3e0-471b-b79d-4eb1cf2a80df\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.990073 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91756acc-c3e0-471b-b79d-4eb1cf2a80df-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl\" (UID: \"91756acc-c3e0-471b-b79d-4eb1cf2a80df\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.990692 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/91756acc-c3e0-471b-b79d-4eb1cf2a80df-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl\" (UID: \"91756acc-c3e0-471b-b79d-4eb1cf2a80df\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl" Nov 26 01:25:10 crc kubenswrapper[4766]: I1126 01:25:10.991203 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91756acc-c3e0-471b-b79d-4eb1cf2a80df-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl\" (UID: \"91756acc-c3e0-471b-b79d-4eb1cf2a80df\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl" Nov 26 01:25:11 crc kubenswrapper[4766]: I1126 01:25:11.003754 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g6p9\" (UniqueName: \"kubernetes.io/projected/91756acc-c3e0-471b-b79d-4eb1cf2a80df-kube-api-access-6g6p9\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl\" (UID: \"91756acc-c3e0-471b-b79d-4eb1cf2a80df\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl" Nov 26 01:25:11 crc kubenswrapper[4766]: I1126 01:25:11.104032 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl" Nov 26 01:25:11 crc kubenswrapper[4766]: I1126 01:25:11.719644 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl"] Nov 26 01:25:12 crc kubenswrapper[4766]: I1126 01:25:12.685100 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl" event={"ID":"91756acc-c3e0-471b-b79d-4eb1cf2a80df","Type":"ContainerStarted","Data":"fca6c85e788bd142804477368af6f510598c9427a0da3512c2dd6abc1b6ca581"} Nov 26 01:25:12 crc kubenswrapper[4766]: I1126 01:25:12.685963 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl" event={"ID":"91756acc-c3e0-471b-b79d-4eb1cf2a80df","Type":"ContainerStarted","Data":"983649c4b1563d3a38fbd8f4908393d9f6b5d3a609e183afd7201c0a1dd0d530"} Nov 26 01:25:12 crc kubenswrapper[4766]: I1126 01:25:12.700288 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gd7g9" podUID="3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf" containerName="registry-server" probeResult="failure" output=< Nov 26 01:25:12 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 01:25:12 crc kubenswrapper[4766]: > Nov 26 01:25:12 crc kubenswrapper[4766]: I1126 01:25:12.711648 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl" podStartSLOduration=2.227208185 podStartE2EDuration="2.711620021s" podCreationTimestamp="2025-11-26 01:25:10 +0000 UTC" firstStartedPulling="2025-11-26 01:25:11.720866587 +0000 UTC m=+3692.569637027" lastFinishedPulling="2025-11-26 01:25:12.205278393 +0000 UTC m=+3693.054048863" observedRunningTime="2025-11-26 01:25:12.703204147 +0000 UTC m=+3693.551974587" watchObservedRunningTime="2025-11-26 01:25:12.711620021 +0000 UTC m=+3693.560390491" Nov 26 01:25:15 crc kubenswrapper[4766]: I1126 01:25:15.134603 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jvvwl" Nov 26 01:25:15 crc kubenswrapper[4766]: I1126 01:25:15.249969 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jvvwl"] Nov 26 01:25:15 crc kubenswrapper[4766]: I1126 01:25:15.371085 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qrsgb"] Nov 26 01:25:15 crc kubenswrapper[4766]: I1126 01:25:15.371635 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qrsgb" podUID="d93cf31b-be67-48b1-a248-ba0df8870fdd" containerName="registry-server" containerID="cri-o://92fd35cfdb6e079155ae81fd73f3eafb0c01a7c95934425a72564d76ba3e085b" gracePeriod=2 Nov 26 01:25:15 crc kubenswrapper[4766]: I1126 01:25:15.738146 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qrsgb" event={"ID":"d93cf31b-be67-48b1-a248-ba0df8870fdd","Type":"ContainerDied","Data":"92fd35cfdb6e079155ae81fd73f3eafb0c01a7c95934425a72564d76ba3e085b"} Nov 26 01:25:15 crc kubenswrapper[4766]: I1126 01:25:15.739214 4766 generic.go:334] "Generic (PLEG): container finished" podID="d93cf31b-be67-48b1-a248-ba0df8870fdd" containerID="92fd35cfdb6e079155ae81fd73f3eafb0c01a7c95934425a72564d76ba3e085b" exitCode=0 Nov 26 01:25:15 crc kubenswrapper[4766]: I1126 01:25:15.907896 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qrsgb" Nov 26 01:25:15 crc kubenswrapper[4766]: I1126 01:25:15.989767 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gshs6\" (UniqueName: \"kubernetes.io/projected/d93cf31b-be67-48b1-a248-ba0df8870fdd-kube-api-access-gshs6\") pod \"d93cf31b-be67-48b1-a248-ba0df8870fdd\" (UID: \"d93cf31b-be67-48b1-a248-ba0df8870fdd\") " Nov 26 01:25:15 crc kubenswrapper[4766]: I1126 01:25:15.990189 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d93cf31b-be67-48b1-a248-ba0df8870fdd-utilities\") pod \"d93cf31b-be67-48b1-a248-ba0df8870fdd\" (UID: \"d93cf31b-be67-48b1-a248-ba0df8870fdd\") " Nov 26 01:25:15 crc kubenswrapper[4766]: I1126 01:25:15.990310 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d93cf31b-be67-48b1-a248-ba0df8870fdd-catalog-content\") pod \"d93cf31b-be67-48b1-a248-ba0df8870fdd\" (UID: \"d93cf31b-be67-48b1-a248-ba0df8870fdd\") " Nov 26 01:25:15 crc kubenswrapper[4766]: I1126 01:25:15.992759 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d93cf31b-be67-48b1-a248-ba0df8870fdd-utilities" (OuterVolumeSpecName: "utilities") pod "d93cf31b-be67-48b1-a248-ba0df8870fdd" (UID: "d93cf31b-be67-48b1-a248-ba0df8870fdd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:25:16 crc kubenswrapper[4766]: I1126 01:25:16.013561 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d93cf31b-be67-48b1-a248-ba0df8870fdd-kube-api-access-gshs6" (OuterVolumeSpecName: "kube-api-access-gshs6") pod "d93cf31b-be67-48b1-a248-ba0df8870fdd" (UID: "d93cf31b-be67-48b1-a248-ba0df8870fdd"). InnerVolumeSpecName "kube-api-access-gshs6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:25:16 crc kubenswrapper[4766]: I1126 01:25:16.058367 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d93cf31b-be67-48b1-a248-ba0df8870fdd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d93cf31b-be67-48b1-a248-ba0df8870fdd" (UID: "d93cf31b-be67-48b1-a248-ba0df8870fdd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:25:16 crc kubenswrapper[4766]: I1126 01:25:16.092796 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gshs6\" (UniqueName: \"kubernetes.io/projected/d93cf31b-be67-48b1-a248-ba0df8870fdd-kube-api-access-gshs6\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:16 crc kubenswrapper[4766]: I1126 01:25:16.092826 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d93cf31b-be67-48b1-a248-ba0df8870fdd-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:16 crc kubenswrapper[4766]: I1126 01:25:16.092836 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d93cf31b-be67-48b1-a248-ba0df8870fdd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:16 crc kubenswrapper[4766]: I1126 01:25:16.753185 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qrsgb" event={"ID":"d93cf31b-be67-48b1-a248-ba0df8870fdd","Type":"ContainerDied","Data":"23f16556b38d18934fd8b69ccffbbdc7cd65a30a96b5a7abedfa6b7f97e7376e"} Nov 26 01:25:16 crc kubenswrapper[4766]: I1126 01:25:16.753233 4766 scope.go:117] "RemoveContainer" containerID="92fd35cfdb6e079155ae81fd73f3eafb0c01a7c95934425a72564d76ba3e085b" Nov 26 01:25:16 crc kubenswrapper[4766]: I1126 01:25:16.753378 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qrsgb" Nov 26 01:25:16 crc kubenswrapper[4766]: I1126 01:25:16.793775 4766 scope.go:117] "RemoveContainer" containerID="2724526881ad660ecc4eac36ae865e4ad120d04d2d05f2e34a339938afcc6da0" Nov 26 01:25:16 crc kubenswrapper[4766]: I1126 01:25:16.795875 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qrsgb"] Nov 26 01:25:16 crc kubenswrapper[4766]: I1126 01:25:16.810355 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qrsgb"] Nov 26 01:25:16 crc kubenswrapper[4766]: I1126 01:25:16.822190 4766 scope.go:117] "RemoveContainer" containerID="46d3ce5e9c1e6baa80f11a2d9811b7c9b65d645dfac64cac84c89b7956313321" Nov 26 01:25:17 crc kubenswrapper[4766]: I1126 01:25:17.828530 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:25:17 crc kubenswrapper[4766]: E1126 01:25:17.828993 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:25:17 crc kubenswrapper[4766]: I1126 01:25:17.842236 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d93cf31b-be67-48b1-a248-ba0df8870fdd" path="/var/lib/kubelet/pods/d93cf31b-be67-48b1-a248-ba0df8870fdd/volumes" Nov 26 01:25:20 crc kubenswrapper[4766]: I1126 01:25:20.799406 4766 generic.go:334] "Generic (PLEG): container finished" podID="91756acc-c3e0-471b-b79d-4eb1cf2a80df" containerID="fca6c85e788bd142804477368af6f510598c9427a0da3512c2dd6abc1b6ca581" exitCode=0 Nov 26 01:25:20 crc kubenswrapper[4766]: I1126 01:25:20.799513 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl" event={"ID":"91756acc-c3e0-471b-b79d-4eb1cf2a80df","Type":"ContainerDied","Data":"fca6c85e788bd142804477368af6f510598c9427a0da3512c2dd6abc1b6ca581"} Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.249822 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.321271 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6p9\" (UniqueName: \"kubernetes.io/projected/91756acc-c3e0-471b-b79d-4eb1cf2a80df-kube-api-access-6g6p9\") pod \"91756acc-c3e0-471b-b79d-4eb1cf2a80df\" (UID: \"91756acc-c3e0-471b-b79d-4eb1cf2a80df\") " Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.321519 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91756acc-c3e0-471b-b79d-4eb1cf2a80df-ssh-key\") pod \"91756acc-c3e0-471b-b79d-4eb1cf2a80df\" (UID: \"91756acc-c3e0-471b-b79d-4eb1cf2a80df\") " Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.321545 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/91756acc-c3e0-471b-b79d-4eb1cf2a80df-ceph\") pod \"91756acc-c3e0-471b-b79d-4eb1cf2a80df\" (UID: \"91756acc-c3e0-471b-b79d-4eb1cf2a80df\") " Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.321567 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91756acc-c3e0-471b-b79d-4eb1cf2a80df-inventory\") pod \"91756acc-c3e0-471b-b79d-4eb1cf2a80df\" (UID: \"91756acc-c3e0-471b-b79d-4eb1cf2a80df\") " Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.328581 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91756acc-c3e0-471b-b79d-4eb1cf2a80df-kube-api-access-6g6p9" (OuterVolumeSpecName: "kube-api-access-6g6p9") pod "91756acc-c3e0-471b-b79d-4eb1cf2a80df" (UID: "91756acc-c3e0-471b-b79d-4eb1cf2a80df"). InnerVolumeSpecName "kube-api-access-6g6p9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.330102 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91756acc-c3e0-471b-b79d-4eb1cf2a80df-ceph" (OuterVolumeSpecName: "ceph") pod "91756acc-c3e0-471b-b79d-4eb1cf2a80df" (UID: "91756acc-c3e0-471b-b79d-4eb1cf2a80df"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.350742 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91756acc-c3e0-471b-b79d-4eb1cf2a80df-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "91756acc-c3e0-471b-b79d-4eb1cf2a80df" (UID: "91756acc-c3e0-471b-b79d-4eb1cf2a80df"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.350905 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91756acc-c3e0-471b-b79d-4eb1cf2a80df-inventory" (OuterVolumeSpecName: "inventory") pod "91756acc-c3e0-471b-b79d-4eb1cf2a80df" (UID: "91756acc-c3e0-471b-b79d-4eb1cf2a80df"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.425172 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6p9\" (UniqueName: \"kubernetes.io/projected/91756acc-c3e0-471b-b79d-4eb1cf2a80df-kube-api-access-6g6p9\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.425213 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91756acc-c3e0-471b-b79d-4eb1cf2a80df-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.425224 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/91756acc-c3e0-471b-b79d-4eb1cf2a80df-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.425233 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91756acc-c3e0-471b-b79d-4eb1cf2a80df-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.701293 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gd7g9" podUID="3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf" containerName="registry-server" probeResult="failure" output=< Nov 26 01:25:22 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 01:25:22 crc kubenswrapper[4766]: > Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.822526 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl" event={"ID":"91756acc-c3e0-471b-b79d-4eb1cf2a80df","Type":"ContainerDied","Data":"983649c4b1563d3a38fbd8f4908393d9f6b5d3a609e183afd7201c0a1dd0d530"} Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.822594 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="983649c4b1563d3a38fbd8f4908393d9f6b5d3a609e183afd7201c0a1dd0d530" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.822606 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.894549 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z"] Nov 26 01:25:22 crc kubenswrapper[4766]: E1126 01:25:22.895012 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d93cf31b-be67-48b1-a248-ba0df8870fdd" containerName="extract-content" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.895032 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d93cf31b-be67-48b1-a248-ba0df8870fdd" containerName="extract-content" Nov 26 01:25:22 crc kubenswrapper[4766]: E1126 01:25:22.895044 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91756acc-c3e0-471b-b79d-4eb1cf2a80df" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.895053 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="91756acc-c3e0-471b-b79d-4eb1cf2a80df" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Nov 26 01:25:22 crc kubenswrapper[4766]: E1126 01:25:22.895068 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d93cf31b-be67-48b1-a248-ba0df8870fdd" containerName="extract-utilities" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.895073 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d93cf31b-be67-48b1-a248-ba0df8870fdd" containerName="extract-utilities" Nov 26 01:25:22 crc kubenswrapper[4766]: E1126 01:25:22.895083 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d93cf31b-be67-48b1-a248-ba0df8870fdd" containerName="registry-server" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.895088 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d93cf31b-be67-48b1-a248-ba0df8870fdd" containerName="registry-server" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.895301 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="91756acc-c3e0-471b-b79d-4eb1cf2a80df" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.895328 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="d93cf31b-be67-48b1-a248-ba0df8870fdd" containerName="registry-server" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.896001 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.899793 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.899868 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.900242 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.900676 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.901896 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.903106 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:25:22 crc kubenswrapper[4766]: I1126 01:25:22.910315 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z"] Nov 26 01:25:23 crc kubenswrapper[4766]: I1126 01:25:23.038016 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8xf9z\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:25:23 crc kubenswrapper[4766]: I1126 01:25:23.038325 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8xf9z\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:25:23 crc kubenswrapper[4766]: I1126 01:25:23.038443 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8xf9z\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:25:23 crc kubenswrapper[4766]: I1126 01:25:23.038544 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8xf9z\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:25:23 crc kubenswrapper[4766]: I1126 01:25:23.038638 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8xf9z\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:25:23 crc kubenswrapper[4766]: I1126 01:25:23.038800 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8qtr\" (UniqueName: \"kubernetes.io/projected/45de00ed-b796-4ac1-8e38-ee601eab6dc0-kube-api-access-z8qtr\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8xf9z\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:25:23 crc kubenswrapper[4766]: I1126 01:25:23.141029 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8xf9z\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:25:23 crc kubenswrapper[4766]: I1126 01:25:23.141321 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8xf9z\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:25:23 crc kubenswrapper[4766]: I1126 01:25:23.141453 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8xf9z\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:25:23 crc kubenswrapper[4766]: I1126 01:25:23.141560 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8xf9z\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:25:23 crc kubenswrapper[4766]: I1126 01:25:23.141677 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8xf9z\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:25:23 crc kubenswrapper[4766]: I1126 01:25:23.141867 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8qtr\" (UniqueName: \"kubernetes.io/projected/45de00ed-b796-4ac1-8e38-ee601eab6dc0-kube-api-access-z8qtr\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8xf9z\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:25:23 crc kubenswrapper[4766]: I1126 01:25:23.142877 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8xf9z\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:25:23 crc kubenswrapper[4766]: I1126 01:25:23.145247 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8xf9z\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:25:23 crc kubenswrapper[4766]: I1126 01:25:23.154141 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8xf9z\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:25:23 crc kubenswrapper[4766]: I1126 01:25:23.154371 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8xf9z\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:25:23 crc kubenswrapper[4766]: I1126 01:25:23.154532 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8xf9z\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:25:23 crc kubenswrapper[4766]: I1126 01:25:23.159941 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8qtr\" (UniqueName: \"kubernetes.io/projected/45de00ed-b796-4ac1-8e38-ee601eab6dc0-kube-api-access-z8qtr\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8xf9z\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:25:23 crc kubenswrapper[4766]: I1126 01:25:23.215259 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:25:23 crc kubenswrapper[4766]: I1126 01:25:23.881757 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z"] Nov 26 01:25:24 crc kubenswrapper[4766]: I1126 01:25:24.865224 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" event={"ID":"45de00ed-b796-4ac1-8e38-ee601eab6dc0","Type":"ContainerStarted","Data":"0c8d8012b1c24fcdf77e152346269a91a9151e1eded5d80bc5df365f95b23d22"} Nov 26 01:25:24 crc kubenswrapper[4766]: I1126 01:25:24.865737 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" event={"ID":"45de00ed-b796-4ac1-8e38-ee601eab6dc0","Type":"ContainerStarted","Data":"40cf7de14a6dd4406207b23590f7458197f30e16ffc28bb2588f34bd2d9edaa7"} Nov 26 01:25:24 crc kubenswrapper[4766]: I1126 01:25:24.898674 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" podStartSLOduration=2.458817747 podStartE2EDuration="2.898641921s" podCreationTimestamp="2025-11-26 01:25:22 +0000 UTC" firstStartedPulling="2025-11-26 01:25:23.906413682 +0000 UTC m=+3704.755184112" lastFinishedPulling="2025-11-26 01:25:24.346237856 +0000 UTC m=+3705.195008286" observedRunningTime="2025-11-26 01:25:24.896131281 +0000 UTC m=+3705.744901711" watchObservedRunningTime="2025-11-26 01:25:24.898641921 +0000 UTC m=+3705.747412351" Nov 26 01:25:29 crc kubenswrapper[4766]: I1126 01:25:29.833149 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:25:29 crc kubenswrapper[4766]: E1126 01:25:29.834010 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:25:31 crc kubenswrapper[4766]: I1126 01:25:31.722385 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gd7g9" Nov 26 01:25:31 crc kubenswrapper[4766]: I1126 01:25:31.791819 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gd7g9" Nov 26 01:25:31 crc kubenswrapper[4766]: I1126 01:25:31.969811 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gd7g9"] Nov 26 01:25:33 crc kubenswrapper[4766]: I1126 01:25:33.005349 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gd7g9" podUID="3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf" containerName="registry-server" containerID="cri-o://21a4e1449393042730a65f9b62cfba2481d6dfab6a2dd2367d08bce002f447f7" gracePeriod=2 Nov 26 01:25:33 crc kubenswrapper[4766]: I1126 01:25:33.520074 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gd7g9" Nov 26 01:25:33 crc kubenswrapper[4766]: I1126 01:25:33.582280 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf-catalog-content\") pod \"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf\" (UID: \"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf\") " Nov 26 01:25:33 crc kubenswrapper[4766]: I1126 01:25:33.582458 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf-utilities\") pod \"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf\" (UID: \"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf\") " Nov 26 01:25:33 crc kubenswrapper[4766]: I1126 01:25:33.582678 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hj92c\" (UniqueName: \"kubernetes.io/projected/3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf-kube-api-access-hj92c\") pod \"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf\" (UID: \"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf\") " Nov 26 01:25:33 crc kubenswrapper[4766]: I1126 01:25:33.583361 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf-utilities" (OuterVolumeSpecName: "utilities") pod "3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf" (UID: "3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:25:33 crc kubenswrapper[4766]: I1126 01:25:33.593811 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf-kube-api-access-hj92c" (OuterVolumeSpecName: "kube-api-access-hj92c") pod "3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf" (UID: "3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf"). InnerVolumeSpecName "kube-api-access-hj92c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:25:33 crc kubenswrapper[4766]: I1126 01:25:33.669996 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf" (UID: "3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:25:33 crc kubenswrapper[4766]: I1126 01:25:33.684843 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:33 crc kubenswrapper[4766]: I1126 01:25:33.684890 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hj92c\" (UniqueName: \"kubernetes.io/projected/3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf-kube-api-access-hj92c\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:33 crc kubenswrapper[4766]: I1126 01:25:33.684904 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 01:25:34 crc kubenswrapper[4766]: I1126 01:25:34.018220 4766 generic.go:334] "Generic (PLEG): container finished" podID="3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf" containerID="21a4e1449393042730a65f9b62cfba2481d6dfab6a2dd2367d08bce002f447f7" exitCode=0 Nov 26 01:25:34 crc kubenswrapper[4766]: I1126 01:25:34.018267 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gd7g9" event={"ID":"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf","Type":"ContainerDied","Data":"21a4e1449393042730a65f9b62cfba2481d6dfab6a2dd2367d08bce002f447f7"} Nov 26 01:25:34 crc kubenswrapper[4766]: I1126 01:25:34.018299 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gd7g9" event={"ID":"3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf","Type":"ContainerDied","Data":"b116cb3bf0f6ebc6b4dc2056949058fe6f8343a7069262cf0b290de3d48f73d1"} Nov 26 01:25:34 crc kubenswrapper[4766]: I1126 01:25:34.018321 4766 scope.go:117] "RemoveContainer" containerID="21a4e1449393042730a65f9b62cfba2481d6dfab6a2dd2367d08bce002f447f7" Nov 26 01:25:34 crc kubenswrapper[4766]: I1126 01:25:34.018339 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gd7g9" Nov 26 01:25:34 crc kubenswrapper[4766]: I1126 01:25:34.039700 4766 scope.go:117] "RemoveContainer" containerID="ddf93ecfd0a2e8c8fa5d5ecb62d0a03ab918594b140b199b6f4d327dadc92a90" Nov 26 01:25:34 crc kubenswrapper[4766]: I1126 01:25:34.052197 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gd7g9"] Nov 26 01:25:34 crc kubenswrapper[4766]: I1126 01:25:34.061267 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gd7g9"] Nov 26 01:25:34 crc kubenswrapper[4766]: I1126 01:25:34.068529 4766 scope.go:117] "RemoveContainer" containerID="b2af83b85a9cc51edd772ea31a1b8e6149c3f9a8c0b3ad3aa01ff0c586420ade" Nov 26 01:25:34 crc kubenswrapper[4766]: I1126 01:25:34.126941 4766 scope.go:117] "RemoveContainer" containerID="21a4e1449393042730a65f9b62cfba2481d6dfab6a2dd2367d08bce002f447f7" Nov 26 01:25:34 crc kubenswrapper[4766]: E1126 01:25:34.127496 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21a4e1449393042730a65f9b62cfba2481d6dfab6a2dd2367d08bce002f447f7\": container with ID starting with 21a4e1449393042730a65f9b62cfba2481d6dfab6a2dd2367d08bce002f447f7 not found: ID does not exist" containerID="21a4e1449393042730a65f9b62cfba2481d6dfab6a2dd2367d08bce002f447f7" Nov 26 01:25:34 crc kubenswrapper[4766]: I1126 01:25:34.127527 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21a4e1449393042730a65f9b62cfba2481d6dfab6a2dd2367d08bce002f447f7"} err="failed to get container status \"21a4e1449393042730a65f9b62cfba2481d6dfab6a2dd2367d08bce002f447f7\": rpc error: code = NotFound desc = could not find container \"21a4e1449393042730a65f9b62cfba2481d6dfab6a2dd2367d08bce002f447f7\": container with ID starting with 21a4e1449393042730a65f9b62cfba2481d6dfab6a2dd2367d08bce002f447f7 not found: ID does not exist" Nov 26 01:25:34 crc kubenswrapper[4766]: I1126 01:25:34.127544 4766 scope.go:117] "RemoveContainer" containerID="ddf93ecfd0a2e8c8fa5d5ecb62d0a03ab918594b140b199b6f4d327dadc92a90" Nov 26 01:25:34 crc kubenswrapper[4766]: E1126 01:25:34.127896 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddf93ecfd0a2e8c8fa5d5ecb62d0a03ab918594b140b199b6f4d327dadc92a90\": container with ID starting with ddf93ecfd0a2e8c8fa5d5ecb62d0a03ab918594b140b199b6f4d327dadc92a90 not found: ID does not exist" containerID="ddf93ecfd0a2e8c8fa5d5ecb62d0a03ab918594b140b199b6f4d327dadc92a90" Nov 26 01:25:34 crc kubenswrapper[4766]: I1126 01:25:34.127933 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddf93ecfd0a2e8c8fa5d5ecb62d0a03ab918594b140b199b6f4d327dadc92a90"} err="failed to get container status \"ddf93ecfd0a2e8c8fa5d5ecb62d0a03ab918594b140b199b6f4d327dadc92a90\": rpc error: code = NotFound desc = could not find container \"ddf93ecfd0a2e8c8fa5d5ecb62d0a03ab918594b140b199b6f4d327dadc92a90\": container with ID starting with ddf93ecfd0a2e8c8fa5d5ecb62d0a03ab918594b140b199b6f4d327dadc92a90 not found: ID does not exist" Nov 26 01:25:34 crc kubenswrapper[4766]: I1126 01:25:34.127956 4766 scope.go:117] "RemoveContainer" containerID="b2af83b85a9cc51edd772ea31a1b8e6149c3f9a8c0b3ad3aa01ff0c586420ade" Nov 26 01:25:34 crc kubenswrapper[4766]: E1126 01:25:34.128355 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2af83b85a9cc51edd772ea31a1b8e6149c3f9a8c0b3ad3aa01ff0c586420ade\": container with ID starting with b2af83b85a9cc51edd772ea31a1b8e6149c3f9a8c0b3ad3aa01ff0c586420ade not found: ID does not exist" containerID="b2af83b85a9cc51edd772ea31a1b8e6149c3f9a8c0b3ad3aa01ff0c586420ade" Nov 26 01:25:34 crc kubenswrapper[4766]: I1126 01:25:34.128394 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2af83b85a9cc51edd772ea31a1b8e6149c3f9a8c0b3ad3aa01ff0c586420ade"} err="failed to get container status \"b2af83b85a9cc51edd772ea31a1b8e6149c3f9a8c0b3ad3aa01ff0c586420ade\": rpc error: code = NotFound desc = could not find container \"b2af83b85a9cc51edd772ea31a1b8e6149c3f9a8c0b3ad3aa01ff0c586420ade\": container with ID starting with b2af83b85a9cc51edd772ea31a1b8e6149c3f9a8c0b3ad3aa01ff0c586420ade not found: ID does not exist" Nov 26 01:25:35 crc kubenswrapper[4766]: I1126 01:25:35.838620 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf" path="/var/lib/kubelet/pods/3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf/volumes" Nov 26 01:25:42 crc kubenswrapper[4766]: I1126 01:25:42.827192 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:25:42 crc kubenswrapper[4766]: E1126 01:25:42.828146 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:25:53 crc kubenswrapper[4766]: I1126 01:25:53.827009 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:25:53 crc kubenswrapper[4766]: E1126 01:25:53.829261 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:26:05 crc kubenswrapper[4766]: I1126 01:26:05.826948 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:26:05 crc kubenswrapper[4766]: E1126 01:26:05.828175 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:26:16 crc kubenswrapper[4766]: I1126 01:26:16.827348 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:26:16 crc kubenswrapper[4766]: E1126 01:26:16.828286 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:26:30 crc kubenswrapper[4766]: I1126 01:26:30.828592 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:26:30 crc kubenswrapper[4766]: E1126 01:26:30.830391 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:26:45 crc kubenswrapper[4766]: I1126 01:26:45.826965 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:26:45 crc kubenswrapper[4766]: E1126 01:26:45.828072 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:26:55 crc kubenswrapper[4766]: I1126 01:26:55.000924 4766 generic.go:334] "Generic (PLEG): container finished" podID="45de00ed-b796-4ac1-8e38-ee601eab6dc0" containerID="0c8d8012b1c24fcdf77e152346269a91a9151e1eded5d80bc5df365f95b23d22" exitCode=0 Nov 26 01:26:55 crc kubenswrapper[4766]: I1126 01:26:55.001069 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" event={"ID":"45de00ed-b796-4ac1-8e38-ee601eab6dc0","Type":"ContainerDied","Data":"0c8d8012b1c24fcdf77e152346269a91a9151e1eded5d80bc5df365f95b23d22"} Nov 26 01:26:56 crc kubenswrapper[4766]: I1126 01:26:56.544512 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:26:56 crc kubenswrapper[4766]: I1126 01:26:56.567243 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ceph\") pod \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " Nov 26 01:26:56 crc kubenswrapper[4766]: I1126 01:26:56.567456 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ssh-key\") pod \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " Nov 26 01:26:56 crc kubenswrapper[4766]: I1126 01:26:56.567569 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-inventory\") pod \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " Nov 26 01:26:56 crc kubenswrapper[4766]: I1126 01:26:56.567647 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ovn-combined-ca-bundle\") pod \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " Nov 26 01:26:56 crc kubenswrapper[4766]: I1126 01:26:56.567720 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8qtr\" (UniqueName: \"kubernetes.io/projected/45de00ed-b796-4ac1-8e38-ee601eab6dc0-kube-api-access-z8qtr\") pod \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " Nov 26 01:26:56 crc kubenswrapper[4766]: I1126 01:26:56.568531 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ovncontroller-config-0\") pod \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\" (UID: \"45de00ed-b796-4ac1-8e38-ee601eab6dc0\") " Nov 26 01:26:56 crc kubenswrapper[4766]: I1126 01:26:56.578708 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ceph" (OuterVolumeSpecName: "ceph") pod "45de00ed-b796-4ac1-8e38-ee601eab6dc0" (UID: "45de00ed-b796-4ac1-8e38-ee601eab6dc0"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:26:56 crc kubenswrapper[4766]: I1126 01:26:56.589825 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "45de00ed-b796-4ac1-8e38-ee601eab6dc0" (UID: "45de00ed-b796-4ac1-8e38-ee601eab6dc0"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:26:56 crc kubenswrapper[4766]: I1126 01:26:56.599339 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45de00ed-b796-4ac1-8e38-ee601eab6dc0-kube-api-access-z8qtr" (OuterVolumeSpecName: "kube-api-access-z8qtr") pod "45de00ed-b796-4ac1-8e38-ee601eab6dc0" (UID: "45de00ed-b796-4ac1-8e38-ee601eab6dc0"). InnerVolumeSpecName "kube-api-access-z8qtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:26:56 crc kubenswrapper[4766]: I1126 01:26:56.615088 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "45de00ed-b796-4ac1-8e38-ee601eab6dc0" (UID: "45de00ed-b796-4ac1-8e38-ee601eab6dc0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:26:56 crc kubenswrapper[4766]: I1126 01:26:56.620425 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-inventory" (OuterVolumeSpecName: "inventory") pod "45de00ed-b796-4ac1-8e38-ee601eab6dc0" (UID: "45de00ed-b796-4ac1-8e38-ee601eab6dc0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:26:56 crc kubenswrapper[4766]: I1126 01:26:56.628288 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "45de00ed-b796-4ac1-8e38-ee601eab6dc0" (UID: "45de00ed-b796-4ac1-8e38-ee601eab6dc0"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:26:56 crc kubenswrapper[4766]: I1126 01:26:56.671965 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:26:56 crc kubenswrapper[4766]: I1126 01:26:56.672007 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:26:56 crc kubenswrapper[4766]: I1126 01:26:56.672024 4766 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:26:56 crc kubenswrapper[4766]: I1126 01:26:56.672044 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8qtr\" (UniqueName: \"kubernetes.io/projected/45de00ed-b796-4ac1-8e38-ee601eab6dc0-kube-api-access-z8qtr\") on node \"crc\" DevicePath \"\"" Nov 26 01:26:56 crc kubenswrapper[4766]: I1126 01:26:56.672056 4766 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:26:56 crc kubenswrapper[4766]: I1126 01:26:56.672067 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/45de00ed-b796-4ac1-8e38-ee601eab6dc0-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.029532 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" event={"ID":"45de00ed-b796-4ac1-8e38-ee601eab6dc0","Type":"ContainerDied","Data":"40cf7de14a6dd4406207b23590f7458197f30e16ffc28bb2588f34bd2d9edaa7"} Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.029589 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40cf7de14a6dd4406207b23590f7458197f30e16ffc28bb2588f34bd2d9edaa7" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.029876 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8xf9z" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.180761 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h"] Nov 26 01:26:57 crc kubenswrapper[4766]: E1126 01:26:57.181349 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45de00ed-b796-4ac1-8e38-ee601eab6dc0" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.181374 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="45de00ed-b796-4ac1-8e38-ee601eab6dc0" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 26 01:26:57 crc kubenswrapper[4766]: E1126 01:26:57.181400 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf" containerName="extract-utilities" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.181411 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf" containerName="extract-utilities" Nov 26 01:26:57 crc kubenswrapper[4766]: E1126 01:26:57.181459 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf" containerName="registry-server" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.181470 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf" containerName="registry-server" Nov 26 01:26:57 crc kubenswrapper[4766]: E1126 01:26:57.181484 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf" containerName="extract-content" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.181494 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf" containerName="extract-content" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.181823 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c62bb64-7d2f-4a5a-a73e-1ac29fbc8faf" containerName="registry-server" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.181874 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="45de00ed-b796-4ac1-8e38-ee601eab6dc0" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.183048 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.185345 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.185403 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.186112 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.187091 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.187612 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.187621 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.188029 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.228692 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h"] Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.284235 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blqnv\" (UniqueName: \"kubernetes.io/projected/3efd579d-7c56-4624-a920-1e40cea0b951-kube-api-access-blqnv\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.284317 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.284474 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.284497 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.284680 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.284728 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.284799 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.386172 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.386250 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.386319 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.386386 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blqnv\" (UniqueName: \"kubernetes.io/projected/3efd579d-7c56-4624-a920-1e40cea0b951-kube-api-access-blqnv\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.386424 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.386495 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.386518 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.389942 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.390608 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.391831 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.397389 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.399277 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.400791 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.415926 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blqnv\" (UniqueName: \"kubernetes.io/projected/3efd579d-7c56-4624-a920-1e40cea0b951-kube-api-access-blqnv\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.506552 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:26:57 crc kubenswrapper[4766]: I1126 01:26:57.827086 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:26:57 crc kubenswrapper[4766]: E1126 01:26:57.827945 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:26:58 crc kubenswrapper[4766]: I1126 01:26:58.082746 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h"] Nov 26 01:26:58 crc kubenswrapper[4766]: W1126 01:26:58.085898 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3efd579d_7c56_4624_a920_1e40cea0b951.slice/crio-60552213f653766db7e33d8487a8f75b4a72f2a296e4e3c6b6dbb79e86381d3f WatchSource:0}: Error finding container 60552213f653766db7e33d8487a8f75b4a72f2a296e4e3c6b6dbb79e86381d3f: Status 404 returned error can't find the container with id 60552213f653766db7e33d8487a8f75b4a72f2a296e4e3c6b6dbb79e86381d3f Nov 26 01:26:59 crc kubenswrapper[4766]: I1126 01:26:59.050819 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" event={"ID":"3efd579d-7c56-4624-a920-1e40cea0b951","Type":"ContainerStarted","Data":"9c7463072f6807c25f87fdfb62520e0580172abc91cf31be81a85eed09f11a44"} Nov 26 01:26:59 crc kubenswrapper[4766]: I1126 01:26:59.051163 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" event={"ID":"3efd579d-7c56-4624-a920-1e40cea0b951","Type":"ContainerStarted","Data":"60552213f653766db7e33d8487a8f75b4a72f2a296e4e3c6b6dbb79e86381d3f"} Nov 26 01:26:59 crc kubenswrapper[4766]: I1126 01:26:59.079785 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" podStartSLOduration=1.49455083 podStartE2EDuration="2.07976405s" podCreationTimestamp="2025-11-26 01:26:57 +0000 UTC" firstStartedPulling="2025-11-26 01:26:58.088797382 +0000 UTC m=+3798.937567843" lastFinishedPulling="2025-11-26 01:26:58.674010603 +0000 UTC m=+3799.522781063" observedRunningTime="2025-11-26 01:26:59.069760257 +0000 UTC m=+3799.918530727" watchObservedRunningTime="2025-11-26 01:26:59.07976405 +0000 UTC m=+3799.928534490" Nov 26 01:27:11 crc kubenswrapper[4766]: I1126 01:27:11.826464 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:27:11 crc kubenswrapper[4766]: E1126 01:27:11.827371 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:27:22 crc kubenswrapper[4766]: I1126 01:27:22.827754 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:27:22 crc kubenswrapper[4766]: E1126 01:27:22.828555 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:27:36 crc kubenswrapper[4766]: I1126 01:27:36.827732 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:27:36 crc kubenswrapper[4766]: E1126 01:27:36.830480 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:27:51 crc kubenswrapper[4766]: I1126 01:27:51.828352 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:27:51 crc kubenswrapper[4766]: E1126 01:27:51.829536 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:28:05 crc kubenswrapper[4766]: I1126 01:28:05.828773 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:28:05 crc kubenswrapper[4766]: E1126 01:28:05.829762 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:28:12 crc kubenswrapper[4766]: I1126 01:28:12.381822 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" podUID="abf46edd-34b9-42e7-94ca-a3066047cb53" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 01:28:17 crc kubenswrapper[4766]: I1126 01:28:17.831475 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:28:17 crc kubenswrapper[4766]: E1126 01:28:17.832990 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:28:24 crc kubenswrapper[4766]: I1126 01:28:24.272892 4766 generic.go:334] "Generic (PLEG): container finished" podID="3efd579d-7c56-4624-a920-1e40cea0b951" containerID="9c7463072f6807c25f87fdfb62520e0580172abc91cf31be81a85eed09f11a44" exitCode=0 Nov 26 01:28:24 crc kubenswrapper[4766]: I1126 01:28:24.272939 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" event={"ID":"3efd579d-7c56-4624-a920-1e40cea0b951","Type":"ContainerDied","Data":"9c7463072f6807c25f87fdfb62520e0580172abc91cf31be81a85eed09f11a44"} Nov 26 01:28:25 crc kubenswrapper[4766]: I1126 01:28:25.867722 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:28:25 crc kubenswrapper[4766]: I1126 01:28:25.927773 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-inventory\") pod \"3efd579d-7c56-4624-a920-1e40cea0b951\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " Nov 26 01:28:25 crc kubenswrapper[4766]: I1126 01:28:25.928810 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blqnv\" (UniqueName: \"kubernetes.io/projected/3efd579d-7c56-4624-a920-1e40cea0b951-kube-api-access-blqnv\") pod \"3efd579d-7c56-4624-a920-1e40cea0b951\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " Nov 26 01:28:25 crc kubenswrapper[4766]: I1126 01:28:25.929132 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-nova-metadata-neutron-config-0\") pod \"3efd579d-7c56-4624-a920-1e40cea0b951\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " Nov 26 01:28:25 crc kubenswrapper[4766]: I1126 01:28:25.929187 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-ceph\") pod \"3efd579d-7c56-4624-a920-1e40cea0b951\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " Nov 26 01:28:25 crc kubenswrapper[4766]: I1126 01:28:25.929305 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-neutron-ovn-metadata-agent-neutron-config-0\") pod \"3efd579d-7c56-4624-a920-1e40cea0b951\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " Nov 26 01:28:25 crc kubenswrapper[4766]: I1126 01:28:25.929336 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-neutron-metadata-combined-ca-bundle\") pod \"3efd579d-7c56-4624-a920-1e40cea0b951\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " Nov 26 01:28:25 crc kubenswrapper[4766]: I1126 01:28:25.929390 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-ssh-key\") pod \"3efd579d-7c56-4624-a920-1e40cea0b951\" (UID: \"3efd579d-7c56-4624-a920-1e40cea0b951\") " Nov 26 01:28:25 crc kubenswrapper[4766]: I1126 01:28:25.937845 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3efd579d-7c56-4624-a920-1e40cea0b951-kube-api-access-blqnv" (OuterVolumeSpecName: "kube-api-access-blqnv") pod "3efd579d-7c56-4624-a920-1e40cea0b951" (UID: "3efd579d-7c56-4624-a920-1e40cea0b951"). InnerVolumeSpecName "kube-api-access-blqnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:28:25 crc kubenswrapper[4766]: I1126 01:28:25.938509 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-ceph" (OuterVolumeSpecName: "ceph") pod "3efd579d-7c56-4624-a920-1e40cea0b951" (UID: "3efd579d-7c56-4624-a920-1e40cea0b951"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:28:25 crc kubenswrapper[4766]: I1126 01:28:25.953733 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "3efd579d-7c56-4624-a920-1e40cea0b951" (UID: "3efd579d-7c56-4624-a920-1e40cea0b951"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:28:25 crc kubenswrapper[4766]: I1126 01:28:25.966820 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3efd579d-7c56-4624-a920-1e40cea0b951" (UID: "3efd579d-7c56-4624-a920-1e40cea0b951"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:28:25 crc kubenswrapper[4766]: I1126 01:28:25.974070 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "3efd579d-7c56-4624-a920-1e40cea0b951" (UID: "3efd579d-7c56-4624-a920-1e40cea0b951"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:28:25 crc kubenswrapper[4766]: I1126 01:28:25.979912 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-inventory" (OuterVolumeSpecName: "inventory") pod "3efd579d-7c56-4624-a920-1e40cea0b951" (UID: "3efd579d-7c56-4624-a920-1e40cea0b951"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:28:25 crc kubenswrapper[4766]: I1126 01:28:25.982415 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "3efd579d-7c56-4624-a920-1e40cea0b951" (UID: "3efd579d-7c56-4624-a920-1e40cea0b951"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.032070 4766 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.032102 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.032114 4766 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.032129 4766 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.032139 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.032148 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3efd579d-7c56-4624-a920-1e40cea0b951-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.032168 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blqnv\" (UniqueName: \"kubernetes.io/projected/3efd579d-7c56-4624-a920-1e40cea0b951-kube-api-access-blqnv\") on node \"crc\" DevicePath \"\"" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.306547 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" event={"ID":"3efd579d-7c56-4624-a920-1e40cea0b951","Type":"ContainerDied","Data":"60552213f653766db7e33d8487a8f75b4a72f2a296e4e3c6b6dbb79e86381d3f"} Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.306616 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60552213f653766db7e33d8487a8f75b4a72f2a296e4e3c6b6dbb79e86381d3f" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.306673 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.467390 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp"] Nov 26 01:28:26 crc kubenswrapper[4766]: E1126 01:28:26.467922 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3efd579d-7c56-4624-a920-1e40cea0b951" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.467938 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="3efd579d-7c56-4624-a920-1e40cea0b951" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.468199 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="3efd579d-7c56-4624-a920-1e40cea0b951" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.469391 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.472601 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.472818 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.472901 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.473049 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.473188 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.473211 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.484308 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp"] Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.542962 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxnn9\" (UniqueName: \"kubernetes.io/projected/64255763-3ccc-42d2-88ae-3c48ecea8c07-kube-api-access-mxnn9\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nswmp\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.543003 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nswmp\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.543049 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nswmp\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.543080 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nswmp\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.543115 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nswmp\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.543139 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nswmp\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.644180 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nswmp\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.644243 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nswmp\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.644268 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nswmp\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.644365 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxnn9\" (UniqueName: \"kubernetes.io/projected/64255763-3ccc-42d2-88ae-3c48ecea8c07-kube-api-access-mxnn9\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nswmp\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.644385 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nswmp\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.644425 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nswmp\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.649884 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nswmp\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.650062 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nswmp\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.650083 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nswmp\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.650326 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nswmp\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.652778 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nswmp\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.665498 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxnn9\" (UniqueName: \"kubernetes.io/projected/64255763-3ccc-42d2-88ae-3c48ecea8c07-kube-api-access-mxnn9\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nswmp\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:28:26 crc kubenswrapper[4766]: I1126 01:28:26.792413 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:28:27 crc kubenswrapper[4766]: I1126 01:28:27.386396 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp"] Nov 26 01:28:27 crc kubenswrapper[4766]: W1126 01:28:27.390161 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64255763_3ccc_42d2_88ae_3c48ecea8c07.slice/crio-963746258be6021f4551b8033cbca9f18d83a2d22964bb65fd25555f70125414 WatchSource:0}: Error finding container 963746258be6021f4551b8033cbca9f18d83a2d22964bb65fd25555f70125414: Status 404 returned error can't find the container with id 963746258be6021f4551b8033cbca9f18d83a2d22964bb65fd25555f70125414 Nov 26 01:28:27 crc kubenswrapper[4766]: I1126 01:28:27.393007 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 01:28:28 crc kubenswrapper[4766]: I1126 01:28:28.335207 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" event={"ID":"64255763-3ccc-42d2-88ae-3c48ecea8c07","Type":"ContainerStarted","Data":"963746258be6021f4551b8033cbca9f18d83a2d22964bb65fd25555f70125414"} Nov 26 01:28:28 crc kubenswrapper[4766]: I1126 01:28:28.367389 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" podStartSLOduration=1.7530205840000002 podStartE2EDuration="2.367358551s" podCreationTimestamp="2025-11-26 01:28:26 +0000 UTC" firstStartedPulling="2025-11-26 01:28:27.39275287 +0000 UTC m=+3888.241523290" lastFinishedPulling="2025-11-26 01:28:28.007090817 +0000 UTC m=+3888.855861257" observedRunningTime="2025-11-26 01:28:28.359362576 +0000 UTC m=+3889.208133006" watchObservedRunningTime="2025-11-26 01:28:28.367358551 +0000 UTC m=+3889.216129011" Nov 26 01:28:29 crc kubenswrapper[4766]: I1126 01:28:29.351129 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" event={"ID":"64255763-3ccc-42d2-88ae-3c48ecea8c07","Type":"ContainerStarted","Data":"32058766d372e79ded5d44dbbad8b501ba9b5dbf818ba226506b93c538301a1e"} Nov 26 01:28:29 crc kubenswrapper[4766]: I1126 01:28:29.843976 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:28:29 crc kubenswrapper[4766]: E1126 01:28:29.844913 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:28:41 crc kubenswrapper[4766]: I1126 01:28:41.827190 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:28:41 crc kubenswrapper[4766]: E1126 01:28:41.828755 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:28:55 crc kubenswrapper[4766]: I1126 01:28:55.827187 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:28:55 crc kubenswrapper[4766]: E1126 01:28:55.828065 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:29:10 crc kubenswrapper[4766]: I1126 01:29:10.826898 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:29:10 crc kubenswrapper[4766]: E1126 01:29:10.827691 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:29:22 crc kubenswrapper[4766]: I1126 01:29:22.827107 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:29:24 crc kubenswrapper[4766]: I1126 01:29:24.165115 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"53c28fab7d2dc492443a92908357f7cd25e699dec76a6009866ab9bb91349133"} Nov 26 01:30:00 crc kubenswrapper[4766]: I1126 01:30:00.164126 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw"] Nov 26 01:30:00 crc kubenswrapper[4766]: I1126 01:30:00.166393 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw" Nov 26 01:30:00 crc kubenswrapper[4766]: I1126 01:30:00.169775 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 26 01:30:00 crc kubenswrapper[4766]: I1126 01:30:00.170506 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 26 01:30:00 crc kubenswrapper[4766]: I1126 01:30:00.182227 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw"] Nov 26 01:30:00 crc kubenswrapper[4766]: I1126 01:30:00.261219 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjrsp\" (UniqueName: \"kubernetes.io/projected/ccb5619d-03b6-4fae-9e5d-950952eb0afe-kube-api-access-vjrsp\") pod \"collect-profiles-29402010-j96pw\" (UID: \"ccb5619d-03b6-4fae-9e5d-950952eb0afe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw" Nov 26 01:30:00 crc kubenswrapper[4766]: I1126 01:30:00.261297 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ccb5619d-03b6-4fae-9e5d-950952eb0afe-config-volume\") pod \"collect-profiles-29402010-j96pw\" (UID: \"ccb5619d-03b6-4fae-9e5d-950952eb0afe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw" Nov 26 01:30:00 crc kubenswrapper[4766]: I1126 01:30:00.261318 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ccb5619d-03b6-4fae-9e5d-950952eb0afe-secret-volume\") pod \"collect-profiles-29402010-j96pw\" (UID: \"ccb5619d-03b6-4fae-9e5d-950952eb0afe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw" Nov 26 01:30:00 crc kubenswrapper[4766]: I1126 01:30:00.363628 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjrsp\" (UniqueName: \"kubernetes.io/projected/ccb5619d-03b6-4fae-9e5d-950952eb0afe-kube-api-access-vjrsp\") pod \"collect-profiles-29402010-j96pw\" (UID: \"ccb5619d-03b6-4fae-9e5d-950952eb0afe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw" Nov 26 01:30:00 crc kubenswrapper[4766]: I1126 01:30:00.363734 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ccb5619d-03b6-4fae-9e5d-950952eb0afe-secret-volume\") pod \"collect-profiles-29402010-j96pw\" (UID: \"ccb5619d-03b6-4fae-9e5d-950952eb0afe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw" Nov 26 01:30:00 crc kubenswrapper[4766]: I1126 01:30:00.363755 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ccb5619d-03b6-4fae-9e5d-950952eb0afe-config-volume\") pod \"collect-profiles-29402010-j96pw\" (UID: \"ccb5619d-03b6-4fae-9e5d-950952eb0afe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw" Nov 26 01:30:00 crc kubenswrapper[4766]: I1126 01:30:00.364936 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ccb5619d-03b6-4fae-9e5d-950952eb0afe-config-volume\") pod \"collect-profiles-29402010-j96pw\" (UID: \"ccb5619d-03b6-4fae-9e5d-950952eb0afe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw" Nov 26 01:30:00 crc kubenswrapper[4766]: I1126 01:30:00.371140 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ccb5619d-03b6-4fae-9e5d-950952eb0afe-secret-volume\") pod \"collect-profiles-29402010-j96pw\" (UID: \"ccb5619d-03b6-4fae-9e5d-950952eb0afe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw" Nov 26 01:30:00 crc kubenswrapper[4766]: I1126 01:30:00.388692 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjrsp\" (UniqueName: \"kubernetes.io/projected/ccb5619d-03b6-4fae-9e5d-950952eb0afe-kube-api-access-vjrsp\") pod \"collect-profiles-29402010-j96pw\" (UID: \"ccb5619d-03b6-4fae-9e5d-950952eb0afe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw" Nov 26 01:30:00 crc kubenswrapper[4766]: I1126 01:30:00.510539 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw" Nov 26 01:30:01 crc kubenswrapper[4766]: I1126 01:30:01.064539 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw"] Nov 26 01:30:01 crc kubenswrapper[4766]: I1126 01:30:01.624366 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw" event={"ID":"ccb5619d-03b6-4fae-9e5d-950952eb0afe","Type":"ContainerStarted","Data":"26e6f8f0c2f7670c708b926520f3d7313f72a5d8dd65b7856b61162c5f475dd5"} Nov 26 01:30:01 crc kubenswrapper[4766]: I1126 01:30:01.625846 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw" event={"ID":"ccb5619d-03b6-4fae-9e5d-950952eb0afe","Type":"ContainerStarted","Data":"ade08315ba8814ca142d2ccf2eeef81d718e1f811a942d5ea58cc5f8d387875c"} Nov 26 01:30:01 crc kubenswrapper[4766]: I1126 01:30:01.654383 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw" podStartSLOduration=1.654356816 podStartE2EDuration="1.654356816s" podCreationTimestamp="2025-11-26 01:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 01:30:01.642830195 +0000 UTC m=+3982.491600635" watchObservedRunningTime="2025-11-26 01:30:01.654356816 +0000 UTC m=+3982.503127246" Nov 26 01:30:02 crc kubenswrapper[4766]: I1126 01:30:02.642144 4766 generic.go:334] "Generic (PLEG): container finished" podID="ccb5619d-03b6-4fae-9e5d-950952eb0afe" containerID="26e6f8f0c2f7670c708b926520f3d7313f72a5d8dd65b7856b61162c5f475dd5" exitCode=0 Nov 26 01:30:02 crc kubenswrapper[4766]: I1126 01:30:02.642238 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw" event={"ID":"ccb5619d-03b6-4fae-9e5d-950952eb0afe","Type":"ContainerDied","Data":"26e6f8f0c2f7670c708b926520f3d7313f72a5d8dd65b7856b61162c5f475dd5"} Nov 26 01:30:04 crc kubenswrapper[4766]: I1126 01:30:04.065986 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw" Nov 26 01:30:04 crc kubenswrapper[4766]: I1126 01:30:04.142113 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjrsp\" (UniqueName: \"kubernetes.io/projected/ccb5619d-03b6-4fae-9e5d-950952eb0afe-kube-api-access-vjrsp\") pod \"ccb5619d-03b6-4fae-9e5d-950952eb0afe\" (UID: \"ccb5619d-03b6-4fae-9e5d-950952eb0afe\") " Nov 26 01:30:04 crc kubenswrapper[4766]: I1126 01:30:04.142304 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ccb5619d-03b6-4fae-9e5d-950952eb0afe-secret-volume\") pod \"ccb5619d-03b6-4fae-9e5d-950952eb0afe\" (UID: \"ccb5619d-03b6-4fae-9e5d-950952eb0afe\") " Nov 26 01:30:04 crc kubenswrapper[4766]: I1126 01:30:04.142341 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ccb5619d-03b6-4fae-9e5d-950952eb0afe-config-volume\") pod \"ccb5619d-03b6-4fae-9e5d-950952eb0afe\" (UID: \"ccb5619d-03b6-4fae-9e5d-950952eb0afe\") " Nov 26 01:30:04 crc kubenswrapper[4766]: I1126 01:30:04.143500 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccb5619d-03b6-4fae-9e5d-950952eb0afe-config-volume" (OuterVolumeSpecName: "config-volume") pod "ccb5619d-03b6-4fae-9e5d-950952eb0afe" (UID: "ccb5619d-03b6-4fae-9e5d-950952eb0afe"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:30:04 crc kubenswrapper[4766]: I1126 01:30:04.151736 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccb5619d-03b6-4fae-9e5d-950952eb0afe-kube-api-access-vjrsp" (OuterVolumeSpecName: "kube-api-access-vjrsp") pod "ccb5619d-03b6-4fae-9e5d-950952eb0afe" (UID: "ccb5619d-03b6-4fae-9e5d-950952eb0afe"). InnerVolumeSpecName "kube-api-access-vjrsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:30:04 crc kubenswrapper[4766]: I1126 01:30:04.152126 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccb5619d-03b6-4fae-9e5d-950952eb0afe-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ccb5619d-03b6-4fae-9e5d-950952eb0afe" (UID: "ccb5619d-03b6-4fae-9e5d-950952eb0afe"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:30:04 crc kubenswrapper[4766]: I1126 01:30:04.244145 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjrsp\" (UniqueName: \"kubernetes.io/projected/ccb5619d-03b6-4fae-9e5d-950952eb0afe-kube-api-access-vjrsp\") on node \"crc\" DevicePath \"\"" Nov 26 01:30:04 crc kubenswrapper[4766]: I1126 01:30:04.244183 4766 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ccb5619d-03b6-4fae-9e5d-950952eb0afe-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 26 01:30:04 crc kubenswrapper[4766]: I1126 01:30:04.244194 4766 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ccb5619d-03b6-4fae-9e5d-950952eb0afe-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 01:30:04 crc kubenswrapper[4766]: I1126 01:30:04.669125 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw" event={"ID":"ccb5619d-03b6-4fae-9e5d-950952eb0afe","Type":"ContainerDied","Data":"ade08315ba8814ca142d2ccf2eeef81d718e1f811a942d5ea58cc5f8d387875c"} Nov 26 01:30:04 crc kubenswrapper[4766]: I1126 01:30:04.669506 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ade08315ba8814ca142d2ccf2eeef81d718e1f811a942d5ea58cc5f8d387875c" Nov 26 01:30:04 crc kubenswrapper[4766]: I1126 01:30:04.669220 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw" Nov 26 01:30:04 crc kubenswrapper[4766]: I1126 01:30:04.766549 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg"] Nov 26 01:30:04 crc kubenswrapper[4766]: I1126 01:30:04.778609 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401965-qggxg"] Nov 26 01:30:05 crc kubenswrapper[4766]: I1126 01:30:05.845921 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32dc2556-2ee6-4b11-90c0-c074b1c2e774" path="/var/lib/kubelet/pods/32dc2556-2ee6-4b11-90c0-c074b1c2e774/volumes" Nov 26 01:31:01 crc kubenswrapper[4766]: I1126 01:31:01.875480 4766 scope.go:117] "RemoveContainer" containerID="0e267bf2eb7b3baa5687c89d55b47b09cf78dd1b9bed01fd6d72dce260665cba" Nov 26 01:31:32 crc kubenswrapper[4766]: I1126 01:31:32.186456 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-w95xh"] Nov 26 01:31:32 crc kubenswrapper[4766]: E1126 01:31:32.187840 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccb5619d-03b6-4fae-9e5d-950952eb0afe" containerName="collect-profiles" Nov 26 01:31:32 crc kubenswrapper[4766]: I1126 01:31:32.187859 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccb5619d-03b6-4fae-9e5d-950952eb0afe" containerName="collect-profiles" Nov 26 01:31:32 crc kubenswrapper[4766]: I1126 01:31:32.188166 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccb5619d-03b6-4fae-9e5d-950952eb0afe" containerName="collect-profiles" Nov 26 01:31:32 crc kubenswrapper[4766]: I1126 01:31:32.190189 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w95xh" Nov 26 01:31:32 crc kubenswrapper[4766]: I1126 01:31:32.243145 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w95xh"] Nov 26 01:31:32 crc kubenswrapper[4766]: I1126 01:31:32.367178 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fb0240b-9aef-450b-b86f-8551dd61fc78-catalog-content\") pod \"community-operators-w95xh\" (UID: \"0fb0240b-9aef-450b-b86f-8551dd61fc78\") " pod="openshift-marketplace/community-operators-w95xh" Nov 26 01:31:32 crc kubenswrapper[4766]: I1126 01:31:32.367486 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fb0240b-9aef-450b-b86f-8551dd61fc78-utilities\") pod \"community-operators-w95xh\" (UID: \"0fb0240b-9aef-450b-b86f-8551dd61fc78\") " pod="openshift-marketplace/community-operators-w95xh" Nov 26 01:31:32 crc kubenswrapper[4766]: I1126 01:31:32.367723 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vk4z\" (UniqueName: \"kubernetes.io/projected/0fb0240b-9aef-450b-b86f-8551dd61fc78-kube-api-access-7vk4z\") pod \"community-operators-w95xh\" (UID: \"0fb0240b-9aef-450b-b86f-8551dd61fc78\") " pod="openshift-marketplace/community-operators-w95xh" Nov 26 01:31:32 crc kubenswrapper[4766]: I1126 01:31:32.469605 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fb0240b-9aef-450b-b86f-8551dd61fc78-catalog-content\") pod \"community-operators-w95xh\" (UID: \"0fb0240b-9aef-450b-b86f-8551dd61fc78\") " pod="openshift-marketplace/community-operators-w95xh" Nov 26 01:31:32 crc kubenswrapper[4766]: I1126 01:31:32.469749 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fb0240b-9aef-450b-b86f-8551dd61fc78-utilities\") pod \"community-operators-w95xh\" (UID: \"0fb0240b-9aef-450b-b86f-8551dd61fc78\") " pod="openshift-marketplace/community-operators-w95xh" Nov 26 01:31:32 crc kubenswrapper[4766]: I1126 01:31:32.469795 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vk4z\" (UniqueName: \"kubernetes.io/projected/0fb0240b-9aef-450b-b86f-8551dd61fc78-kube-api-access-7vk4z\") pod \"community-operators-w95xh\" (UID: \"0fb0240b-9aef-450b-b86f-8551dd61fc78\") " pod="openshift-marketplace/community-operators-w95xh" Nov 26 01:31:32 crc kubenswrapper[4766]: I1126 01:31:32.470289 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fb0240b-9aef-450b-b86f-8551dd61fc78-catalog-content\") pod \"community-operators-w95xh\" (UID: \"0fb0240b-9aef-450b-b86f-8551dd61fc78\") " pod="openshift-marketplace/community-operators-w95xh" Nov 26 01:31:32 crc kubenswrapper[4766]: I1126 01:31:32.470321 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fb0240b-9aef-450b-b86f-8551dd61fc78-utilities\") pod \"community-operators-w95xh\" (UID: \"0fb0240b-9aef-450b-b86f-8551dd61fc78\") " pod="openshift-marketplace/community-operators-w95xh" Nov 26 01:31:32 crc kubenswrapper[4766]: I1126 01:31:32.631519 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vk4z\" (UniqueName: \"kubernetes.io/projected/0fb0240b-9aef-450b-b86f-8551dd61fc78-kube-api-access-7vk4z\") pod \"community-operators-w95xh\" (UID: \"0fb0240b-9aef-450b-b86f-8551dd61fc78\") " pod="openshift-marketplace/community-operators-w95xh" Nov 26 01:31:32 crc kubenswrapper[4766]: I1126 01:31:32.826776 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w95xh" Nov 26 01:31:33 crc kubenswrapper[4766]: I1126 01:31:33.309367 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w95xh"] Nov 26 01:31:33 crc kubenswrapper[4766]: I1126 01:31:33.812802 4766 generic.go:334] "Generic (PLEG): container finished" podID="0fb0240b-9aef-450b-b86f-8551dd61fc78" containerID="2a1be017453c53219866b5e9669eced0ac01ede13f5e0cdd872df3ddb4ec21f2" exitCode=0 Nov 26 01:31:33 crc kubenswrapper[4766]: I1126 01:31:33.813149 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w95xh" event={"ID":"0fb0240b-9aef-450b-b86f-8551dd61fc78","Type":"ContainerDied","Data":"2a1be017453c53219866b5e9669eced0ac01ede13f5e0cdd872df3ddb4ec21f2"} Nov 26 01:31:33 crc kubenswrapper[4766]: I1126 01:31:33.813188 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w95xh" event={"ID":"0fb0240b-9aef-450b-b86f-8551dd61fc78","Type":"ContainerStarted","Data":"ddddd5b3fa4d1a5369dd193c5b9a3c788da5f4479a3f0452fae5acf415112703"} Nov 26 01:31:35 crc kubenswrapper[4766]: I1126 01:31:35.847137 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w95xh" event={"ID":"0fb0240b-9aef-450b-b86f-8551dd61fc78","Type":"ContainerStarted","Data":"7def179ddbd5237df3bfccccdfea7842f407359cccd0132403fef1c32c9580a6"} Nov 26 01:31:36 crc kubenswrapper[4766]: I1126 01:31:36.862390 4766 generic.go:334] "Generic (PLEG): container finished" podID="0fb0240b-9aef-450b-b86f-8551dd61fc78" containerID="7def179ddbd5237df3bfccccdfea7842f407359cccd0132403fef1c32c9580a6" exitCode=0 Nov 26 01:31:36 crc kubenswrapper[4766]: I1126 01:31:36.862862 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w95xh" event={"ID":"0fb0240b-9aef-450b-b86f-8551dd61fc78","Type":"ContainerDied","Data":"7def179ddbd5237df3bfccccdfea7842f407359cccd0132403fef1c32c9580a6"} Nov 26 01:31:37 crc kubenswrapper[4766]: I1126 01:31:37.878547 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w95xh" event={"ID":"0fb0240b-9aef-450b-b86f-8551dd61fc78","Type":"ContainerStarted","Data":"71c07b29d3ec16e030c9fdba490dad2911bb168b996c642ffedd37a6c2c14241"} Nov 26 01:31:37 crc kubenswrapper[4766]: I1126 01:31:37.903741 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-w95xh" podStartSLOduration=2.436328883 podStartE2EDuration="5.903720853s" podCreationTimestamp="2025-11-26 01:31:32 +0000 UTC" firstStartedPulling="2025-11-26 01:31:33.815854256 +0000 UTC m=+4074.664624716" lastFinishedPulling="2025-11-26 01:31:37.283246256 +0000 UTC m=+4078.132016686" observedRunningTime="2025-11-26 01:31:37.900286949 +0000 UTC m=+4078.749057429" watchObservedRunningTime="2025-11-26 01:31:37.903720853 +0000 UTC m=+4078.752491293" Nov 26 01:31:41 crc kubenswrapper[4766]: I1126 01:31:41.484911 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:31:41 crc kubenswrapper[4766]: I1126 01:31:41.485515 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:31:42 crc kubenswrapper[4766]: I1126 01:31:42.827719 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-w95xh" Nov 26 01:31:42 crc kubenswrapper[4766]: I1126 01:31:42.828085 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-w95xh" Nov 26 01:31:42 crc kubenswrapper[4766]: I1126 01:31:42.929088 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-w95xh" Nov 26 01:31:43 crc kubenswrapper[4766]: I1126 01:31:43.009456 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-w95xh" Nov 26 01:31:43 crc kubenswrapper[4766]: I1126 01:31:43.192888 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w95xh"] Nov 26 01:31:44 crc kubenswrapper[4766]: I1126 01:31:44.973289 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-w95xh" podUID="0fb0240b-9aef-450b-b86f-8551dd61fc78" containerName="registry-server" containerID="cri-o://71c07b29d3ec16e030c9fdba490dad2911bb168b996c642ffedd37a6c2c14241" gracePeriod=2 Nov 26 01:31:45 crc kubenswrapper[4766]: I1126 01:31:45.991064 4766 generic.go:334] "Generic (PLEG): container finished" podID="0fb0240b-9aef-450b-b86f-8551dd61fc78" containerID="71c07b29d3ec16e030c9fdba490dad2911bb168b996c642ffedd37a6c2c14241" exitCode=0 Nov 26 01:31:45 crc kubenswrapper[4766]: I1126 01:31:45.991162 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w95xh" event={"ID":"0fb0240b-9aef-450b-b86f-8551dd61fc78","Type":"ContainerDied","Data":"71c07b29d3ec16e030c9fdba490dad2911bb168b996c642ffedd37a6c2c14241"} Nov 26 01:31:46 crc kubenswrapper[4766]: I1126 01:31:46.432500 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w95xh" Nov 26 01:31:46 crc kubenswrapper[4766]: I1126 01:31:46.495955 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fb0240b-9aef-450b-b86f-8551dd61fc78-catalog-content\") pod \"0fb0240b-9aef-450b-b86f-8551dd61fc78\" (UID: \"0fb0240b-9aef-450b-b86f-8551dd61fc78\") " Nov 26 01:31:46 crc kubenswrapper[4766]: I1126 01:31:46.496361 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fb0240b-9aef-450b-b86f-8551dd61fc78-utilities\") pod \"0fb0240b-9aef-450b-b86f-8551dd61fc78\" (UID: \"0fb0240b-9aef-450b-b86f-8551dd61fc78\") " Nov 26 01:31:46 crc kubenswrapper[4766]: I1126 01:31:46.496421 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vk4z\" (UniqueName: \"kubernetes.io/projected/0fb0240b-9aef-450b-b86f-8551dd61fc78-kube-api-access-7vk4z\") pod \"0fb0240b-9aef-450b-b86f-8551dd61fc78\" (UID: \"0fb0240b-9aef-450b-b86f-8551dd61fc78\") " Nov 26 01:31:46 crc kubenswrapper[4766]: I1126 01:31:46.499266 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fb0240b-9aef-450b-b86f-8551dd61fc78-utilities" (OuterVolumeSpecName: "utilities") pod "0fb0240b-9aef-450b-b86f-8551dd61fc78" (UID: "0fb0240b-9aef-450b-b86f-8551dd61fc78"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:31:46 crc kubenswrapper[4766]: I1126 01:31:46.505740 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fb0240b-9aef-450b-b86f-8551dd61fc78-kube-api-access-7vk4z" (OuterVolumeSpecName: "kube-api-access-7vk4z") pod "0fb0240b-9aef-450b-b86f-8551dd61fc78" (UID: "0fb0240b-9aef-450b-b86f-8551dd61fc78"). InnerVolumeSpecName "kube-api-access-7vk4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:31:46 crc kubenswrapper[4766]: I1126 01:31:46.558461 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fb0240b-9aef-450b-b86f-8551dd61fc78-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0fb0240b-9aef-450b-b86f-8551dd61fc78" (UID: "0fb0240b-9aef-450b-b86f-8551dd61fc78"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:31:46 crc kubenswrapper[4766]: I1126 01:31:46.599195 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fb0240b-9aef-450b-b86f-8551dd61fc78-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 01:31:46 crc kubenswrapper[4766]: I1126 01:31:46.599236 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fb0240b-9aef-450b-b86f-8551dd61fc78-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 01:31:46 crc kubenswrapper[4766]: I1126 01:31:46.599252 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vk4z\" (UniqueName: \"kubernetes.io/projected/0fb0240b-9aef-450b-b86f-8551dd61fc78-kube-api-access-7vk4z\") on node \"crc\" DevicePath \"\"" Nov 26 01:31:47 crc kubenswrapper[4766]: I1126 01:31:47.010985 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w95xh" event={"ID":"0fb0240b-9aef-450b-b86f-8551dd61fc78","Type":"ContainerDied","Data":"ddddd5b3fa4d1a5369dd193c5b9a3c788da5f4479a3f0452fae5acf415112703"} Nov 26 01:31:47 crc kubenswrapper[4766]: I1126 01:31:47.011062 4766 scope.go:117] "RemoveContainer" containerID="71c07b29d3ec16e030c9fdba490dad2911bb168b996c642ffedd37a6c2c14241" Nov 26 01:31:47 crc kubenswrapper[4766]: I1126 01:31:47.011148 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w95xh" Nov 26 01:31:47 crc kubenswrapper[4766]: I1126 01:31:47.065731 4766 scope.go:117] "RemoveContainer" containerID="7def179ddbd5237df3bfccccdfea7842f407359cccd0132403fef1c32c9580a6" Nov 26 01:31:47 crc kubenswrapper[4766]: I1126 01:31:47.093761 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w95xh"] Nov 26 01:31:47 crc kubenswrapper[4766]: I1126 01:31:47.112507 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-w95xh"] Nov 26 01:31:47 crc kubenswrapper[4766]: I1126 01:31:47.113090 4766 scope.go:117] "RemoveContainer" containerID="2a1be017453c53219866b5e9669eced0ac01ede13f5e0cdd872df3ddb4ec21f2" Nov 26 01:31:47 crc kubenswrapper[4766]: I1126 01:31:47.852735 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fb0240b-9aef-450b-b86f-8551dd61fc78" path="/var/lib/kubelet/pods/0fb0240b-9aef-450b-b86f-8551dd61fc78/volumes" Nov 26 01:31:48 crc kubenswrapper[4766]: I1126 01:31:48.808842 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="44f36c8b-51b8-4bce-afbb-1962eb317dec" containerName="galera" probeResult="failure" output="command timed out" Nov 26 01:31:48 crc kubenswrapper[4766]: I1126 01:31:48.809102 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="44f36c8b-51b8-4bce-afbb-1962eb317dec" containerName="galera" probeResult="failure" output="command timed out" Nov 26 01:32:11 crc kubenswrapper[4766]: I1126 01:32:11.479562 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:32:11 crc kubenswrapper[4766]: I1126 01:32:11.480142 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:32:19 crc kubenswrapper[4766]: I1126 01:32:19.503340 4766 generic.go:334] "Generic (PLEG): container finished" podID="64255763-3ccc-42d2-88ae-3c48ecea8c07" containerID="32058766d372e79ded5d44dbbad8b501ba9b5dbf818ba226506b93c538301a1e" exitCode=0 Nov 26 01:32:19 crc kubenswrapper[4766]: I1126 01:32:19.504107 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" event={"ID":"64255763-3ccc-42d2-88ae-3c48ecea8c07","Type":"ContainerDied","Data":"32058766d372e79ded5d44dbbad8b501ba9b5dbf818ba226506b93c538301a1e"} Nov 26 01:32:20 crc kubenswrapper[4766]: I1126 01:32:20.992820 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.032642 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-libvirt-secret-0\") pod \"64255763-3ccc-42d2-88ae-3c48ecea8c07\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.032725 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxnn9\" (UniqueName: \"kubernetes.io/projected/64255763-3ccc-42d2-88ae-3c48ecea8c07-kube-api-access-mxnn9\") pod \"64255763-3ccc-42d2-88ae-3c48ecea8c07\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.032823 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-ceph\") pod \"64255763-3ccc-42d2-88ae-3c48ecea8c07\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.032873 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-ssh-key\") pod \"64255763-3ccc-42d2-88ae-3c48ecea8c07\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.032988 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-libvirt-combined-ca-bundle\") pod \"64255763-3ccc-42d2-88ae-3c48ecea8c07\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.033084 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-inventory\") pod \"64255763-3ccc-42d2-88ae-3c48ecea8c07\" (UID: \"64255763-3ccc-42d2-88ae-3c48ecea8c07\") " Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.039057 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64255763-3ccc-42d2-88ae-3c48ecea8c07-kube-api-access-mxnn9" (OuterVolumeSpecName: "kube-api-access-mxnn9") pod "64255763-3ccc-42d2-88ae-3c48ecea8c07" (UID: "64255763-3ccc-42d2-88ae-3c48ecea8c07"). InnerVolumeSpecName "kube-api-access-mxnn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.040900 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-ceph" (OuterVolumeSpecName: "ceph") pod "64255763-3ccc-42d2-88ae-3c48ecea8c07" (UID: "64255763-3ccc-42d2-88ae-3c48ecea8c07"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.041966 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "64255763-3ccc-42d2-88ae-3c48ecea8c07" (UID: "64255763-3ccc-42d2-88ae-3c48ecea8c07"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.071146 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "64255763-3ccc-42d2-88ae-3c48ecea8c07" (UID: "64255763-3ccc-42d2-88ae-3c48ecea8c07"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.085712 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "64255763-3ccc-42d2-88ae-3c48ecea8c07" (UID: "64255763-3ccc-42d2-88ae-3c48ecea8c07"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.100086 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-inventory" (OuterVolumeSpecName: "inventory") pod "64255763-3ccc-42d2-88ae-3c48ecea8c07" (UID: "64255763-3ccc-42d2-88ae-3c48ecea8c07"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.134808 4766 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.134842 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.134855 4766 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.134868 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxnn9\" (UniqueName: \"kubernetes.io/projected/64255763-3ccc-42d2-88ae-3c48ecea8c07-kube-api-access-mxnn9\") on node \"crc\" DevicePath \"\"" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.134879 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.134890 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/64255763-3ccc-42d2-88ae-3c48ecea8c07-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.533607 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" event={"ID":"64255763-3ccc-42d2-88ae-3c48ecea8c07","Type":"ContainerDied","Data":"963746258be6021f4551b8033cbca9f18d83a2d22964bb65fd25555f70125414"} Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.533726 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="963746258be6021f4551b8033cbca9f18d83a2d22964bb65fd25555f70125414" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.533838 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nswmp" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.667568 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964"] Nov 26 01:32:21 crc kubenswrapper[4766]: E1126 01:32:21.668320 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fb0240b-9aef-450b-b86f-8551dd61fc78" containerName="extract-content" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.668337 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fb0240b-9aef-450b-b86f-8551dd61fc78" containerName="extract-content" Nov 26 01:32:21 crc kubenswrapper[4766]: E1126 01:32:21.668353 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64255763-3ccc-42d2-88ae-3c48ecea8c07" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.668365 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="64255763-3ccc-42d2-88ae-3c48ecea8c07" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 26 01:32:21 crc kubenswrapper[4766]: E1126 01:32:21.668407 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fb0240b-9aef-450b-b86f-8551dd61fc78" containerName="extract-utilities" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.668416 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fb0240b-9aef-450b-b86f-8551dd61fc78" containerName="extract-utilities" Nov 26 01:32:21 crc kubenswrapper[4766]: E1126 01:32:21.668437 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fb0240b-9aef-450b-b86f-8551dd61fc78" containerName="registry-server" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.668445 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fb0240b-9aef-450b-b86f-8551dd61fc78" containerName="registry-server" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.668996 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fb0240b-9aef-450b-b86f-8551dd61fc78" containerName="registry-server" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.669024 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="64255763-3ccc-42d2-88ae-3c48ecea8c07" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.669917 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.673928 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.674091 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.674123 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.674224 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.674371 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.674614 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.674814 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.674959 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.675239 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.683413 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964"] Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.743868 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.744075 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.744203 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/8da0c4a5-4c26-46de-8408-a8b562ba8f76-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.744325 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.744431 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.744529 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.744635 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.744742 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.744773 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.744873 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlhgm\" (UniqueName: \"kubernetes.io/projected/8da0c4a5-4c26-46de-8408-a8b562ba8f76-kube-api-access-wlhgm\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.745002 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.847150 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.847309 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.847398 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.847486 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.847716 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlhgm\" (UniqueName: \"kubernetes.io/projected/8da0c4a5-4c26-46de-8408-a8b562ba8f76-kube-api-access-wlhgm\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.847847 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.848016 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.848179 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.848311 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/8da0c4a5-4c26-46de-8408-a8b562ba8f76-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.848369 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.848445 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.848853 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.851512 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/8da0c4a5-4c26-46de-8408-a8b562ba8f76-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.853439 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.853969 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.855640 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.856343 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.857045 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.858232 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.860185 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.860848 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:21 crc kubenswrapper[4766]: I1126 01:32:21.865151 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlhgm\" (UniqueName: \"kubernetes.io/projected/8da0c4a5-4c26-46de-8408-a8b562ba8f76-kube-api-access-wlhgm\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:22 crc kubenswrapper[4766]: I1126 01:32:22.008819 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:32:22 crc kubenswrapper[4766]: I1126 01:32:22.602800 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964"] Nov 26 01:32:23 crc kubenswrapper[4766]: I1126 01:32:23.557953 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" event={"ID":"8da0c4a5-4c26-46de-8408-a8b562ba8f76","Type":"ContainerStarted","Data":"4ecbd58c6bd96e44ea7745ce725b13265105c7ac1c1a8350cfe5c801ce3499f4"} Nov 26 01:32:23 crc kubenswrapper[4766]: I1126 01:32:23.558269 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" event={"ID":"8da0c4a5-4c26-46de-8408-a8b562ba8f76","Type":"ContainerStarted","Data":"d2a703a3ac62ed41996caad696bda01e5e5748dee7834b5bd41b703bfdccf466"} Nov 26 01:32:23 crc kubenswrapper[4766]: I1126 01:32:23.604120 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" podStartSLOduration=2.119857543 podStartE2EDuration="2.60409782s" podCreationTimestamp="2025-11-26 01:32:21 +0000 UTC" firstStartedPulling="2025-11-26 01:32:22.595277573 +0000 UTC m=+4123.444048013" lastFinishedPulling="2025-11-26 01:32:23.07951786 +0000 UTC m=+4123.928288290" observedRunningTime="2025-11-26 01:32:23.58648546 +0000 UTC m=+4124.435255910" watchObservedRunningTime="2025-11-26 01:32:23.60409782 +0000 UTC m=+4124.452868260" Nov 26 01:32:41 crc kubenswrapper[4766]: I1126 01:32:41.479342 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:32:41 crc kubenswrapper[4766]: I1126 01:32:41.480135 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:32:41 crc kubenswrapper[4766]: I1126 01:32:41.480209 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 01:32:41 crc kubenswrapper[4766]: I1126 01:32:41.481533 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"53c28fab7d2dc492443a92908357f7cd25e699dec76a6009866ab9bb91349133"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 01:32:41 crc kubenswrapper[4766]: I1126 01:32:41.481603 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://53c28fab7d2dc492443a92908357f7cd25e699dec76a6009866ab9bb91349133" gracePeriod=600 Nov 26 01:32:41 crc kubenswrapper[4766]: I1126 01:32:41.827952 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="53c28fab7d2dc492443a92908357f7cd25e699dec76a6009866ab9bb91349133" exitCode=0 Nov 26 01:32:41 crc kubenswrapper[4766]: I1126 01:32:41.839253 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"53c28fab7d2dc492443a92908357f7cd25e699dec76a6009866ab9bb91349133"} Nov 26 01:32:41 crc kubenswrapper[4766]: I1126 01:32:41.839309 4766 scope.go:117] "RemoveContainer" containerID="7338e0f9e870120881007f94e741776a7118e16106eeaebbf573909312312f24" Nov 26 01:32:42 crc kubenswrapper[4766]: I1126 01:32:42.845508 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657"} Nov 26 01:34:41 crc kubenswrapper[4766]: I1126 01:34:41.479619 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:34:41 crc kubenswrapper[4766]: I1126 01:34:41.480319 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:35:11 crc kubenswrapper[4766]: I1126 01:35:11.484294 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:35:11 crc kubenswrapper[4766]: I1126 01:35:11.485254 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:35:18 crc kubenswrapper[4766]: I1126 01:35:18.400551 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dzvnw"] Nov 26 01:35:18 crc kubenswrapper[4766]: I1126 01:35:18.403937 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dzvnw" Nov 26 01:35:18 crc kubenswrapper[4766]: I1126 01:35:18.423506 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dzvnw"] Nov 26 01:35:18 crc kubenswrapper[4766]: I1126 01:35:18.424886 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjgb4\" (UniqueName: \"kubernetes.io/projected/03c0dd64-5cd2-4ec9-add0-45d28db2e1e0-kube-api-access-fjgb4\") pod \"certified-operators-dzvnw\" (UID: \"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0\") " pod="openshift-marketplace/certified-operators-dzvnw" Nov 26 01:35:18 crc kubenswrapper[4766]: I1126 01:35:18.425032 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03c0dd64-5cd2-4ec9-add0-45d28db2e1e0-utilities\") pod \"certified-operators-dzvnw\" (UID: \"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0\") " pod="openshift-marketplace/certified-operators-dzvnw" Nov 26 01:35:18 crc kubenswrapper[4766]: I1126 01:35:18.425141 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03c0dd64-5cd2-4ec9-add0-45d28db2e1e0-catalog-content\") pod \"certified-operators-dzvnw\" (UID: \"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0\") " pod="openshift-marketplace/certified-operators-dzvnw" Nov 26 01:35:18 crc kubenswrapper[4766]: I1126 01:35:18.527052 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03c0dd64-5cd2-4ec9-add0-45d28db2e1e0-catalog-content\") pod \"certified-operators-dzvnw\" (UID: \"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0\") " pod="openshift-marketplace/certified-operators-dzvnw" Nov 26 01:35:18 crc kubenswrapper[4766]: I1126 01:35:18.527180 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjgb4\" (UniqueName: \"kubernetes.io/projected/03c0dd64-5cd2-4ec9-add0-45d28db2e1e0-kube-api-access-fjgb4\") pod \"certified-operators-dzvnw\" (UID: \"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0\") " pod="openshift-marketplace/certified-operators-dzvnw" Nov 26 01:35:18 crc kubenswrapper[4766]: I1126 01:35:18.527609 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03c0dd64-5cd2-4ec9-add0-45d28db2e1e0-catalog-content\") pod \"certified-operators-dzvnw\" (UID: \"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0\") " pod="openshift-marketplace/certified-operators-dzvnw" Nov 26 01:35:18 crc kubenswrapper[4766]: I1126 01:35:18.527628 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03c0dd64-5cd2-4ec9-add0-45d28db2e1e0-utilities\") pod \"certified-operators-dzvnw\" (UID: \"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0\") " pod="openshift-marketplace/certified-operators-dzvnw" Nov 26 01:35:18 crc kubenswrapper[4766]: I1126 01:35:18.527982 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03c0dd64-5cd2-4ec9-add0-45d28db2e1e0-utilities\") pod \"certified-operators-dzvnw\" (UID: \"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0\") " pod="openshift-marketplace/certified-operators-dzvnw" Nov 26 01:35:18 crc kubenswrapper[4766]: I1126 01:35:18.568700 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjgb4\" (UniqueName: \"kubernetes.io/projected/03c0dd64-5cd2-4ec9-add0-45d28db2e1e0-kube-api-access-fjgb4\") pod \"certified-operators-dzvnw\" (UID: \"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0\") " pod="openshift-marketplace/certified-operators-dzvnw" Nov 26 01:35:18 crc kubenswrapper[4766]: I1126 01:35:18.729682 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dzvnw" Nov 26 01:35:19 crc kubenswrapper[4766]: I1126 01:35:19.287257 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dzvnw"] Nov 26 01:35:19 crc kubenswrapper[4766]: W1126 01:35:19.294414 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03c0dd64_5cd2_4ec9_add0_45d28db2e1e0.slice/crio-880f938190a740d66dc52c8152d04b0a440e0f73f5ae8904ab02f9df3c3e738a WatchSource:0}: Error finding container 880f938190a740d66dc52c8152d04b0a440e0f73f5ae8904ab02f9df3c3e738a: Status 404 returned error can't find the container with id 880f938190a740d66dc52c8152d04b0a440e0f73f5ae8904ab02f9df3c3e738a Nov 26 01:35:19 crc kubenswrapper[4766]: I1126 01:35:19.988681 4766 generic.go:334] "Generic (PLEG): container finished" podID="03c0dd64-5cd2-4ec9-add0-45d28db2e1e0" containerID="64d3767ccaf134301962316ec76fae5d44dd88c95bf4d6391b88cadb91965adf" exitCode=0 Nov 26 01:35:19 crc kubenswrapper[4766]: I1126 01:35:19.988814 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dzvnw" event={"ID":"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0","Type":"ContainerDied","Data":"64d3767ccaf134301962316ec76fae5d44dd88c95bf4d6391b88cadb91965adf"} Nov 26 01:35:19 crc kubenswrapper[4766]: I1126 01:35:19.989076 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dzvnw" event={"ID":"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0","Type":"ContainerStarted","Data":"880f938190a740d66dc52c8152d04b0a440e0f73f5ae8904ab02f9df3c3e738a"} Nov 26 01:35:19 crc kubenswrapper[4766]: I1126 01:35:19.992151 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 01:35:21 crc kubenswrapper[4766]: I1126 01:35:21.002687 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dzvnw" event={"ID":"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0","Type":"ContainerStarted","Data":"6d1a4ba4e83318e36e4e64798dd3fc0aeb5eb01e46ebad55b430926613a3a92d"} Nov 26 01:35:22 crc kubenswrapper[4766]: I1126 01:35:22.014072 4766 generic.go:334] "Generic (PLEG): container finished" podID="03c0dd64-5cd2-4ec9-add0-45d28db2e1e0" containerID="6d1a4ba4e83318e36e4e64798dd3fc0aeb5eb01e46ebad55b430926613a3a92d" exitCode=0 Nov 26 01:35:22 crc kubenswrapper[4766]: I1126 01:35:22.014136 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dzvnw" event={"ID":"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0","Type":"ContainerDied","Data":"6d1a4ba4e83318e36e4e64798dd3fc0aeb5eb01e46ebad55b430926613a3a92d"} Nov 26 01:35:23 crc kubenswrapper[4766]: I1126 01:35:23.027148 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dzvnw" event={"ID":"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0","Type":"ContainerStarted","Data":"e7dc565a4e5c3f78870cd835f1d750cef2352d82666b680a05458e55ccc79a1d"} Nov 26 01:35:23 crc kubenswrapper[4766]: I1126 01:35:23.066162 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dzvnw" podStartSLOduration=2.637847633 podStartE2EDuration="5.066133947s" podCreationTimestamp="2025-11-26 01:35:18 +0000 UTC" firstStartedPulling="2025-11-26 01:35:19.991862287 +0000 UTC m=+4300.840632737" lastFinishedPulling="2025-11-26 01:35:22.420148611 +0000 UTC m=+4303.268919051" observedRunningTime="2025-11-26 01:35:23.048613739 +0000 UTC m=+4303.897384189" watchObservedRunningTime="2025-11-26 01:35:23.066133947 +0000 UTC m=+4303.914904417" Nov 26 01:35:28 crc kubenswrapper[4766]: I1126 01:35:28.729913 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dzvnw" Nov 26 01:35:28 crc kubenswrapper[4766]: I1126 01:35:28.732130 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dzvnw" Nov 26 01:35:28 crc kubenswrapper[4766]: I1126 01:35:28.817447 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dzvnw" Nov 26 01:35:29 crc kubenswrapper[4766]: I1126 01:35:29.185473 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dzvnw" Nov 26 01:35:29 crc kubenswrapper[4766]: I1126 01:35:29.243975 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dzvnw"] Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.138128 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dzvnw" podUID="03c0dd64-5cd2-4ec9-add0-45d28db2e1e0" containerName="registry-server" containerID="cri-o://e7dc565a4e5c3f78870cd835f1d750cef2352d82666b680a05458e55ccc79a1d" gracePeriod=2 Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.486187 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8h7zp"] Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.488590 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8h7zp" Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.501179 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8h7zp"] Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.558473 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bc01db8-70da-4234-9c0b-247908a59973-catalog-content\") pod \"redhat-marketplace-8h7zp\" (UID: \"0bc01db8-70da-4234-9c0b-247908a59973\") " pod="openshift-marketplace/redhat-marketplace-8h7zp" Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.558882 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bc01db8-70da-4234-9c0b-247908a59973-utilities\") pod \"redhat-marketplace-8h7zp\" (UID: \"0bc01db8-70da-4234-9c0b-247908a59973\") " pod="openshift-marketplace/redhat-marketplace-8h7zp" Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.558955 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz9dt\" (UniqueName: \"kubernetes.io/projected/0bc01db8-70da-4234-9c0b-247908a59973-kube-api-access-gz9dt\") pod \"redhat-marketplace-8h7zp\" (UID: \"0bc01db8-70da-4234-9c0b-247908a59973\") " pod="openshift-marketplace/redhat-marketplace-8h7zp" Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.660444 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bc01db8-70da-4234-9c0b-247908a59973-utilities\") pod \"redhat-marketplace-8h7zp\" (UID: \"0bc01db8-70da-4234-9c0b-247908a59973\") " pod="openshift-marketplace/redhat-marketplace-8h7zp" Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.660509 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz9dt\" (UniqueName: \"kubernetes.io/projected/0bc01db8-70da-4234-9c0b-247908a59973-kube-api-access-gz9dt\") pod \"redhat-marketplace-8h7zp\" (UID: \"0bc01db8-70da-4234-9c0b-247908a59973\") " pod="openshift-marketplace/redhat-marketplace-8h7zp" Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.660594 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bc01db8-70da-4234-9c0b-247908a59973-catalog-content\") pod \"redhat-marketplace-8h7zp\" (UID: \"0bc01db8-70da-4234-9c0b-247908a59973\") " pod="openshift-marketplace/redhat-marketplace-8h7zp" Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.661055 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bc01db8-70da-4234-9c0b-247908a59973-utilities\") pod \"redhat-marketplace-8h7zp\" (UID: \"0bc01db8-70da-4234-9c0b-247908a59973\") " pod="openshift-marketplace/redhat-marketplace-8h7zp" Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.664099 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bc01db8-70da-4234-9c0b-247908a59973-catalog-content\") pod \"redhat-marketplace-8h7zp\" (UID: \"0bc01db8-70da-4234-9c0b-247908a59973\") " pod="openshift-marketplace/redhat-marketplace-8h7zp" Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.681602 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz9dt\" (UniqueName: \"kubernetes.io/projected/0bc01db8-70da-4234-9c0b-247908a59973-kube-api-access-gz9dt\") pod \"redhat-marketplace-8h7zp\" (UID: \"0bc01db8-70da-4234-9c0b-247908a59973\") " pod="openshift-marketplace/redhat-marketplace-8h7zp" Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.740078 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dzvnw" Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.819856 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8h7zp" Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.863054 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03c0dd64-5cd2-4ec9-add0-45d28db2e1e0-catalog-content\") pod \"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0\" (UID: \"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0\") " Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.863140 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjgb4\" (UniqueName: \"kubernetes.io/projected/03c0dd64-5cd2-4ec9-add0-45d28db2e1e0-kube-api-access-fjgb4\") pod \"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0\" (UID: \"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0\") " Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.863257 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03c0dd64-5cd2-4ec9-add0-45d28db2e1e0-utilities\") pod \"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0\" (UID: \"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0\") " Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.864269 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03c0dd64-5cd2-4ec9-add0-45d28db2e1e0-utilities" (OuterVolumeSpecName: "utilities") pod "03c0dd64-5cd2-4ec9-add0-45d28db2e1e0" (UID: "03c0dd64-5cd2-4ec9-add0-45d28db2e1e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.867824 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03c0dd64-5cd2-4ec9-add0-45d28db2e1e0-kube-api-access-fjgb4" (OuterVolumeSpecName: "kube-api-access-fjgb4") pod "03c0dd64-5cd2-4ec9-add0-45d28db2e1e0" (UID: "03c0dd64-5cd2-4ec9-add0-45d28db2e1e0"). InnerVolumeSpecName "kube-api-access-fjgb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.915991 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03c0dd64-5cd2-4ec9-add0-45d28db2e1e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03c0dd64-5cd2-4ec9-add0-45d28db2e1e0" (UID: "03c0dd64-5cd2-4ec9-add0-45d28db2e1e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.966192 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03c0dd64-5cd2-4ec9-add0-45d28db2e1e0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.966228 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjgb4\" (UniqueName: \"kubernetes.io/projected/03c0dd64-5cd2-4ec9-add0-45d28db2e1e0-kube-api-access-fjgb4\") on node \"crc\" DevicePath \"\"" Nov 26 01:35:31 crc kubenswrapper[4766]: I1126 01:35:31.966238 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03c0dd64-5cd2-4ec9-add0-45d28db2e1e0-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 01:35:32 crc kubenswrapper[4766]: I1126 01:35:32.149494 4766 generic.go:334] "Generic (PLEG): container finished" podID="03c0dd64-5cd2-4ec9-add0-45d28db2e1e0" containerID="e7dc565a4e5c3f78870cd835f1d750cef2352d82666b680a05458e55ccc79a1d" exitCode=0 Nov 26 01:35:32 crc kubenswrapper[4766]: I1126 01:35:32.149526 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dzvnw" event={"ID":"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0","Type":"ContainerDied","Data":"e7dc565a4e5c3f78870cd835f1d750cef2352d82666b680a05458e55ccc79a1d"} Nov 26 01:35:32 crc kubenswrapper[4766]: I1126 01:35:32.149589 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dzvnw" event={"ID":"03c0dd64-5cd2-4ec9-add0-45d28db2e1e0","Type":"ContainerDied","Data":"880f938190a740d66dc52c8152d04b0a440e0f73f5ae8904ab02f9df3c3e738a"} Nov 26 01:35:32 crc kubenswrapper[4766]: I1126 01:35:32.149588 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dzvnw" Nov 26 01:35:32 crc kubenswrapper[4766]: I1126 01:35:32.149609 4766 scope.go:117] "RemoveContainer" containerID="e7dc565a4e5c3f78870cd835f1d750cef2352d82666b680a05458e55ccc79a1d" Nov 26 01:35:32 crc kubenswrapper[4766]: I1126 01:35:32.169969 4766 scope.go:117] "RemoveContainer" containerID="6d1a4ba4e83318e36e4e64798dd3fc0aeb5eb01e46ebad55b430926613a3a92d" Nov 26 01:35:32 crc kubenswrapper[4766]: I1126 01:35:32.188570 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dzvnw"] Nov 26 01:35:32 crc kubenswrapper[4766]: I1126 01:35:32.199907 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dzvnw"] Nov 26 01:35:32 crc kubenswrapper[4766]: I1126 01:35:32.212188 4766 scope.go:117] "RemoveContainer" containerID="64d3767ccaf134301962316ec76fae5d44dd88c95bf4d6391b88cadb91965adf" Nov 26 01:35:32 crc kubenswrapper[4766]: I1126 01:35:32.248323 4766 scope.go:117] "RemoveContainer" containerID="e7dc565a4e5c3f78870cd835f1d750cef2352d82666b680a05458e55ccc79a1d" Nov 26 01:35:32 crc kubenswrapper[4766]: E1126 01:35:32.249001 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7dc565a4e5c3f78870cd835f1d750cef2352d82666b680a05458e55ccc79a1d\": container with ID starting with e7dc565a4e5c3f78870cd835f1d750cef2352d82666b680a05458e55ccc79a1d not found: ID does not exist" containerID="e7dc565a4e5c3f78870cd835f1d750cef2352d82666b680a05458e55ccc79a1d" Nov 26 01:35:32 crc kubenswrapper[4766]: I1126 01:35:32.249066 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7dc565a4e5c3f78870cd835f1d750cef2352d82666b680a05458e55ccc79a1d"} err="failed to get container status \"e7dc565a4e5c3f78870cd835f1d750cef2352d82666b680a05458e55ccc79a1d\": rpc error: code = NotFound desc = could not find container \"e7dc565a4e5c3f78870cd835f1d750cef2352d82666b680a05458e55ccc79a1d\": container with ID starting with e7dc565a4e5c3f78870cd835f1d750cef2352d82666b680a05458e55ccc79a1d not found: ID does not exist" Nov 26 01:35:32 crc kubenswrapper[4766]: I1126 01:35:32.249108 4766 scope.go:117] "RemoveContainer" containerID="6d1a4ba4e83318e36e4e64798dd3fc0aeb5eb01e46ebad55b430926613a3a92d" Nov 26 01:35:32 crc kubenswrapper[4766]: E1126 01:35:32.252631 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d1a4ba4e83318e36e4e64798dd3fc0aeb5eb01e46ebad55b430926613a3a92d\": container with ID starting with 6d1a4ba4e83318e36e4e64798dd3fc0aeb5eb01e46ebad55b430926613a3a92d not found: ID does not exist" containerID="6d1a4ba4e83318e36e4e64798dd3fc0aeb5eb01e46ebad55b430926613a3a92d" Nov 26 01:35:32 crc kubenswrapper[4766]: I1126 01:35:32.253397 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d1a4ba4e83318e36e4e64798dd3fc0aeb5eb01e46ebad55b430926613a3a92d"} err="failed to get container status \"6d1a4ba4e83318e36e4e64798dd3fc0aeb5eb01e46ebad55b430926613a3a92d\": rpc error: code = NotFound desc = could not find container \"6d1a4ba4e83318e36e4e64798dd3fc0aeb5eb01e46ebad55b430926613a3a92d\": container with ID starting with 6d1a4ba4e83318e36e4e64798dd3fc0aeb5eb01e46ebad55b430926613a3a92d not found: ID does not exist" Nov 26 01:35:32 crc kubenswrapper[4766]: I1126 01:35:32.253513 4766 scope.go:117] "RemoveContainer" containerID="64d3767ccaf134301962316ec76fae5d44dd88c95bf4d6391b88cadb91965adf" Nov 26 01:35:32 crc kubenswrapper[4766]: E1126 01:35:32.255058 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64d3767ccaf134301962316ec76fae5d44dd88c95bf4d6391b88cadb91965adf\": container with ID starting with 64d3767ccaf134301962316ec76fae5d44dd88c95bf4d6391b88cadb91965adf not found: ID does not exist" containerID="64d3767ccaf134301962316ec76fae5d44dd88c95bf4d6391b88cadb91965adf" Nov 26 01:35:32 crc kubenswrapper[4766]: I1126 01:35:32.255272 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64d3767ccaf134301962316ec76fae5d44dd88c95bf4d6391b88cadb91965adf"} err="failed to get container status \"64d3767ccaf134301962316ec76fae5d44dd88c95bf4d6391b88cadb91965adf\": rpc error: code = NotFound desc = could not find container \"64d3767ccaf134301962316ec76fae5d44dd88c95bf4d6391b88cadb91965adf\": container with ID starting with 64d3767ccaf134301962316ec76fae5d44dd88c95bf4d6391b88cadb91965adf not found: ID does not exist" Nov 26 01:35:32 crc kubenswrapper[4766]: I1126 01:35:32.298748 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8h7zp"] Nov 26 01:35:33 crc kubenswrapper[4766]: I1126 01:35:33.163136 4766 generic.go:334] "Generic (PLEG): container finished" podID="0bc01db8-70da-4234-9c0b-247908a59973" containerID="0a33efb6c4241616bb1bb7bd73f5a07dc8dbdee77ccb66ceaf31053785ac6a3c" exitCode=0 Nov 26 01:35:33 crc kubenswrapper[4766]: I1126 01:35:33.163600 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8h7zp" event={"ID":"0bc01db8-70da-4234-9c0b-247908a59973","Type":"ContainerDied","Data":"0a33efb6c4241616bb1bb7bd73f5a07dc8dbdee77ccb66ceaf31053785ac6a3c"} Nov 26 01:35:33 crc kubenswrapper[4766]: I1126 01:35:33.163644 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8h7zp" event={"ID":"0bc01db8-70da-4234-9c0b-247908a59973","Type":"ContainerStarted","Data":"0e09c1990e0a9a99480412cad1c4c512eedab44d19cbd0c91da6bbdd996ef6dc"} Nov 26 01:35:33 crc kubenswrapper[4766]: I1126 01:35:33.858916 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03c0dd64-5cd2-4ec9-add0-45d28db2e1e0" path="/var/lib/kubelet/pods/03c0dd64-5cd2-4ec9-add0-45d28db2e1e0/volumes" Nov 26 01:35:34 crc kubenswrapper[4766]: I1126 01:35:34.181108 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8h7zp" event={"ID":"0bc01db8-70da-4234-9c0b-247908a59973","Type":"ContainerStarted","Data":"68d26cece3de1aa4a0b3bb20b50daf58eab631e178166cef61221319caba9d4f"} Nov 26 01:35:35 crc kubenswrapper[4766]: I1126 01:35:35.199472 4766 generic.go:334] "Generic (PLEG): container finished" podID="0bc01db8-70da-4234-9c0b-247908a59973" containerID="68d26cece3de1aa4a0b3bb20b50daf58eab631e178166cef61221319caba9d4f" exitCode=0 Nov 26 01:35:35 crc kubenswrapper[4766]: I1126 01:35:35.199537 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8h7zp" event={"ID":"0bc01db8-70da-4234-9c0b-247908a59973","Type":"ContainerDied","Data":"68d26cece3de1aa4a0b3bb20b50daf58eab631e178166cef61221319caba9d4f"} Nov 26 01:35:36 crc kubenswrapper[4766]: I1126 01:35:36.215922 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8h7zp" event={"ID":"0bc01db8-70da-4234-9c0b-247908a59973","Type":"ContainerStarted","Data":"9bfde96554995ead8cd320e3273b4602803144dfbb61483a44aa21a8eaaa97ba"} Nov 26 01:35:36 crc kubenswrapper[4766]: I1126 01:35:36.266202 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8h7zp" podStartSLOduration=2.813319484 podStartE2EDuration="5.266180197s" podCreationTimestamp="2025-11-26 01:35:31 +0000 UTC" firstStartedPulling="2025-11-26 01:35:33.165558324 +0000 UTC m=+4314.014328754" lastFinishedPulling="2025-11-26 01:35:35.618419027 +0000 UTC m=+4316.467189467" observedRunningTime="2025-11-26 01:35:36.245776548 +0000 UTC m=+4317.094546988" watchObservedRunningTime="2025-11-26 01:35:36.266180197 +0000 UTC m=+4317.114950627" Nov 26 01:35:41 crc kubenswrapper[4766]: I1126 01:35:41.479984 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:35:41 crc kubenswrapper[4766]: I1126 01:35:41.480874 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:35:41 crc kubenswrapper[4766]: I1126 01:35:41.480964 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 01:35:41 crc kubenswrapper[4766]: I1126 01:35:41.482300 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 01:35:41 crc kubenswrapper[4766]: I1126 01:35:41.482408 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" gracePeriod=600 Nov 26 01:35:41 crc kubenswrapper[4766]: E1126 01:35:41.612545 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:35:41 crc kubenswrapper[4766]: I1126 01:35:41.820555 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8h7zp" Nov 26 01:35:41 crc kubenswrapper[4766]: I1126 01:35:41.820617 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8h7zp" Nov 26 01:35:41 crc kubenswrapper[4766]: I1126 01:35:41.903090 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8h7zp" Nov 26 01:35:42 crc kubenswrapper[4766]: I1126 01:35:42.285964 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" exitCode=0 Nov 26 01:35:42 crc kubenswrapper[4766]: I1126 01:35:42.286032 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657"} Nov 26 01:35:42 crc kubenswrapper[4766]: I1126 01:35:42.286087 4766 scope.go:117] "RemoveContainer" containerID="53c28fab7d2dc492443a92908357f7cd25e699dec76a6009866ab9bb91349133" Nov 26 01:35:42 crc kubenswrapper[4766]: I1126 01:35:42.286905 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:35:42 crc kubenswrapper[4766]: E1126 01:35:42.287346 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:35:42 crc kubenswrapper[4766]: I1126 01:35:42.364959 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8h7zp" Nov 26 01:35:42 crc kubenswrapper[4766]: I1126 01:35:42.425027 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8h7zp"] Nov 26 01:35:44 crc kubenswrapper[4766]: I1126 01:35:44.311829 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8h7zp" podUID="0bc01db8-70da-4234-9c0b-247908a59973" containerName="registry-server" containerID="cri-o://9bfde96554995ead8cd320e3273b4602803144dfbb61483a44aa21a8eaaa97ba" gracePeriod=2 Nov 26 01:35:44 crc kubenswrapper[4766]: I1126 01:35:44.816022 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8h7zp" Nov 26 01:35:44 crc kubenswrapper[4766]: I1126 01:35:44.970568 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bc01db8-70da-4234-9c0b-247908a59973-utilities\") pod \"0bc01db8-70da-4234-9c0b-247908a59973\" (UID: \"0bc01db8-70da-4234-9c0b-247908a59973\") " Nov 26 01:35:44 crc kubenswrapper[4766]: I1126 01:35:44.971217 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gz9dt\" (UniqueName: \"kubernetes.io/projected/0bc01db8-70da-4234-9c0b-247908a59973-kube-api-access-gz9dt\") pod \"0bc01db8-70da-4234-9c0b-247908a59973\" (UID: \"0bc01db8-70da-4234-9c0b-247908a59973\") " Nov 26 01:35:44 crc kubenswrapper[4766]: I1126 01:35:44.971285 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bc01db8-70da-4234-9c0b-247908a59973-catalog-content\") pod \"0bc01db8-70da-4234-9c0b-247908a59973\" (UID: \"0bc01db8-70da-4234-9c0b-247908a59973\") " Nov 26 01:35:44 crc kubenswrapper[4766]: I1126 01:35:44.972205 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bc01db8-70da-4234-9c0b-247908a59973-utilities" (OuterVolumeSpecName: "utilities") pod "0bc01db8-70da-4234-9c0b-247908a59973" (UID: "0bc01db8-70da-4234-9c0b-247908a59973"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:35:44 crc kubenswrapper[4766]: I1126 01:35:44.972758 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bc01db8-70da-4234-9c0b-247908a59973-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 01:35:44 crc kubenswrapper[4766]: I1126 01:35:44.976999 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bc01db8-70da-4234-9c0b-247908a59973-kube-api-access-gz9dt" (OuterVolumeSpecName: "kube-api-access-gz9dt") pod "0bc01db8-70da-4234-9c0b-247908a59973" (UID: "0bc01db8-70da-4234-9c0b-247908a59973"). InnerVolumeSpecName "kube-api-access-gz9dt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:35:44 crc kubenswrapper[4766]: I1126 01:35:44.989788 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bc01db8-70da-4234-9c0b-247908a59973-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0bc01db8-70da-4234-9c0b-247908a59973" (UID: "0bc01db8-70da-4234-9c0b-247908a59973"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:35:45 crc kubenswrapper[4766]: I1126 01:35:45.075920 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gz9dt\" (UniqueName: \"kubernetes.io/projected/0bc01db8-70da-4234-9c0b-247908a59973-kube-api-access-gz9dt\") on node \"crc\" DevicePath \"\"" Nov 26 01:35:45 crc kubenswrapper[4766]: I1126 01:35:45.075971 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bc01db8-70da-4234-9c0b-247908a59973-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 01:35:45 crc kubenswrapper[4766]: I1126 01:35:45.327109 4766 generic.go:334] "Generic (PLEG): container finished" podID="0bc01db8-70da-4234-9c0b-247908a59973" containerID="9bfde96554995ead8cd320e3273b4602803144dfbb61483a44aa21a8eaaa97ba" exitCode=0 Nov 26 01:35:45 crc kubenswrapper[4766]: I1126 01:35:45.327175 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8h7zp" event={"ID":"0bc01db8-70da-4234-9c0b-247908a59973","Type":"ContainerDied","Data":"9bfde96554995ead8cd320e3273b4602803144dfbb61483a44aa21a8eaaa97ba"} Nov 26 01:35:45 crc kubenswrapper[4766]: I1126 01:35:45.327217 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8h7zp" event={"ID":"0bc01db8-70da-4234-9c0b-247908a59973","Type":"ContainerDied","Data":"0e09c1990e0a9a99480412cad1c4c512eedab44d19cbd0c91da6bbdd996ef6dc"} Nov 26 01:35:45 crc kubenswrapper[4766]: I1126 01:35:45.327245 4766 scope.go:117] "RemoveContainer" containerID="9bfde96554995ead8cd320e3273b4602803144dfbb61483a44aa21a8eaaa97ba" Nov 26 01:35:45 crc kubenswrapper[4766]: I1126 01:35:45.327477 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8h7zp" Nov 26 01:35:45 crc kubenswrapper[4766]: I1126 01:35:45.361308 4766 scope.go:117] "RemoveContainer" containerID="68d26cece3de1aa4a0b3bb20b50daf58eab631e178166cef61221319caba9d4f" Nov 26 01:35:45 crc kubenswrapper[4766]: I1126 01:35:45.395845 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8h7zp"] Nov 26 01:35:45 crc kubenswrapper[4766]: I1126 01:35:45.406895 4766 scope.go:117] "RemoveContainer" containerID="0a33efb6c4241616bb1bb7bd73f5a07dc8dbdee77ccb66ceaf31053785ac6a3c" Nov 26 01:35:45 crc kubenswrapper[4766]: I1126 01:35:45.415200 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8h7zp"] Nov 26 01:35:45 crc kubenswrapper[4766]: I1126 01:35:45.452927 4766 scope.go:117] "RemoveContainer" containerID="9bfde96554995ead8cd320e3273b4602803144dfbb61483a44aa21a8eaaa97ba" Nov 26 01:35:45 crc kubenswrapper[4766]: E1126 01:35:45.453334 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bfde96554995ead8cd320e3273b4602803144dfbb61483a44aa21a8eaaa97ba\": container with ID starting with 9bfde96554995ead8cd320e3273b4602803144dfbb61483a44aa21a8eaaa97ba not found: ID does not exist" containerID="9bfde96554995ead8cd320e3273b4602803144dfbb61483a44aa21a8eaaa97ba" Nov 26 01:35:45 crc kubenswrapper[4766]: I1126 01:35:45.453365 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bfde96554995ead8cd320e3273b4602803144dfbb61483a44aa21a8eaaa97ba"} err="failed to get container status \"9bfde96554995ead8cd320e3273b4602803144dfbb61483a44aa21a8eaaa97ba\": rpc error: code = NotFound desc = could not find container \"9bfde96554995ead8cd320e3273b4602803144dfbb61483a44aa21a8eaaa97ba\": container with ID starting with 9bfde96554995ead8cd320e3273b4602803144dfbb61483a44aa21a8eaaa97ba not found: ID does not exist" Nov 26 01:35:45 crc kubenswrapper[4766]: I1126 01:35:45.453382 4766 scope.go:117] "RemoveContainer" containerID="68d26cece3de1aa4a0b3bb20b50daf58eab631e178166cef61221319caba9d4f" Nov 26 01:35:45 crc kubenswrapper[4766]: E1126 01:35:45.453893 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68d26cece3de1aa4a0b3bb20b50daf58eab631e178166cef61221319caba9d4f\": container with ID starting with 68d26cece3de1aa4a0b3bb20b50daf58eab631e178166cef61221319caba9d4f not found: ID does not exist" containerID="68d26cece3de1aa4a0b3bb20b50daf58eab631e178166cef61221319caba9d4f" Nov 26 01:35:45 crc kubenswrapper[4766]: I1126 01:35:45.453958 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68d26cece3de1aa4a0b3bb20b50daf58eab631e178166cef61221319caba9d4f"} err="failed to get container status \"68d26cece3de1aa4a0b3bb20b50daf58eab631e178166cef61221319caba9d4f\": rpc error: code = NotFound desc = could not find container \"68d26cece3de1aa4a0b3bb20b50daf58eab631e178166cef61221319caba9d4f\": container with ID starting with 68d26cece3de1aa4a0b3bb20b50daf58eab631e178166cef61221319caba9d4f not found: ID does not exist" Nov 26 01:35:45 crc kubenswrapper[4766]: I1126 01:35:45.454001 4766 scope.go:117] "RemoveContainer" containerID="0a33efb6c4241616bb1bb7bd73f5a07dc8dbdee77ccb66ceaf31053785ac6a3c" Nov 26 01:35:45 crc kubenswrapper[4766]: E1126 01:35:45.454277 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a33efb6c4241616bb1bb7bd73f5a07dc8dbdee77ccb66ceaf31053785ac6a3c\": container with ID starting with 0a33efb6c4241616bb1bb7bd73f5a07dc8dbdee77ccb66ceaf31053785ac6a3c not found: ID does not exist" containerID="0a33efb6c4241616bb1bb7bd73f5a07dc8dbdee77ccb66ceaf31053785ac6a3c" Nov 26 01:35:45 crc kubenswrapper[4766]: I1126 01:35:45.454302 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a33efb6c4241616bb1bb7bd73f5a07dc8dbdee77ccb66ceaf31053785ac6a3c"} err="failed to get container status \"0a33efb6c4241616bb1bb7bd73f5a07dc8dbdee77ccb66ceaf31053785ac6a3c\": rpc error: code = NotFound desc = could not find container \"0a33efb6c4241616bb1bb7bd73f5a07dc8dbdee77ccb66ceaf31053785ac6a3c\": container with ID starting with 0a33efb6c4241616bb1bb7bd73f5a07dc8dbdee77ccb66ceaf31053785ac6a3c not found: ID does not exist" Nov 26 01:35:45 crc kubenswrapper[4766]: I1126 01:35:45.841427 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bc01db8-70da-4234-9c0b-247908a59973" path="/var/lib/kubelet/pods/0bc01db8-70da-4234-9c0b-247908a59973/volumes" Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.009907 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hws2v"] Nov 26 01:35:49 crc kubenswrapper[4766]: E1126 01:35:49.011084 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bc01db8-70da-4234-9c0b-247908a59973" containerName="registry-server" Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.011105 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bc01db8-70da-4234-9c0b-247908a59973" containerName="registry-server" Nov 26 01:35:49 crc kubenswrapper[4766]: E1126 01:35:49.011157 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bc01db8-70da-4234-9c0b-247908a59973" containerName="extract-utilities" Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.011167 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bc01db8-70da-4234-9c0b-247908a59973" containerName="extract-utilities" Nov 26 01:35:49 crc kubenswrapper[4766]: E1126 01:35:49.011195 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03c0dd64-5cd2-4ec9-add0-45d28db2e1e0" containerName="extract-content" Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.011204 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="03c0dd64-5cd2-4ec9-add0-45d28db2e1e0" containerName="extract-content" Nov 26 01:35:49 crc kubenswrapper[4766]: E1126 01:35:49.011220 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bc01db8-70da-4234-9c0b-247908a59973" containerName="extract-content" Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.011228 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bc01db8-70da-4234-9c0b-247908a59973" containerName="extract-content" Nov 26 01:35:49 crc kubenswrapper[4766]: E1126 01:35:49.011241 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03c0dd64-5cd2-4ec9-add0-45d28db2e1e0" containerName="registry-server" Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.011248 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="03c0dd64-5cd2-4ec9-add0-45d28db2e1e0" containerName="registry-server" Nov 26 01:35:49 crc kubenswrapper[4766]: E1126 01:35:49.011262 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03c0dd64-5cd2-4ec9-add0-45d28db2e1e0" containerName="extract-utilities" Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.011268 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="03c0dd64-5cd2-4ec9-add0-45d28db2e1e0" containerName="extract-utilities" Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.011514 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bc01db8-70da-4234-9c0b-247908a59973" containerName="registry-server" Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.011544 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="03c0dd64-5cd2-4ec9-add0-45d28db2e1e0" containerName="registry-server" Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.015342 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hws2v" Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.038433 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hws2v"] Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.177916 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e883b02b-3590-4212-a526-b1e9c916b3c7-catalog-content\") pod \"redhat-operators-hws2v\" (UID: \"e883b02b-3590-4212-a526-b1e9c916b3c7\") " pod="openshift-marketplace/redhat-operators-hws2v" Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.177964 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e883b02b-3590-4212-a526-b1e9c916b3c7-utilities\") pod \"redhat-operators-hws2v\" (UID: \"e883b02b-3590-4212-a526-b1e9c916b3c7\") " pod="openshift-marketplace/redhat-operators-hws2v" Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.178022 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmlhf\" (UniqueName: \"kubernetes.io/projected/e883b02b-3590-4212-a526-b1e9c916b3c7-kube-api-access-mmlhf\") pod \"redhat-operators-hws2v\" (UID: \"e883b02b-3590-4212-a526-b1e9c916b3c7\") " pod="openshift-marketplace/redhat-operators-hws2v" Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.279369 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e883b02b-3590-4212-a526-b1e9c916b3c7-catalog-content\") pod \"redhat-operators-hws2v\" (UID: \"e883b02b-3590-4212-a526-b1e9c916b3c7\") " pod="openshift-marketplace/redhat-operators-hws2v" Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.279403 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e883b02b-3590-4212-a526-b1e9c916b3c7-utilities\") pod \"redhat-operators-hws2v\" (UID: \"e883b02b-3590-4212-a526-b1e9c916b3c7\") " pod="openshift-marketplace/redhat-operators-hws2v" Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.279463 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmlhf\" (UniqueName: \"kubernetes.io/projected/e883b02b-3590-4212-a526-b1e9c916b3c7-kube-api-access-mmlhf\") pod \"redhat-operators-hws2v\" (UID: \"e883b02b-3590-4212-a526-b1e9c916b3c7\") " pod="openshift-marketplace/redhat-operators-hws2v" Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.279999 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e883b02b-3590-4212-a526-b1e9c916b3c7-catalog-content\") pod \"redhat-operators-hws2v\" (UID: \"e883b02b-3590-4212-a526-b1e9c916b3c7\") " pod="openshift-marketplace/redhat-operators-hws2v" Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.280058 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e883b02b-3590-4212-a526-b1e9c916b3c7-utilities\") pod \"redhat-operators-hws2v\" (UID: \"e883b02b-3590-4212-a526-b1e9c916b3c7\") " pod="openshift-marketplace/redhat-operators-hws2v" Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.300295 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmlhf\" (UniqueName: \"kubernetes.io/projected/e883b02b-3590-4212-a526-b1e9c916b3c7-kube-api-access-mmlhf\") pod \"redhat-operators-hws2v\" (UID: \"e883b02b-3590-4212-a526-b1e9c916b3c7\") " pod="openshift-marketplace/redhat-operators-hws2v" Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.341712 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hws2v" Nov 26 01:35:49 crc kubenswrapper[4766]: I1126 01:35:49.819234 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hws2v"] Nov 26 01:35:50 crc kubenswrapper[4766]: I1126 01:35:50.378324 4766 generic.go:334] "Generic (PLEG): container finished" podID="e883b02b-3590-4212-a526-b1e9c916b3c7" containerID="881534ee1f296d90b545b3e461427bfb58098299c5292fa013a1096afe89e456" exitCode=0 Nov 26 01:35:50 crc kubenswrapper[4766]: I1126 01:35:50.378384 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hws2v" event={"ID":"e883b02b-3590-4212-a526-b1e9c916b3c7","Type":"ContainerDied","Data":"881534ee1f296d90b545b3e461427bfb58098299c5292fa013a1096afe89e456"} Nov 26 01:35:50 crc kubenswrapper[4766]: I1126 01:35:50.378619 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hws2v" event={"ID":"e883b02b-3590-4212-a526-b1e9c916b3c7","Type":"ContainerStarted","Data":"dbbec034ed041f805b83bcc7d1ca768a427319392553ef9cf590f5ae9e0bc0ee"} Nov 26 01:35:51 crc kubenswrapper[4766]: I1126 01:35:51.403593 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hws2v" event={"ID":"e883b02b-3590-4212-a526-b1e9c916b3c7","Type":"ContainerStarted","Data":"05f5a925e29b4ec5a96111203508e7b1c361ef801d7ed1852c1ac3bba745c4b4"} Nov 26 01:35:54 crc kubenswrapper[4766]: I1126 01:35:54.826816 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:35:54 crc kubenswrapper[4766]: E1126 01:35:54.827715 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:35:55 crc kubenswrapper[4766]: I1126 01:35:55.456983 4766 generic.go:334] "Generic (PLEG): container finished" podID="e883b02b-3590-4212-a526-b1e9c916b3c7" containerID="05f5a925e29b4ec5a96111203508e7b1c361ef801d7ed1852c1ac3bba745c4b4" exitCode=0 Nov 26 01:35:55 crc kubenswrapper[4766]: I1126 01:35:55.457068 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hws2v" event={"ID":"e883b02b-3590-4212-a526-b1e9c916b3c7","Type":"ContainerDied","Data":"05f5a925e29b4ec5a96111203508e7b1c361ef801d7ed1852c1ac3bba745c4b4"} Nov 26 01:35:56 crc kubenswrapper[4766]: I1126 01:35:56.469183 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hws2v" event={"ID":"e883b02b-3590-4212-a526-b1e9c916b3c7","Type":"ContainerStarted","Data":"fe955e8f995733d2d214e098f2221dafdbeade9c4349fc34b60acc36186a4f2a"} Nov 26 01:35:56 crc kubenswrapper[4766]: I1126 01:35:56.532241 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hws2v" podStartSLOduration=3.015723688 podStartE2EDuration="8.53222424s" podCreationTimestamp="2025-11-26 01:35:48 +0000 UTC" firstStartedPulling="2025-11-26 01:35:50.379892649 +0000 UTC m=+4331.228663079" lastFinishedPulling="2025-11-26 01:35:55.896393171 +0000 UTC m=+4336.745163631" observedRunningTime="2025-11-26 01:35:56.522221285 +0000 UTC m=+4337.370991725" watchObservedRunningTime="2025-11-26 01:35:56.53222424 +0000 UTC m=+4337.380994670" Nov 26 01:35:59 crc kubenswrapper[4766]: I1126 01:35:59.342062 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hws2v" Nov 26 01:35:59 crc kubenswrapper[4766]: I1126 01:35:59.342543 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hws2v" Nov 26 01:36:00 crc kubenswrapper[4766]: I1126 01:36:00.395373 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hws2v" podUID="e883b02b-3590-4212-a526-b1e9c916b3c7" containerName="registry-server" probeResult="failure" output=< Nov 26 01:36:00 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 01:36:00 crc kubenswrapper[4766]: > Nov 26 01:36:04 crc kubenswrapper[4766]: I1126 01:36:04.122211 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-756fcfc485-ktbfm" podUID="8d171003-64ac-4108-957b-70f4225359c5" containerName="neutron-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Nov 26 01:36:05 crc kubenswrapper[4766]: I1126 01:36:05.826923 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:36:05 crc kubenswrapper[4766]: E1126 01:36:05.827643 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:36:09 crc kubenswrapper[4766]: I1126 01:36:09.417669 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hws2v" Nov 26 01:36:09 crc kubenswrapper[4766]: I1126 01:36:09.499971 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hws2v" Nov 26 01:36:09 crc kubenswrapper[4766]: I1126 01:36:09.665073 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hws2v"] Nov 26 01:36:10 crc kubenswrapper[4766]: I1126 01:36:10.662549 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hws2v" podUID="e883b02b-3590-4212-a526-b1e9c916b3c7" containerName="registry-server" containerID="cri-o://fe955e8f995733d2d214e098f2221dafdbeade9c4349fc34b60acc36186a4f2a" gracePeriod=2 Nov 26 01:36:11 crc kubenswrapper[4766]: I1126 01:36:11.227606 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hws2v" Nov 26 01:36:11 crc kubenswrapper[4766]: I1126 01:36:11.302950 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmlhf\" (UniqueName: \"kubernetes.io/projected/e883b02b-3590-4212-a526-b1e9c916b3c7-kube-api-access-mmlhf\") pod \"e883b02b-3590-4212-a526-b1e9c916b3c7\" (UID: \"e883b02b-3590-4212-a526-b1e9c916b3c7\") " Nov 26 01:36:11 crc kubenswrapper[4766]: I1126 01:36:11.303062 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e883b02b-3590-4212-a526-b1e9c916b3c7-utilities\") pod \"e883b02b-3590-4212-a526-b1e9c916b3c7\" (UID: \"e883b02b-3590-4212-a526-b1e9c916b3c7\") " Nov 26 01:36:11 crc kubenswrapper[4766]: I1126 01:36:11.303097 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e883b02b-3590-4212-a526-b1e9c916b3c7-catalog-content\") pod \"e883b02b-3590-4212-a526-b1e9c916b3c7\" (UID: \"e883b02b-3590-4212-a526-b1e9c916b3c7\") " Nov 26 01:36:11 crc kubenswrapper[4766]: I1126 01:36:11.308023 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e883b02b-3590-4212-a526-b1e9c916b3c7-utilities" (OuterVolumeSpecName: "utilities") pod "e883b02b-3590-4212-a526-b1e9c916b3c7" (UID: "e883b02b-3590-4212-a526-b1e9c916b3c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:36:11 crc kubenswrapper[4766]: I1126 01:36:11.406234 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e883b02b-3590-4212-a526-b1e9c916b3c7-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 01:36:11 crc kubenswrapper[4766]: I1126 01:36:11.419258 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e883b02b-3590-4212-a526-b1e9c916b3c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e883b02b-3590-4212-a526-b1e9c916b3c7" (UID: "e883b02b-3590-4212-a526-b1e9c916b3c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:36:11 crc kubenswrapper[4766]: I1126 01:36:11.508336 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e883b02b-3590-4212-a526-b1e9c916b3c7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 01:36:11 crc kubenswrapper[4766]: I1126 01:36:11.674945 4766 generic.go:334] "Generic (PLEG): container finished" podID="e883b02b-3590-4212-a526-b1e9c916b3c7" containerID="fe955e8f995733d2d214e098f2221dafdbeade9c4349fc34b60acc36186a4f2a" exitCode=0 Nov 26 01:36:11 crc kubenswrapper[4766]: I1126 01:36:11.674995 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hws2v" event={"ID":"e883b02b-3590-4212-a526-b1e9c916b3c7","Type":"ContainerDied","Data":"fe955e8f995733d2d214e098f2221dafdbeade9c4349fc34b60acc36186a4f2a"} Nov 26 01:36:11 crc kubenswrapper[4766]: I1126 01:36:11.675035 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hws2v" event={"ID":"e883b02b-3590-4212-a526-b1e9c916b3c7","Type":"ContainerDied","Data":"dbbec034ed041f805b83bcc7d1ca768a427319392553ef9cf590f5ae9e0bc0ee"} Nov 26 01:36:11 crc kubenswrapper[4766]: I1126 01:36:11.675056 4766 scope.go:117] "RemoveContainer" containerID="fe955e8f995733d2d214e098f2221dafdbeade9c4349fc34b60acc36186a4f2a" Nov 26 01:36:11 crc kubenswrapper[4766]: I1126 01:36:11.675058 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hws2v" Nov 26 01:36:11 crc kubenswrapper[4766]: I1126 01:36:11.705450 4766 scope.go:117] "RemoveContainer" containerID="05f5a925e29b4ec5a96111203508e7b1c361ef801d7ed1852c1ac3bba745c4b4" Nov 26 01:36:12 crc kubenswrapper[4766]: I1126 01:36:12.006114 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e883b02b-3590-4212-a526-b1e9c916b3c7-kube-api-access-mmlhf" (OuterVolumeSpecName: "kube-api-access-mmlhf") pod "e883b02b-3590-4212-a526-b1e9c916b3c7" (UID: "e883b02b-3590-4212-a526-b1e9c916b3c7"). InnerVolumeSpecName "kube-api-access-mmlhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:36:12 crc kubenswrapper[4766]: I1126 01:36:12.023435 4766 scope.go:117] "RemoveContainer" containerID="881534ee1f296d90b545b3e461427bfb58098299c5292fa013a1096afe89e456" Nov 26 01:36:12 crc kubenswrapper[4766]: I1126 01:36:12.023434 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmlhf\" (UniqueName: \"kubernetes.io/projected/e883b02b-3590-4212-a526-b1e9c916b3c7-kube-api-access-mmlhf\") on node \"crc\" DevicePath \"\"" Nov 26 01:36:12 crc kubenswrapper[4766]: I1126 01:36:12.225772 4766 scope.go:117] "RemoveContainer" containerID="fe955e8f995733d2d214e098f2221dafdbeade9c4349fc34b60acc36186a4f2a" Nov 26 01:36:12 crc kubenswrapper[4766]: E1126 01:36:12.226128 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe955e8f995733d2d214e098f2221dafdbeade9c4349fc34b60acc36186a4f2a\": container with ID starting with fe955e8f995733d2d214e098f2221dafdbeade9c4349fc34b60acc36186a4f2a not found: ID does not exist" containerID="fe955e8f995733d2d214e098f2221dafdbeade9c4349fc34b60acc36186a4f2a" Nov 26 01:36:12 crc kubenswrapper[4766]: I1126 01:36:12.226159 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe955e8f995733d2d214e098f2221dafdbeade9c4349fc34b60acc36186a4f2a"} err="failed to get container status \"fe955e8f995733d2d214e098f2221dafdbeade9c4349fc34b60acc36186a4f2a\": rpc error: code = NotFound desc = could not find container \"fe955e8f995733d2d214e098f2221dafdbeade9c4349fc34b60acc36186a4f2a\": container with ID starting with fe955e8f995733d2d214e098f2221dafdbeade9c4349fc34b60acc36186a4f2a not found: ID does not exist" Nov 26 01:36:12 crc kubenswrapper[4766]: I1126 01:36:12.226175 4766 scope.go:117] "RemoveContainer" containerID="05f5a925e29b4ec5a96111203508e7b1c361ef801d7ed1852c1ac3bba745c4b4" Nov 26 01:36:12 crc kubenswrapper[4766]: E1126 01:36:12.228537 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05f5a925e29b4ec5a96111203508e7b1c361ef801d7ed1852c1ac3bba745c4b4\": container with ID starting with 05f5a925e29b4ec5a96111203508e7b1c361ef801d7ed1852c1ac3bba745c4b4 not found: ID does not exist" containerID="05f5a925e29b4ec5a96111203508e7b1c361ef801d7ed1852c1ac3bba745c4b4" Nov 26 01:36:12 crc kubenswrapper[4766]: I1126 01:36:12.228584 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05f5a925e29b4ec5a96111203508e7b1c361ef801d7ed1852c1ac3bba745c4b4"} err="failed to get container status \"05f5a925e29b4ec5a96111203508e7b1c361ef801d7ed1852c1ac3bba745c4b4\": rpc error: code = NotFound desc = could not find container \"05f5a925e29b4ec5a96111203508e7b1c361ef801d7ed1852c1ac3bba745c4b4\": container with ID starting with 05f5a925e29b4ec5a96111203508e7b1c361ef801d7ed1852c1ac3bba745c4b4 not found: ID does not exist" Nov 26 01:36:12 crc kubenswrapper[4766]: I1126 01:36:12.228613 4766 scope.go:117] "RemoveContainer" containerID="881534ee1f296d90b545b3e461427bfb58098299c5292fa013a1096afe89e456" Nov 26 01:36:12 crc kubenswrapper[4766]: E1126 01:36:12.229112 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"881534ee1f296d90b545b3e461427bfb58098299c5292fa013a1096afe89e456\": container with ID starting with 881534ee1f296d90b545b3e461427bfb58098299c5292fa013a1096afe89e456 not found: ID does not exist" containerID="881534ee1f296d90b545b3e461427bfb58098299c5292fa013a1096afe89e456" Nov 26 01:36:12 crc kubenswrapper[4766]: I1126 01:36:12.229157 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"881534ee1f296d90b545b3e461427bfb58098299c5292fa013a1096afe89e456"} err="failed to get container status \"881534ee1f296d90b545b3e461427bfb58098299c5292fa013a1096afe89e456\": rpc error: code = NotFound desc = could not find container \"881534ee1f296d90b545b3e461427bfb58098299c5292fa013a1096afe89e456\": container with ID starting with 881534ee1f296d90b545b3e461427bfb58098299c5292fa013a1096afe89e456 not found: ID does not exist" Nov 26 01:36:12 crc kubenswrapper[4766]: I1126 01:36:12.321069 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hws2v"] Nov 26 01:36:12 crc kubenswrapper[4766]: I1126 01:36:12.337627 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hws2v"] Nov 26 01:36:13 crc kubenswrapper[4766]: I1126 01:36:13.859915 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e883b02b-3590-4212-a526-b1e9c916b3c7" path="/var/lib/kubelet/pods/e883b02b-3590-4212-a526-b1e9c916b3c7/volumes" Nov 26 01:36:20 crc kubenswrapper[4766]: I1126 01:36:20.827939 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:36:20 crc kubenswrapper[4766]: E1126 01:36:20.830216 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:36:31 crc kubenswrapper[4766]: I1126 01:36:31.828012 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:36:31 crc kubenswrapper[4766]: E1126 01:36:31.828878 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:36:44 crc kubenswrapper[4766]: I1126 01:36:44.827509 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:36:44 crc kubenswrapper[4766]: E1126 01:36:44.828268 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:36:55 crc kubenswrapper[4766]: I1126 01:36:55.827470 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:36:55 crc kubenswrapper[4766]: E1126 01:36:55.830341 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:37:05 crc kubenswrapper[4766]: I1126 01:37:05.389081 4766 generic.go:334] "Generic (PLEG): container finished" podID="8da0c4a5-4c26-46de-8408-a8b562ba8f76" containerID="4ecbd58c6bd96e44ea7745ce725b13265105c7ac1c1a8350cfe5c801ce3499f4" exitCode=0 Nov 26 01:37:05 crc kubenswrapper[4766]: I1126 01:37:05.389734 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" event={"ID":"8da0c4a5-4c26-46de-8408-a8b562ba8f76","Type":"ContainerDied","Data":"4ecbd58c6bd96e44ea7745ce725b13265105c7ac1c1a8350cfe5c801ce3499f4"} Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.000130 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.033014 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-migration-ssh-key-1\") pod \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.033219 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-extra-config-0\") pod \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.033429 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-inventory\") pod \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.034597 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-cell1-compute-config-1\") pod \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.035256 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-custom-ceph-combined-ca-bundle\") pod \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.035314 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlhgm\" (UniqueName: \"kubernetes.io/projected/8da0c4a5-4c26-46de-8408-a8b562ba8f76-kube-api-access-wlhgm\") pod \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.035571 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-migration-ssh-key-0\") pod \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.035620 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-ssh-key\") pod \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.035676 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-ceph\") pod \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.035697 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/8da0c4a5-4c26-46de-8408-a8b562ba8f76-ceph-nova-0\") pod \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.035783 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-cell1-compute-config-0\") pod \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\" (UID: \"8da0c4a5-4c26-46de-8408-a8b562ba8f76\") " Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.041714 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8da0c4a5-4c26-46de-8408-a8b562ba8f76-kube-api-access-wlhgm" (OuterVolumeSpecName: "kube-api-access-wlhgm") pod "8da0c4a5-4c26-46de-8408-a8b562ba8f76" (UID: "8da0c4a5-4c26-46de-8408-a8b562ba8f76"). InnerVolumeSpecName "kube-api-access-wlhgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.047094 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "8da0c4a5-4c26-46de-8408-a8b562ba8f76" (UID: "8da0c4a5-4c26-46de-8408-a8b562ba8f76"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.065149 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "8da0c4a5-4c26-46de-8408-a8b562ba8f76" (UID: "8da0c4a5-4c26-46de-8408-a8b562ba8f76"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.068743 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "8da0c4a5-4c26-46de-8408-a8b562ba8f76" (UID: "8da0c4a5-4c26-46de-8408-a8b562ba8f76"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.077084 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-ceph" (OuterVolumeSpecName: "ceph") pod "8da0c4a5-4c26-46de-8408-a8b562ba8f76" (UID: "8da0c4a5-4c26-46de-8408-a8b562ba8f76"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.086717 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "8da0c4a5-4c26-46de-8408-a8b562ba8f76" (UID: "8da0c4a5-4c26-46de-8408-a8b562ba8f76"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.093013 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8da0c4a5-4c26-46de-8408-a8b562ba8f76" (UID: "8da0c4a5-4c26-46de-8408-a8b562ba8f76"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.098909 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "8da0c4a5-4c26-46de-8408-a8b562ba8f76" (UID: "8da0c4a5-4c26-46de-8408-a8b562ba8f76"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.099402 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8da0c4a5-4c26-46de-8408-a8b562ba8f76-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "8da0c4a5-4c26-46de-8408-a8b562ba8f76" (UID: "8da0c4a5-4c26-46de-8408-a8b562ba8f76"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.102829 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "8da0c4a5-4c26-46de-8408-a8b562ba8f76" (UID: "8da0c4a5-4c26-46de-8408-a8b562ba8f76"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.107203 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-inventory" (OuterVolumeSpecName: "inventory") pod "8da0c4a5-4c26-46de-8408-a8b562ba8f76" (UID: "8da0c4a5-4c26-46de-8408-a8b562ba8f76"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.138762 4766 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.138798 4766 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.138808 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.138819 4766 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.138828 4766 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.138838 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlhgm\" (UniqueName: \"kubernetes.io/projected/8da0c4a5-4c26-46de-8408-a8b562ba8f76-kube-api-access-wlhgm\") on node \"crc\" DevicePath \"\"" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.138846 4766 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.138855 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.138863 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.138870 4766 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/8da0c4a5-4c26-46de-8408-a8b562ba8f76-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.138880 4766 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/8da0c4a5-4c26-46de-8408-a8b562ba8f76-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.413618 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" event={"ID":"8da0c4a5-4c26-46de-8408-a8b562ba8f76","Type":"ContainerDied","Data":"d2a703a3ac62ed41996caad696bda01e5e5748dee7834b5bd41b703bfdccf466"} Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.414116 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2a703a3ac62ed41996caad696bda01e5e5748dee7834b5bd41b703bfdccf466" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.413684 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.564733 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx"] Nov 26 01:37:07 crc kubenswrapper[4766]: E1126 01:37:07.565430 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8da0c4a5-4c26-46de-8408-a8b562ba8f76" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.565464 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="8da0c4a5-4c26-46de-8408-a8b562ba8f76" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Nov 26 01:37:07 crc kubenswrapper[4766]: E1126 01:37:07.565523 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e883b02b-3590-4212-a526-b1e9c916b3c7" containerName="registry-server" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.565537 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e883b02b-3590-4212-a526-b1e9c916b3c7" containerName="registry-server" Nov 26 01:37:07 crc kubenswrapper[4766]: E1126 01:37:07.565589 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e883b02b-3590-4212-a526-b1e9c916b3c7" containerName="extract-content" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.565606 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e883b02b-3590-4212-a526-b1e9c916b3c7" containerName="extract-content" Nov 26 01:37:07 crc kubenswrapper[4766]: E1126 01:37:07.565644 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e883b02b-3590-4212-a526-b1e9c916b3c7" containerName="extract-utilities" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.565692 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e883b02b-3590-4212-a526-b1e9c916b3c7" containerName="extract-utilities" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.566132 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="e883b02b-3590-4212-a526-b1e9c916b3c7" containerName="registry-server" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.566174 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="8da0c4a5-4c26-46de-8408-a8b562ba8f76" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.567484 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.571137 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.571331 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.571708 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.571763 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.571843 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.574754 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.582257 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx"] Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.652352 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.652400 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.652435 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.652457 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceph\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.652524 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.652558 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.652607 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.652679 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wcdx\" (UniqueName: \"kubernetes.io/projected/eef488e4-2c28-4247-a8f4-b8e9008d518e-kube-api-access-9wcdx\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.755075 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.755214 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wcdx\" (UniqueName: \"kubernetes.io/projected/eef488e4-2c28-4247-a8f4-b8e9008d518e-kube-api-access-9wcdx\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.755315 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.755348 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.755396 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.755427 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceph\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.755469 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.755503 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.760796 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.761329 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.761432 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.762307 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.764153 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.766911 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.769346 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceph\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.780565 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wcdx\" (UniqueName: \"kubernetes.io/projected/eef488e4-2c28-4247-a8f4-b8e9008d518e-kube-api-access-9wcdx\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g6znx\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:07 crc kubenswrapper[4766]: I1126 01:37:07.897006 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:37:08 crc kubenswrapper[4766]: I1126 01:37:08.471981 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx"] Nov 26 01:37:08 crc kubenswrapper[4766]: I1126 01:37:08.826939 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:37:08 crc kubenswrapper[4766]: E1126 01:37:08.827366 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:37:09 crc kubenswrapper[4766]: I1126 01:37:09.440851 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" event={"ID":"eef488e4-2c28-4247-a8f4-b8e9008d518e","Type":"ContainerStarted","Data":"6cf98dd3de7c7de7fd48a38947f713d37b0e1446fb81ab33269d7274e49b3042"} Nov 26 01:37:09 crc kubenswrapper[4766]: I1126 01:37:09.441311 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" event={"ID":"eef488e4-2c28-4247-a8f4-b8e9008d518e","Type":"ContainerStarted","Data":"fa78979048843c160a7af051c231c9dbe2d79e6f6eefd5095204bcd291168a3d"} Nov 26 01:37:09 crc kubenswrapper[4766]: I1126 01:37:09.475797 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" podStartSLOduration=1.926140411 podStartE2EDuration="2.475776574s" podCreationTimestamp="2025-11-26 01:37:07 +0000 UTC" firstStartedPulling="2025-11-26 01:37:08.512902878 +0000 UTC m=+4409.361673348" lastFinishedPulling="2025-11-26 01:37:09.062539081 +0000 UTC m=+4409.911309511" observedRunningTime="2025-11-26 01:37:09.460912791 +0000 UTC m=+4410.309683231" watchObservedRunningTime="2025-11-26 01:37:09.475776574 +0000 UTC m=+4410.324547004" Nov 26 01:37:23 crc kubenswrapper[4766]: I1126 01:37:23.827167 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:37:23 crc kubenswrapper[4766]: E1126 01:37:23.827926 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:37:34 crc kubenswrapper[4766]: I1126 01:37:34.827007 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:37:34 crc kubenswrapper[4766]: E1126 01:37:34.829208 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:37:45 crc kubenswrapper[4766]: I1126 01:37:45.826556 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:37:45 crc kubenswrapper[4766]: E1126 01:37:45.827575 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:37:58 crc kubenswrapper[4766]: I1126 01:37:58.835822 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:37:58 crc kubenswrapper[4766]: E1126 01:37:58.837858 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:38:09 crc kubenswrapper[4766]: I1126 01:38:09.837750 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:38:09 crc kubenswrapper[4766]: E1126 01:38:09.838823 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:38:22 crc kubenswrapper[4766]: I1126 01:38:22.827165 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:38:22 crc kubenswrapper[4766]: E1126 01:38:22.829538 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:38:37 crc kubenswrapper[4766]: I1126 01:38:37.826856 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:38:37 crc kubenswrapper[4766]: E1126 01:38:37.827724 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:38:48 crc kubenswrapper[4766]: I1126 01:38:48.828363 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:38:48 crc kubenswrapper[4766]: E1126 01:38:48.829597 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:39:01 crc kubenswrapper[4766]: I1126 01:39:01.827309 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:39:01 crc kubenswrapper[4766]: E1126 01:39:01.828439 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:39:13 crc kubenswrapper[4766]: I1126 01:39:13.827499 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:39:13 crc kubenswrapper[4766]: E1126 01:39:13.828331 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:39:24 crc kubenswrapper[4766]: I1126 01:39:24.827628 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:39:24 crc kubenswrapper[4766]: E1126 01:39:24.829050 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:39:38 crc kubenswrapper[4766]: I1126 01:39:38.827284 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:39:38 crc kubenswrapper[4766]: E1126 01:39:38.828518 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:39:50 crc kubenswrapper[4766]: I1126 01:39:50.826840 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:39:50 crc kubenswrapper[4766]: E1126 01:39:50.827734 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:40:01 crc kubenswrapper[4766]: I1126 01:40:01.827000 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:40:01 crc kubenswrapper[4766]: E1126 01:40:01.828116 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:40:13 crc kubenswrapper[4766]: I1126 01:40:13.827174 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:40:13 crc kubenswrapper[4766]: E1126 01:40:13.828354 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:40:27 crc kubenswrapper[4766]: I1126 01:40:27.828453 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:40:27 crc kubenswrapper[4766]: E1126 01:40:27.829796 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:40:42 crc kubenswrapper[4766]: I1126 01:40:42.827080 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:40:43 crc kubenswrapper[4766]: I1126 01:40:43.211441 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"e9fdc469b39326ae18e21b32546bbc3ad05efbc565ef046911377215b98ca5e8"} Nov 26 01:40:49 crc kubenswrapper[4766]: I1126 01:40:49.293213 4766 generic.go:334] "Generic (PLEG): container finished" podID="eef488e4-2c28-4247-a8f4-b8e9008d518e" containerID="6cf98dd3de7c7de7fd48a38947f713d37b0e1446fb81ab33269d7274e49b3042" exitCode=0 Nov 26 01:40:49 crc kubenswrapper[4766]: I1126 01:40:49.293355 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" event={"ID":"eef488e4-2c28-4247-a8f4-b8e9008d518e","Type":"ContainerDied","Data":"6cf98dd3de7c7de7fd48a38947f713d37b0e1446fb81ab33269d7274e49b3042"} Nov 26 01:40:50 crc kubenswrapper[4766]: I1126 01:40:50.819913 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:40:50 crc kubenswrapper[4766]: I1126 01:40:50.922904 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wcdx\" (UniqueName: \"kubernetes.io/projected/eef488e4-2c28-4247-a8f4-b8e9008d518e-kube-api-access-9wcdx\") pod \"eef488e4-2c28-4247-a8f4-b8e9008d518e\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " Nov 26 01:40:50 crc kubenswrapper[4766]: I1126 01:40:50.922976 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceilometer-compute-config-data-0\") pod \"eef488e4-2c28-4247-a8f4-b8e9008d518e\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " Nov 26 01:40:50 crc kubenswrapper[4766]: I1126 01:40:50.923001 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceilometer-compute-config-data-1\") pod \"eef488e4-2c28-4247-a8f4-b8e9008d518e\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " Nov 26 01:40:50 crc kubenswrapper[4766]: I1126 01:40:50.923048 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-telemetry-combined-ca-bundle\") pod \"eef488e4-2c28-4247-a8f4-b8e9008d518e\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " Nov 26 01:40:50 crc kubenswrapper[4766]: I1126 01:40:50.923209 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-inventory\") pod \"eef488e4-2c28-4247-a8f4-b8e9008d518e\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " Nov 26 01:40:50 crc kubenswrapper[4766]: I1126 01:40:50.923245 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceilometer-compute-config-data-2\") pod \"eef488e4-2c28-4247-a8f4-b8e9008d518e\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " Nov 26 01:40:50 crc kubenswrapper[4766]: I1126 01:40:50.923342 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ssh-key\") pod \"eef488e4-2c28-4247-a8f4-b8e9008d518e\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " Nov 26 01:40:50 crc kubenswrapper[4766]: I1126 01:40:50.923397 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceph\") pod \"eef488e4-2c28-4247-a8f4-b8e9008d518e\" (UID: \"eef488e4-2c28-4247-a8f4-b8e9008d518e\") " Nov 26 01:40:50 crc kubenswrapper[4766]: I1126 01:40:50.930861 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "eef488e4-2c28-4247-a8f4-b8e9008d518e" (UID: "eef488e4-2c28-4247-a8f4-b8e9008d518e"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:40:50 crc kubenswrapper[4766]: I1126 01:40:50.930988 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eef488e4-2c28-4247-a8f4-b8e9008d518e-kube-api-access-9wcdx" (OuterVolumeSpecName: "kube-api-access-9wcdx") pod "eef488e4-2c28-4247-a8f4-b8e9008d518e" (UID: "eef488e4-2c28-4247-a8f4-b8e9008d518e"). InnerVolumeSpecName "kube-api-access-9wcdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:40:50 crc kubenswrapper[4766]: I1126 01:40:50.934378 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceph" (OuterVolumeSpecName: "ceph") pod "eef488e4-2c28-4247-a8f4-b8e9008d518e" (UID: "eef488e4-2c28-4247-a8f4-b8e9008d518e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:40:50 crc kubenswrapper[4766]: I1126 01:40:50.956305 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "eef488e4-2c28-4247-a8f4-b8e9008d518e" (UID: "eef488e4-2c28-4247-a8f4-b8e9008d518e"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:40:50 crc kubenswrapper[4766]: I1126 01:40:50.959278 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "eef488e4-2c28-4247-a8f4-b8e9008d518e" (UID: "eef488e4-2c28-4247-a8f4-b8e9008d518e"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:40:50 crc kubenswrapper[4766]: I1126 01:40:50.962304 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-inventory" (OuterVolumeSpecName: "inventory") pod "eef488e4-2c28-4247-a8f4-b8e9008d518e" (UID: "eef488e4-2c28-4247-a8f4-b8e9008d518e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:40:50 crc kubenswrapper[4766]: I1126 01:40:50.964983 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "eef488e4-2c28-4247-a8f4-b8e9008d518e" (UID: "eef488e4-2c28-4247-a8f4-b8e9008d518e"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:40:50 crc kubenswrapper[4766]: I1126 01:40:50.969114 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "eef488e4-2c28-4247-a8f4-b8e9008d518e" (UID: "eef488e4-2c28-4247-a8f4-b8e9008d518e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.025492 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.025943 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.025954 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wcdx\" (UniqueName: \"kubernetes.io/projected/eef488e4-2c28-4247-a8f4-b8e9008d518e-kube-api-access-9wcdx\") on node \"crc\" DevicePath \"\"" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.025966 4766 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.025978 4766 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.025989 4766 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.026001 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.026013 4766 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/eef488e4-2c28-4247-a8f4-b8e9008d518e-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.322457 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" event={"ID":"eef488e4-2c28-4247-a8f4-b8e9008d518e","Type":"ContainerDied","Data":"fa78979048843c160a7af051c231c9dbe2d79e6f6eefd5095204bcd291168a3d"} Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.322497 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa78979048843c160a7af051c231c9dbe2d79e6f6eefd5095204bcd291168a3d" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.322537 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g6znx" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.464577 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8"] Nov 26 01:40:51 crc kubenswrapper[4766]: E1126 01:40:51.464975 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eef488e4-2c28-4247-a8f4-b8e9008d518e" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.464995 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="eef488e4-2c28-4247-a8f4-b8e9008d518e" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.465231 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="eef488e4-2c28-4247-a8f4-b8e9008d518e" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.465943 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.468267 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.468333 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.468410 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.468490 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.468973 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-ipmi-config-data" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.469540 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.486818 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8"] Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.535118 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.535163 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.535352 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.535407 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.535616 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceph\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.535946 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.535997 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.536040 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xbfc\" (UniqueName: \"kubernetes.io/projected/0667c3b6-bdae-4a36-acce-947435c51d91-kube-api-access-8xbfc\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.638205 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.638301 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.638356 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xbfc\" (UniqueName: \"kubernetes.io/projected/0667c3b6-bdae-4a36-acce-947435c51d91-kube-api-access-8xbfc\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.638411 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.638452 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.638572 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.638624 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.638850 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceph\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.642949 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.643078 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.643038 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.644014 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.644688 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.646381 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.656104 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceph\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.656818 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xbfc\" (UniqueName: \"kubernetes.io/projected/0667c3b6-bdae-4a36-acce-947435c51d91-kube-api-access-8xbfc\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:51 crc kubenswrapper[4766]: I1126 01:40:51.794403 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:40:52 crc kubenswrapper[4766]: I1126 01:40:52.452206 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8"] Nov 26 01:40:52 crc kubenswrapper[4766]: I1126 01:40:52.452679 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 01:40:53 crc kubenswrapper[4766]: I1126 01:40:53.343987 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" event={"ID":"0667c3b6-bdae-4a36-acce-947435c51d91","Type":"ContainerStarted","Data":"0f94161a57a5bca5d170e4dbccf98b55d615240bce36a4bce31ed35936861532"} Nov 26 01:40:53 crc kubenswrapper[4766]: I1126 01:40:53.344362 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" event={"ID":"0667c3b6-bdae-4a36-acce-947435c51d91","Type":"ContainerStarted","Data":"c19d53c062de62678e1935f632f5b42fe43c6bb00873c5ed518cf8105d7b4b86"} Nov 26 01:40:53 crc kubenswrapper[4766]: I1126 01:40:53.375229 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" podStartSLOduration=1.8790560589999998 podStartE2EDuration="2.375207546s" podCreationTimestamp="2025-11-26 01:40:51 +0000 UTC" firstStartedPulling="2025-11-26 01:40:52.452082631 +0000 UTC m=+4633.300853101" lastFinishedPulling="2025-11-26 01:40:52.948234158 +0000 UTC m=+4633.797004588" observedRunningTime="2025-11-26 01:40:53.365056908 +0000 UTC m=+4634.213827348" watchObservedRunningTime="2025-11-26 01:40:53.375207546 +0000 UTC m=+4634.223977976" Nov 26 01:41:36 crc kubenswrapper[4766]: I1126 01:41:36.144137 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6vj5j"] Nov 26 01:41:36 crc kubenswrapper[4766]: I1126 01:41:36.147961 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vj5j" Nov 26 01:41:36 crc kubenswrapper[4766]: I1126 01:41:36.168933 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6vj5j"] Nov 26 01:41:36 crc kubenswrapper[4766]: I1126 01:41:36.208162 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1f0994c-ef19-451f-bc61-fe7b66c3399b-catalog-content\") pod \"community-operators-6vj5j\" (UID: \"e1f0994c-ef19-451f-bc61-fe7b66c3399b\") " pod="openshift-marketplace/community-operators-6vj5j" Nov 26 01:41:36 crc kubenswrapper[4766]: I1126 01:41:36.208503 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1f0994c-ef19-451f-bc61-fe7b66c3399b-utilities\") pod \"community-operators-6vj5j\" (UID: \"e1f0994c-ef19-451f-bc61-fe7b66c3399b\") " pod="openshift-marketplace/community-operators-6vj5j" Nov 26 01:41:36 crc kubenswrapper[4766]: I1126 01:41:36.208734 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9m7n\" (UniqueName: \"kubernetes.io/projected/e1f0994c-ef19-451f-bc61-fe7b66c3399b-kube-api-access-l9m7n\") pod \"community-operators-6vj5j\" (UID: \"e1f0994c-ef19-451f-bc61-fe7b66c3399b\") " pod="openshift-marketplace/community-operators-6vj5j" Nov 26 01:41:36 crc kubenswrapper[4766]: I1126 01:41:36.310825 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1f0994c-ef19-451f-bc61-fe7b66c3399b-catalog-content\") pod \"community-operators-6vj5j\" (UID: \"e1f0994c-ef19-451f-bc61-fe7b66c3399b\") " pod="openshift-marketplace/community-operators-6vj5j" Nov 26 01:41:36 crc kubenswrapper[4766]: I1126 01:41:36.310946 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1f0994c-ef19-451f-bc61-fe7b66c3399b-utilities\") pod \"community-operators-6vj5j\" (UID: \"e1f0994c-ef19-451f-bc61-fe7b66c3399b\") " pod="openshift-marketplace/community-operators-6vj5j" Nov 26 01:41:36 crc kubenswrapper[4766]: I1126 01:41:36.311135 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9m7n\" (UniqueName: \"kubernetes.io/projected/e1f0994c-ef19-451f-bc61-fe7b66c3399b-kube-api-access-l9m7n\") pod \"community-operators-6vj5j\" (UID: \"e1f0994c-ef19-451f-bc61-fe7b66c3399b\") " pod="openshift-marketplace/community-operators-6vj5j" Nov 26 01:41:36 crc kubenswrapper[4766]: I1126 01:41:36.311535 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1f0994c-ef19-451f-bc61-fe7b66c3399b-utilities\") pod \"community-operators-6vj5j\" (UID: \"e1f0994c-ef19-451f-bc61-fe7b66c3399b\") " pod="openshift-marketplace/community-operators-6vj5j" Nov 26 01:41:36 crc kubenswrapper[4766]: I1126 01:41:36.311613 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1f0994c-ef19-451f-bc61-fe7b66c3399b-catalog-content\") pod \"community-operators-6vj5j\" (UID: \"e1f0994c-ef19-451f-bc61-fe7b66c3399b\") " pod="openshift-marketplace/community-operators-6vj5j" Nov 26 01:41:36 crc kubenswrapper[4766]: I1126 01:41:36.336272 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9m7n\" (UniqueName: \"kubernetes.io/projected/e1f0994c-ef19-451f-bc61-fe7b66c3399b-kube-api-access-l9m7n\") pod \"community-operators-6vj5j\" (UID: \"e1f0994c-ef19-451f-bc61-fe7b66c3399b\") " pod="openshift-marketplace/community-operators-6vj5j" Nov 26 01:41:36 crc kubenswrapper[4766]: I1126 01:41:36.492509 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vj5j" Nov 26 01:41:37 crc kubenswrapper[4766]: I1126 01:41:37.071287 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6vj5j"] Nov 26 01:41:37 crc kubenswrapper[4766]: W1126 01:41:37.072638 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1f0994c_ef19_451f_bc61_fe7b66c3399b.slice/crio-1fa051e2174c77a15743c540d3609665556be2e146f8629b2a1666d2a45b8536 WatchSource:0}: Error finding container 1fa051e2174c77a15743c540d3609665556be2e146f8629b2a1666d2a45b8536: Status 404 returned error can't find the container with id 1fa051e2174c77a15743c540d3609665556be2e146f8629b2a1666d2a45b8536 Nov 26 01:41:37 crc kubenswrapper[4766]: I1126 01:41:37.919581 4766 generic.go:334] "Generic (PLEG): container finished" podID="e1f0994c-ef19-451f-bc61-fe7b66c3399b" containerID="bdd30c50c15c92c9196f32b1acd8e19cab061e0e51ba94456b42a946a02959ea" exitCode=0 Nov 26 01:41:37 crc kubenswrapper[4766]: I1126 01:41:37.919863 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vj5j" event={"ID":"e1f0994c-ef19-451f-bc61-fe7b66c3399b","Type":"ContainerDied","Data":"bdd30c50c15c92c9196f32b1acd8e19cab061e0e51ba94456b42a946a02959ea"} Nov 26 01:41:37 crc kubenswrapper[4766]: I1126 01:41:37.919946 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vj5j" event={"ID":"e1f0994c-ef19-451f-bc61-fe7b66c3399b","Type":"ContainerStarted","Data":"1fa051e2174c77a15743c540d3609665556be2e146f8629b2a1666d2a45b8536"} Nov 26 01:41:42 crc kubenswrapper[4766]: I1126 01:41:42.977619 4766 generic.go:334] "Generic (PLEG): container finished" podID="e1f0994c-ef19-451f-bc61-fe7b66c3399b" containerID="de13291e6a82272c7d4d7e3e04d12dfddd701320e8992049e87bd2ea768c470c" exitCode=0 Nov 26 01:41:42 crc kubenswrapper[4766]: I1126 01:41:42.977730 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vj5j" event={"ID":"e1f0994c-ef19-451f-bc61-fe7b66c3399b","Type":"ContainerDied","Data":"de13291e6a82272c7d4d7e3e04d12dfddd701320e8992049e87bd2ea768c470c"} Nov 26 01:41:43 crc kubenswrapper[4766]: I1126 01:41:43.989959 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vj5j" event={"ID":"e1f0994c-ef19-451f-bc61-fe7b66c3399b","Type":"ContainerStarted","Data":"0f2bdbb3f91f9b7ac50fe687581f5b65a68cbfb8dd101010b975f0f0f50b2cbe"} Nov 26 01:41:44 crc kubenswrapper[4766]: I1126 01:41:44.008194 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6vj5j" podStartSLOduration=2.5108251 podStartE2EDuration="8.008178333s" podCreationTimestamp="2025-11-26 01:41:36 +0000 UTC" firstStartedPulling="2025-11-26 01:41:37.922661122 +0000 UTC m=+4678.771431552" lastFinishedPulling="2025-11-26 01:41:43.420014315 +0000 UTC m=+4684.268784785" observedRunningTime="2025-11-26 01:41:44.004910183 +0000 UTC m=+4684.853680623" watchObservedRunningTime="2025-11-26 01:41:44.008178333 +0000 UTC m=+4684.856948763" Nov 26 01:41:46 crc kubenswrapper[4766]: I1126 01:41:46.493359 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6vj5j" Nov 26 01:41:46 crc kubenswrapper[4766]: I1126 01:41:46.493829 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6vj5j" Nov 26 01:41:46 crc kubenswrapper[4766]: I1126 01:41:46.562581 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6vj5j" Nov 26 01:41:56 crc kubenswrapper[4766]: I1126 01:41:56.574994 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6vj5j" Nov 26 01:41:56 crc kubenswrapper[4766]: I1126 01:41:56.680724 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6vj5j"] Nov 26 01:41:56 crc kubenswrapper[4766]: I1126 01:41:56.736198 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6vhw6"] Nov 26 01:41:56 crc kubenswrapper[4766]: I1126 01:41:56.736532 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6vhw6" podUID="8951948c-6794-4707-b88e-9e7b506c7df0" containerName="registry-server" containerID="cri-o://0d9da0dedb1cb7c18aea33f003bb734bd7f7ca3849370a469e8929e09b890095" gracePeriod=2 Nov 26 01:41:57 crc kubenswrapper[4766]: I1126 01:41:57.191743 4766 generic.go:334] "Generic (PLEG): container finished" podID="8951948c-6794-4707-b88e-9e7b506c7df0" containerID="0d9da0dedb1cb7c18aea33f003bb734bd7f7ca3849370a469e8929e09b890095" exitCode=0 Nov 26 01:41:57 crc kubenswrapper[4766]: I1126 01:41:57.191817 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vhw6" event={"ID":"8951948c-6794-4707-b88e-9e7b506c7df0","Type":"ContainerDied","Data":"0d9da0dedb1cb7c18aea33f003bb734bd7f7ca3849370a469e8929e09b890095"} Nov 26 01:41:57 crc kubenswrapper[4766]: I1126 01:41:57.192263 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vhw6" event={"ID":"8951948c-6794-4707-b88e-9e7b506c7df0","Type":"ContainerDied","Data":"120d87266e1db442f2f0ec2b5b9c21d972c21dffc4badb557fb82360e4eaf848"} Nov 26 01:41:57 crc kubenswrapper[4766]: I1126 01:41:57.192282 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="120d87266e1db442f2f0ec2b5b9c21d972c21dffc4badb557fb82360e4eaf848" Nov 26 01:41:57 crc kubenswrapper[4766]: I1126 01:41:57.241929 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vhw6" Nov 26 01:41:57 crc kubenswrapper[4766]: I1126 01:41:57.349505 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rndnx\" (UniqueName: \"kubernetes.io/projected/8951948c-6794-4707-b88e-9e7b506c7df0-kube-api-access-rndnx\") pod \"8951948c-6794-4707-b88e-9e7b506c7df0\" (UID: \"8951948c-6794-4707-b88e-9e7b506c7df0\") " Nov 26 01:41:57 crc kubenswrapper[4766]: I1126 01:41:57.349913 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8951948c-6794-4707-b88e-9e7b506c7df0-utilities\") pod \"8951948c-6794-4707-b88e-9e7b506c7df0\" (UID: \"8951948c-6794-4707-b88e-9e7b506c7df0\") " Nov 26 01:41:57 crc kubenswrapper[4766]: I1126 01:41:57.350380 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8951948c-6794-4707-b88e-9e7b506c7df0-catalog-content\") pod \"8951948c-6794-4707-b88e-9e7b506c7df0\" (UID: \"8951948c-6794-4707-b88e-9e7b506c7df0\") " Nov 26 01:41:57 crc kubenswrapper[4766]: I1126 01:41:57.351902 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8951948c-6794-4707-b88e-9e7b506c7df0-utilities" (OuterVolumeSpecName: "utilities") pod "8951948c-6794-4707-b88e-9e7b506c7df0" (UID: "8951948c-6794-4707-b88e-9e7b506c7df0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:41:57 crc kubenswrapper[4766]: I1126 01:41:57.360473 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8951948c-6794-4707-b88e-9e7b506c7df0-kube-api-access-rndnx" (OuterVolumeSpecName: "kube-api-access-rndnx") pod "8951948c-6794-4707-b88e-9e7b506c7df0" (UID: "8951948c-6794-4707-b88e-9e7b506c7df0"). InnerVolumeSpecName "kube-api-access-rndnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:41:57 crc kubenswrapper[4766]: I1126 01:41:57.413141 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8951948c-6794-4707-b88e-9e7b506c7df0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8951948c-6794-4707-b88e-9e7b506c7df0" (UID: "8951948c-6794-4707-b88e-9e7b506c7df0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:41:57 crc kubenswrapper[4766]: I1126 01:41:57.452787 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8951948c-6794-4707-b88e-9e7b506c7df0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 01:41:57 crc kubenswrapper[4766]: I1126 01:41:57.452820 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rndnx\" (UniqueName: \"kubernetes.io/projected/8951948c-6794-4707-b88e-9e7b506c7df0-kube-api-access-rndnx\") on node \"crc\" DevicePath \"\"" Nov 26 01:41:57 crc kubenswrapper[4766]: I1126 01:41:57.452833 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8951948c-6794-4707-b88e-9e7b506c7df0-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 01:41:58 crc kubenswrapper[4766]: I1126 01:41:58.207261 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vhw6" Nov 26 01:41:58 crc kubenswrapper[4766]: I1126 01:41:58.263527 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6vhw6"] Nov 26 01:41:58 crc kubenswrapper[4766]: I1126 01:41:58.282246 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6vhw6"] Nov 26 01:41:59 crc kubenswrapper[4766]: I1126 01:41:59.842822 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8951948c-6794-4707-b88e-9e7b506c7df0" path="/var/lib/kubelet/pods/8951948c-6794-4707-b88e-9e7b506c7df0/volumes" Nov 26 01:42:02 crc kubenswrapper[4766]: I1126 01:42:02.951754 4766 scope.go:117] "RemoveContainer" containerID="0d9da0dedb1cb7c18aea33f003bb734bd7f7ca3849370a469e8929e09b890095" Nov 26 01:42:02 crc kubenswrapper[4766]: I1126 01:42:02.995686 4766 scope.go:117] "RemoveContainer" containerID="22997b585a6d84f3f360ddb7f52b0dcc030828cdce4794210dc6451b9ee841f2" Nov 26 01:42:03 crc kubenswrapper[4766]: I1126 01:42:03.424535 4766 scope.go:117] "RemoveContainer" containerID="95a5a7028bd9bfaf61e3fee7917eb6da7156d1f2ca4fba89488f54f6d26c0bb6" Nov 26 01:43:11 crc kubenswrapper[4766]: I1126 01:43:11.479541 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:43:11 crc kubenswrapper[4766]: I1126 01:43:11.480141 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:43:41 crc kubenswrapper[4766]: I1126 01:43:41.479190 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:43:41 crc kubenswrapper[4766]: I1126 01:43:41.479749 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:43:44 crc kubenswrapper[4766]: I1126 01:43:44.662971 4766 generic.go:334] "Generic (PLEG): container finished" podID="0667c3b6-bdae-4a36-acce-947435c51d91" containerID="0f94161a57a5bca5d170e4dbccf98b55d615240bce36a4bce31ed35936861532" exitCode=0 Nov 26 01:43:44 crc kubenswrapper[4766]: I1126 01:43:44.663082 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" event={"ID":"0667c3b6-bdae-4a36-acce-947435c51d91","Type":"ContainerDied","Data":"0f94161a57a5bca5d170e4dbccf98b55d615240bce36a4bce31ed35936861532"} Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.325731 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.429868 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceilometer-ipmi-config-data-1\") pod \"0667c3b6-bdae-4a36-acce-947435c51d91\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.430160 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceilometer-ipmi-config-data-2\") pod \"0667c3b6-bdae-4a36-acce-947435c51d91\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.430384 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-telemetry-power-monitoring-combined-ca-bundle\") pod \"0667c3b6-bdae-4a36-acce-947435c51d91\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.430452 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ssh-key\") pod \"0667c3b6-bdae-4a36-acce-947435c51d91\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.430507 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceilometer-ipmi-config-data-0\") pod \"0667c3b6-bdae-4a36-acce-947435c51d91\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.430538 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-inventory\") pod \"0667c3b6-bdae-4a36-acce-947435c51d91\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.430586 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xbfc\" (UniqueName: \"kubernetes.io/projected/0667c3b6-bdae-4a36-acce-947435c51d91-kube-api-access-8xbfc\") pod \"0667c3b6-bdae-4a36-acce-947435c51d91\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.430635 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceph\") pod \"0667c3b6-bdae-4a36-acce-947435c51d91\" (UID: \"0667c3b6-bdae-4a36-acce-947435c51d91\") " Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.436793 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceph" (OuterVolumeSpecName: "ceph") pod "0667c3b6-bdae-4a36-acce-947435c51d91" (UID: "0667c3b6-bdae-4a36-acce-947435c51d91"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.437209 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "0667c3b6-bdae-4a36-acce-947435c51d91" (UID: "0667c3b6-bdae-4a36-acce-947435c51d91"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.440403 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0667c3b6-bdae-4a36-acce-947435c51d91-kube-api-access-8xbfc" (OuterVolumeSpecName: "kube-api-access-8xbfc") pod "0667c3b6-bdae-4a36-acce-947435c51d91" (UID: "0667c3b6-bdae-4a36-acce-947435c51d91"). InnerVolumeSpecName "kube-api-access-8xbfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.460286 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-inventory" (OuterVolumeSpecName: "inventory") pod "0667c3b6-bdae-4a36-acce-947435c51d91" (UID: "0667c3b6-bdae-4a36-acce-947435c51d91"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.468674 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0667c3b6-bdae-4a36-acce-947435c51d91" (UID: "0667c3b6-bdae-4a36-acce-947435c51d91"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.472343 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceilometer-ipmi-config-data-0" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-0") pod "0667c3b6-bdae-4a36-acce-947435c51d91" (UID: "0667c3b6-bdae-4a36-acce-947435c51d91"). InnerVolumeSpecName "ceilometer-ipmi-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.479470 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceilometer-ipmi-config-data-1" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-1") pod "0667c3b6-bdae-4a36-acce-947435c51d91" (UID: "0667c3b6-bdae-4a36-acce-947435c51d91"). InnerVolumeSpecName "ceilometer-ipmi-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.486286 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceilometer-ipmi-config-data-2" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-2") pod "0667c3b6-bdae-4a36-acce-947435c51d91" (UID: "0667c3b6-bdae-4a36-acce-947435c51d91"). InnerVolumeSpecName "ceilometer-ipmi-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.532467 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xbfc\" (UniqueName: \"kubernetes.io/projected/0667c3b6-bdae-4a36-acce-947435c51d91-kube-api-access-8xbfc\") on node \"crc\" DevicePath \"\"" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.532695 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.532756 4766 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceilometer-ipmi-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.532835 4766 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceilometer-ipmi-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.532904 4766 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.532960 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.533018 4766 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-ceilometer-ipmi-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.533073 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0667c3b6-bdae-4a36-acce-947435c51d91-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.687235 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" event={"ID":"0667c3b6-bdae-4a36-acce-947435c51d91","Type":"ContainerDied","Data":"c19d53c062de62678e1935f632f5b42fe43c6bb00873c5ed518cf8105d7b4b86"} Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.687274 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c19d53c062de62678e1935f632f5b42fe43c6bb00873c5ed518cf8105d7b4b86" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.687335 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.872806 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7"] Nov 26 01:43:46 crc kubenswrapper[4766]: E1126 01:43:46.874934 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8951948c-6794-4707-b88e-9e7b506c7df0" containerName="extract-utilities" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.874963 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="8951948c-6794-4707-b88e-9e7b506c7df0" containerName="extract-utilities" Nov 26 01:43:46 crc kubenswrapper[4766]: E1126 01:43:46.875009 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8951948c-6794-4707-b88e-9e7b506c7df0" containerName="extract-content" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.875022 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="8951948c-6794-4707-b88e-9e7b506c7df0" containerName="extract-content" Nov 26 01:43:46 crc kubenswrapper[4766]: E1126 01:43:46.875066 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0667c3b6-bdae-4a36-acce-947435c51d91" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.875080 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0667c3b6-bdae-4a36-acce-947435c51d91" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Nov 26 01:43:46 crc kubenswrapper[4766]: E1126 01:43:46.875119 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8951948c-6794-4707-b88e-9e7b506c7df0" containerName="registry-server" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.875130 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="8951948c-6794-4707-b88e-9e7b506c7df0" containerName="registry-server" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.875533 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="8951948c-6794-4707-b88e-9e7b506c7df0" containerName="registry-server" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.875581 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="0667c3b6-bdae-4a36-acce-947435c51d91" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.876691 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.879579 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g9q2t" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.880634 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.881107 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-compute-config-data" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.883684 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7"] Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.884863 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.884904 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.884952 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.947024 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-25nr7\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.947692 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gjts\" (UniqueName: \"kubernetes.io/projected/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-kube-api-access-9gjts\") pod \"logging-edpm-deployment-openstack-edpm-ipam-25nr7\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.947758 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-ceph\") pod \"logging-edpm-deployment-openstack-edpm-ipam-25nr7\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.947812 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-25nr7\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.947943 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-25nr7\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:43:46 crc kubenswrapper[4766]: I1126 01:43:46.948059 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-25nr7\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:43:47 crc kubenswrapper[4766]: I1126 01:43:47.049633 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gjts\" (UniqueName: \"kubernetes.io/projected/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-kube-api-access-9gjts\") pod \"logging-edpm-deployment-openstack-edpm-ipam-25nr7\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:43:47 crc kubenswrapper[4766]: I1126 01:43:47.050016 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-ceph\") pod \"logging-edpm-deployment-openstack-edpm-ipam-25nr7\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:43:47 crc kubenswrapper[4766]: I1126 01:43:47.050231 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-25nr7\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:43:47 crc kubenswrapper[4766]: I1126 01:43:47.050608 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-25nr7\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:43:47 crc kubenswrapper[4766]: I1126 01:43:47.050987 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-25nr7\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:43:47 crc kubenswrapper[4766]: I1126 01:43:47.051380 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-25nr7\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:43:47 crc kubenswrapper[4766]: I1126 01:43:47.055293 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-ceph\") pod \"logging-edpm-deployment-openstack-edpm-ipam-25nr7\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:43:47 crc kubenswrapper[4766]: I1126 01:43:47.055370 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-25nr7\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:43:47 crc kubenswrapper[4766]: I1126 01:43:47.056761 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-25nr7\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:43:47 crc kubenswrapper[4766]: I1126 01:43:47.059165 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-25nr7\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:43:47 crc kubenswrapper[4766]: I1126 01:43:47.062132 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-25nr7\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:43:47 crc kubenswrapper[4766]: I1126 01:43:47.066027 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gjts\" (UniqueName: \"kubernetes.io/projected/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-kube-api-access-9gjts\") pod \"logging-edpm-deployment-openstack-edpm-ipam-25nr7\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:43:47 crc kubenswrapper[4766]: I1126 01:43:47.203520 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:43:47 crc kubenswrapper[4766]: I1126 01:43:47.860482 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7"] Nov 26 01:43:48 crc kubenswrapper[4766]: I1126 01:43:48.724077 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" event={"ID":"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a","Type":"ContainerStarted","Data":"f6622c43cb3d089d52c051d81da5e666a1c785046b1e83b5209fc6ea9769a2c7"} Nov 26 01:43:49 crc kubenswrapper[4766]: I1126 01:43:49.737004 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" event={"ID":"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a","Type":"ContainerStarted","Data":"58775869a990d06970c499dee03eaa71f008113e4de5227673a8017d6014cc26"} Nov 26 01:43:49 crc kubenswrapper[4766]: I1126 01:43:49.766575 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" podStartSLOduration=3.284846883 podStartE2EDuration="3.76655834s" podCreationTimestamp="2025-11-26 01:43:46 +0000 UTC" firstStartedPulling="2025-11-26 01:43:48.621974882 +0000 UTC m=+4809.470745352" lastFinishedPulling="2025-11-26 01:43:49.103686349 +0000 UTC m=+4809.952456809" observedRunningTime="2025-11-26 01:43:49.765608847 +0000 UTC m=+4810.614379277" watchObservedRunningTime="2025-11-26 01:43:49.76655834 +0000 UTC m=+4810.615328770" Nov 26 01:44:03 crc kubenswrapper[4766]: I1126 01:44:03.957913 4766 generic.go:334] "Generic (PLEG): container finished" podID="85810b6d-fde1-48a2-b55d-b6c9c72d6b6a" containerID="58775869a990d06970c499dee03eaa71f008113e4de5227673a8017d6014cc26" exitCode=0 Nov 26 01:44:03 crc kubenswrapper[4766]: I1126 01:44:03.958018 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" event={"ID":"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a","Type":"ContainerDied","Data":"58775869a990d06970c499dee03eaa71f008113e4de5227673a8017d6014cc26"} Nov 26 01:44:05 crc kubenswrapper[4766]: I1126 01:44:05.604495 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:44:05 crc kubenswrapper[4766]: I1126 01:44:05.717670 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-ssh-key\") pod \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " Nov 26 01:44:05 crc kubenswrapper[4766]: I1126 01:44:05.717831 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-ceph\") pod \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " Nov 26 01:44:05 crc kubenswrapper[4766]: I1126 01:44:05.717916 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-logging-compute-config-data-1\") pod \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " Nov 26 01:44:05 crc kubenswrapper[4766]: I1126 01:44:05.717945 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-inventory\") pod \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " Nov 26 01:44:05 crc kubenswrapper[4766]: I1126 01:44:05.718052 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gjts\" (UniqueName: \"kubernetes.io/projected/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-kube-api-access-9gjts\") pod \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " Nov 26 01:44:05 crc kubenswrapper[4766]: I1126 01:44:05.718198 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-logging-compute-config-data-0\") pod \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\" (UID: \"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a\") " Nov 26 01:44:05 crc kubenswrapper[4766]: I1126 01:44:05.723878 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-kube-api-access-9gjts" (OuterVolumeSpecName: "kube-api-access-9gjts") pod "85810b6d-fde1-48a2-b55d-b6c9c72d6b6a" (UID: "85810b6d-fde1-48a2-b55d-b6c9c72d6b6a"). InnerVolumeSpecName "kube-api-access-9gjts". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:44:05 crc kubenswrapper[4766]: I1126 01:44:05.724287 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-ceph" (OuterVolumeSpecName: "ceph") pod "85810b6d-fde1-48a2-b55d-b6c9c72d6b6a" (UID: "85810b6d-fde1-48a2-b55d-b6c9c72d6b6a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:44:05 crc kubenswrapper[4766]: I1126 01:44:05.762227 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "85810b6d-fde1-48a2-b55d-b6c9c72d6b6a" (UID: "85810b6d-fde1-48a2-b55d-b6c9c72d6b6a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:44:05 crc kubenswrapper[4766]: I1126 01:44:05.764311 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-inventory" (OuterVolumeSpecName: "inventory") pod "85810b6d-fde1-48a2-b55d-b6c9c72d6b6a" (UID: "85810b6d-fde1-48a2-b55d-b6c9c72d6b6a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:44:05 crc kubenswrapper[4766]: I1126 01:44:05.775808 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-logging-compute-config-data-1" (OuterVolumeSpecName: "logging-compute-config-data-1") pod "85810b6d-fde1-48a2-b55d-b6c9c72d6b6a" (UID: "85810b6d-fde1-48a2-b55d-b6c9c72d6b6a"). InnerVolumeSpecName "logging-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:44:05 crc kubenswrapper[4766]: I1126 01:44:05.785988 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-logging-compute-config-data-0" (OuterVolumeSpecName: "logging-compute-config-data-0") pod "85810b6d-fde1-48a2-b55d-b6c9c72d6b6a" (UID: "85810b6d-fde1-48a2-b55d-b6c9c72d6b6a"). InnerVolumeSpecName "logging-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:44:05 crc kubenswrapper[4766]: I1126 01:44:05.821105 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:05 crc kubenswrapper[4766]: I1126 01:44:05.821136 4766 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-logging-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:05 crc kubenswrapper[4766]: I1126 01:44:05.821152 4766 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:05 crc kubenswrapper[4766]: I1126 01:44:05.821166 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gjts\" (UniqueName: \"kubernetes.io/projected/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-kube-api-access-9gjts\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:05 crc kubenswrapper[4766]: I1126 01:44:05.821179 4766 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-logging-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:05 crc kubenswrapper[4766]: I1126 01:44:05.821189 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85810b6d-fde1-48a2-b55d-b6c9c72d6b6a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:06 crc kubenswrapper[4766]: I1126 01:44:06.018144 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" event={"ID":"85810b6d-fde1-48a2-b55d-b6c9c72d6b6a","Type":"ContainerDied","Data":"f6622c43cb3d089d52c051d81da5e666a1c785046b1e83b5209fc6ea9769a2c7"} Nov 26 01:44:06 crc kubenswrapper[4766]: I1126 01:44:06.018191 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6622c43cb3d089d52c051d81da5e666a1c785046b1e83b5209fc6ea9769a2c7" Nov 26 01:44:06 crc kubenswrapper[4766]: I1126 01:44:06.018223 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-25nr7" Nov 26 01:44:11 crc kubenswrapper[4766]: I1126 01:44:11.479932 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:44:11 crc kubenswrapper[4766]: I1126 01:44:11.480539 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:44:11 crc kubenswrapper[4766]: I1126 01:44:11.480614 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 01:44:11 crc kubenswrapper[4766]: I1126 01:44:11.481861 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e9fdc469b39326ae18e21b32546bbc3ad05efbc565ef046911377215b98ca5e8"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 01:44:11 crc kubenswrapper[4766]: I1126 01:44:11.481945 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://e9fdc469b39326ae18e21b32546bbc3ad05efbc565ef046911377215b98ca5e8" gracePeriod=600 Nov 26 01:44:12 crc kubenswrapper[4766]: I1126 01:44:12.092189 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="e9fdc469b39326ae18e21b32546bbc3ad05efbc565ef046911377215b98ca5e8" exitCode=0 Nov 26 01:44:12 crc kubenswrapper[4766]: I1126 01:44:12.092253 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"e9fdc469b39326ae18e21b32546bbc3ad05efbc565ef046911377215b98ca5e8"} Nov 26 01:44:12 crc kubenswrapper[4766]: I1126 01:44:12.092456 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479"} Nov 26 01:44:12 crc kubenswrapper[4766]: I1126 01:44:12.092486 4766 scope.go:117] "RemoveContainer" containerID="6e55eae70a0fd72d693176aff272a2cd048703a594b7a4e004e92fee110fb657" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.705831 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 26 01:44:22 crc kubenswrapper[4766]: E1126 01:44:22.706947 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85810b6d-fde1-48a2-b55d-b6c9c72d6b6a" containerName="logging-edpm-deployment-openstack-edpm-ipam" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.706968 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="85810b6d-fde1-48a2-b55d-b6c9c72d6b6a" containerName="logging-edpm-deployment-openstack-edpm-ipam" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.707191 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="85810b6d-fde1-48a2-b55d-b6c9c72d6b6a" containerName="logging-edpm-deployment-openstack-edpm-ipam" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.708628 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.710387 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.711834 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.723200 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.737595 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.737673 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.737691 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.737724 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-dev\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.737757 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgljl\" (UniqueName: \"kubernetes.io/projected/5687a3ac-6850-4533-9c40-b0954d98c628-kube-api-access-wgljl\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.737782 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5687a3ac-6850-4533-9c40-b0954d98c628-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.737822 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-run\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.737850 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5687a3ac-6850-4533-9c40-b0954d98c628-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.737875 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5687a3ac-6850-4533-9c40-b0954d98c628-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.737912 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.738310 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.738356 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5687a3ac-6850-4533-9c40-b0954d98c628-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.738379 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5687a3ac-6850-4533-9c40-b0954d98c628-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.738395 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.738446 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.738519 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-sys\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.789795 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.791882 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.797180 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.801960 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.840926 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-run\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.841178 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.841278 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.841387 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5687a3ac-6850-4533-9c40-b0954d98c628-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.841524 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5687a3ac-6850-4533-9c40-b0954d98c628-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.841672 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.841892 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.842041 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-sys\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.842113 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-sys\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.842219 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.842347 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2c89caf0-3bca-4172-9a9b-27657d46a1c5-ceph\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.842459 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.842559 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c89caf0-3bca-4172-9a9b-27657d46a1c5-scripts\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.842690 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.842800 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.842915 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-dev\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.843012 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-lib-modules\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.843114 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-etc-nvme\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.843199 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.843310 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgljl\" (UniqueName: \"kubernetes.io/projected/5687a3ac-6850-4533-9c40-b0954d98c628-kube-api-access-wgljl\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.843485 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5687a3ac-6850-4533-9c40-b0954d98c628-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.843600 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-run\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.843741 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5687a3ac-6850-4533-9c40-b0954d98c628-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.844562 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c89caf0-3bca-4172-9a9b-27657d46a1c5-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.844682 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-dev\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.844784 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl2r7\" (UniqueName: \"kubernetes.io/projected/2c89caf0-3bca-4172-9a9b-27657d46a1c5-kube-api-access-rl2r7\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.844881 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5687a3ac-6850-4533-9c40-b0954d98c628-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.844978 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2c89caf0-3bca-4172-9a9b-27657d46a1c5-config-data-custom\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.845153 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.845294 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c89caf0-3bca-4172-9a9b-27657d46a1c5-config-data\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.845404 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.845541 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.843884 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-dev\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.843880 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-sys\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.844889 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.843883 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.843989 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-run\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.844423 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.844791 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.843861 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.843842 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.846354 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5687a3ac-6850-4533-9c40-b0954d98c628-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.853076 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5687a3ac-6850-4533-9c40-b0954d98c628-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.874423 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5687a3ac-6850-4533-9c40-b0954d98c628-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.874798 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5687a3ac-6850-4533-9c40-b0954d98c628-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.874506 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5687a3ac-6850-4533-9c40-b0954d98c628-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.876899 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgljl\" (UniqueName: \"kubernetes.io/projected/5687a3ac-6850-4533-9c40-b0954d98c628-kube-api-access-wgljl\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.898538 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5687a3ac-6850-4533-9c40-b0954d98c628-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"5687a3ac-6850-4533-9c40-b0954d98c628\") " pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.948230 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-sys\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.948274 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.948303 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2c89caf0-3bca-4172-9a9b-27657d46a1c5-ceph\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.948329 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c89caf0-3bca-4172-9a9b-27657d46a1c5-scripts\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.948374 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-sys\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.948436 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-lib-modules\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.948391 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-lib-modules\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.948437 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.948484 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-etc-nvme\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.948521 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.948612 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c89caf0-3bca-4172-9a9b-27657d46a1c5-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.948649 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-dev\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.948706 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl2r7\" (UniqueName: \"kubernetes.io/projected/2c89caf0-3bca-4172-9a9b-27657d46a1c5-kube-api-access-rl2r7\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.948761 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2c89caf0-3bca-4172-9a9b-27657d46a1c5-config-data-custom\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.948841 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c89caf0-3bca-4172-9a9b-27657d46a1c5-config-data\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.948867 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.948890 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.948947 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-dev\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.948962 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-run\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.949006 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.949027 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.949066 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-etc-nvme\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.949105 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.949284 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.949332 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-run\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.949377 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2c89caf0-3bca-4172-9a9b-27657d46a1c5-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.952974 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c89caf0-3bca-4172-9a9b-27657d46a1c5-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.953042 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2c89caf0-3bca-4172-9a9b-27657d46a1c5-config-data-custom\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.953154 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c89caf0-3bca-4172-9a9b-27657d46a1c5-scripts\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.953672 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2c89caf0-3bca-4172-9a9b-27657d46a1c5-ceph\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.955011 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c89caf0-3bca-4172-9a9b-27657d46a1c5-config-data\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:22 crc kubenswrapper[4766]: I1126 01:44:22.968230 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl2r7\" (UniqueName: \"kubernetes.io/projected/2c89caf0-3bca-4172-9a9b-27657d46a1c5-kube-api-access-rl2r7\") pod \"cinder-backup-0\" (UID: \"2c89caf0-3bca-4172-9a9b-27657d46a1c5\") " pod="openstack/cinder-backup-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.031770 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.138417 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.582967 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.590895 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.599370 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.599707 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-7t54l" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.600277 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.601566 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.609218 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.652825 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.654640 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.664198 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.664596 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.677542 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.677587 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99765d7e-a086-4d6a-9f06-3847a50c658e-logs\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.677609 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.677627 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-scripts\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.677652 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99765d7e-a086-4d6a-9f06-3847a50c658e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.677700 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.677741 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.677757 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/672183ef-ddc2-4b72-b127-e1953eaa6549-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.677789 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmvnv\" (UniqueName: \"kubernetes.io/projected/672183ef-ddc2-4b72-b127-e1953eaa6549-kube-api-access-qmvnv\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.677815 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.677858 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.677885 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.677909 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/672183ef-ddc2-4b72-b127-e1953eaa6549-logs\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.677937 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hs6rd\" (UniqueName: \"kubernetes.io/projected/99765d7e-a086-4d6a-9f06-3847a50c658e-kube-api-access-hs6rd\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.677965 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/99765d7e-a086-4d6a-9f06-3847a50c658e-ceph\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.677981 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/672183ef-ddc2-4b72-b127-e1953eaa6549-ceph\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.677998 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-config-data\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.678014 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.711100 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-8lxnr"] Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.713003 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-8lxnr" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.736363 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.764627 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-8lxnr"] Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.779967 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.780047 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.780078 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/672183ef-ddc2-4b72-b127-e1953eaa6549-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.780111 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmvnv\" (UniqueName: \"kubernetes.io/projected/672183ef-ddc2-4b72-b127-e1953eaa6549-kube-api-access-qmvnv\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.780135 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.780181 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.780215 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e-operator-scripts\") pod \"manila-db-create-8lxnr\" (UID: \"aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e\") " pod="openstack/manila-db-create-8lxnr" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.780234 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.780256 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwn9k\" (UniqueName: \"kubernetes.io/projected/aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e-kube-api-access-hwn9k\") pod \"manila-db-create-8lxnr\" (UID: \"aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e\") " pod="openstack/manila-db-create-8lxnr" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.780275 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/672183ef-ddc2-4b72-b127-e1953eaa6549-logs\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.780300 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hs6rd\" (UniqueName: \"kubernetes.io/projected/99765d7e-a086-4d6a-9f06-3847a50c658e-kube-api-access-hs6rd\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.780328 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/99765d7e-a086-4d6a-9f06-3847a50c658e-ceph\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.780346 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/672183ef-ddc2-4b72-b127-e1953eaa6549-ceph\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.780364 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-config-data\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.780379 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.780409 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.780429 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99765d7e-a086-4d6a-9f06-3847a50c658e-logs\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.780448 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.780463 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-scripts\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.780483 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99765d7e-a086-4d6a-9f06-3847a50c658e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.780988 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99765d7e-a086-4d6a-9f06-3847a50c658e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.781082 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.781156 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/672183ef-ddc2-4b72-b127-e1953eaa6549-logs\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.781639 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-7368-account-create-update-qcqxb"] Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.783105 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-7368-account-create-update-qcqxb" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.784762 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.788290 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.793572 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/672183ef-ddc2-4b72-b127-e1953eaa6549-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.793710 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.794270 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99765d7e-a086-4d6a-9f06-3847a50c658e-logs\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.800029 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.801742 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-scripts\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.806909 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-7368-account-create-update-qcqxb"] Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.819773 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.821145 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-config-data\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.821757 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/672183ef-ddc2-4b72-b127-e1953eaa6549-ceph\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.822015 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.822324 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/99765d7e-a086-4d6a-9f06-3847a50c658e-ceph\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.824332 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.838466 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hs6rd\" (UniqueName: \"kubernetes.io/projected/99765d7e-a086-4d6a-9f06-3847a50c658e-kube-api-access-hs6rd\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.842166 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.862849 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmvnv\" (UniqueName: \"kubernetes.io/projected/672183ef-ddc2-4b72-b127-e1953eaa6549-kube-api-access-qmvnv\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.888972 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e-operator-scripts\") pod \"manila-db-create-8lxnr\" (UID: \"aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e\") " pod="openstack/manila-db-create-8lxnr" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.889622 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwn9k\" (UniqueName: \"kubernetes.io/projected/aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e-kube-api-access-hwn9k\") pod \"manila-db-create-8lxnr\" (UID: \"aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e\") " pod="openstack/manila-db-create-8lxnr" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.891331 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e-operator-scripts\") pod \"manila-db-create-8lxnr\" (UID: \"aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e\") " pod="openstack/manila-db-create-8lxnr" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.904141 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.913835 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.926800 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.930705 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwn9k\" (UniqueName: \"kubernetes.io/projected/aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e-kube-api-access-hwn9k\") pod \"manila-db-create-8lxnr\" (UID: \"aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e\") " pod="openstack/manila-db-create-8lxnr" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.937631 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.952906 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7f5b87f5c-z9b2b"] Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.962611 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f5b87f5c-z9b2b" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.975081 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.975346 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-p9nbj" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.975513 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 26 01:44:23 crc kubenswrapper[4766]: I1126 01:44:23.975876 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:23.998137 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dcf71c2f-342b-4253-b16e-da87cebb754f-operator-scripts\") pod \"manila-7368-account-create-update-qcqxb\" (UID: \"dcf71c2f-342b-4253-b16e-da87cebb754f\") " pod="openstack/manila-7368-account-create-update-qcqxb" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:23.998180 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq72v\" (UniqueName: \"kubernetes.io/projected/dcf71c2f-342b-4253-b16e-da87cebb754f-kube-api-access-kq72v\") pod \"manila-7368-account-create-update-qcqxb\" (UID: \"dcf71c2f-342b-4253-b16e-da87cebb754f\") " pod="openstack/manila-7368-account-create-update-qcqxb" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:23.999260 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.008918 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f5b87f5c-z9b2b"] Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.052862 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-8lxnr" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.054785 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7f7f64877f-4sskz"] Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.059008 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f7f64877f-4sskz" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.080699 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 01:44:24 crc kubenswrapper[4766]: W1126 01:44:24.082402 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c89caf0_3bca_4172_9a9b_27657d46a1c5.slice/crio-ec76d97e065efed40aaf41a81c129ed16bfc0d42de06b9e5ed3e21cdef21e8fd WatchSource:0}: Error finding container ec76d97e065efed40aaf41a81c129ed16bfc0d42de06b9e5ed3e21cdef21e8fd: Status 404 returned error can't find the container with id ec76d97e065efed40aaf41a81c129ed16bfc0d42de06b9e5ed3e21cdef21e8fd Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.090976 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f7f64877f-4sskz"] Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.102342 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c577b8bb-a3c5-405d-8da8-0289e8483ba5-scripts\") pod \"horizon-7f5b87f5c-z9b2b\" (UID: \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\") " pod="openstack/horizon-7f5b87f5c-z9b2b" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.102404 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c577b8bb-a3c5-405d-8da8-0289e8483ba5-horizon-secret-key\") pod \"horizon-7f5b87f5c-z9b2b\" (UID: \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\") " pod="openstack/horizon-7f5b87f5c-z9b2b" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.102437 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c577b8bb-a3c5-405d-8da8-0289e8483ba5-config-data\") pod \"horizon-7f5b87f5c-z9b2b\" (UID: \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\") " pod="openstack/horizon-7f5b87f5c-z9b2b" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.102476 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnlf9\" (UniqueName: \"kubernetes.io/projected/c577b8bb-a3c5-405d-8da8-0289e8483ba5-kube-api-access-fnlf9\") pod \"horizon-7f5b87f5c-z9b2b\" (UID: \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\") " pod="openstack/horizon-7f5b87f5c-z9b2b" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.102512 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dcf71c2f-342b-4253-b16e-da87cebb754f-operator-scripts\") pod \"manila-7368-account-create-update-qcqxb\" (UID: \"dcf71c2f-342b-4253-b16e-da87cebb754f\") " pod="openstack/manila-7368-account-create-update-qcqxb" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.102537 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq72v\" (UniqueName: \"kubernetes.io/projected/dcf71c2f-342b-4253-b16e-da87cebb754f-kube-api-access-kq72v\") pod \"manila-7368-account-create-update-qcqxb\" (UID: \"dcf71c2f-342b-4253-b16e-da87cebb754f\") " pod="openstack/manila-7368-account-create-update-qcqxb" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.102576 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c577b8bb-a3c5-405d-8da8-0289e8483ba5-logs\") pod \"horizon-7f5b87f5c-z9b2b\" (UID: \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\") " pod="openstack/horizon-7f5b87f5c-z9b2b" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.103502 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dcf71c2f-342b-4253-b16e-da87cebb754f-operator-scripts\") pod \"manila-7368-account-create-update-qcqxb\" (UID: \"dcf71c2f-342b-4253-b16e-da87cebb754f\") " pod="openstack/manila-7368-account-create-update-qcqxb" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.108609 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.134012 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq72v\" (UniqueName: \"kubernetes.io/projected/dcf71c2f-342b-4253-b16e-da87cebb754f-kube-api-access-kq72v\") pod \"manila-7368-account-create-update-qcqxb\" (UID: \"dcf71c2f-342b-4253-b16e-da87cebb754f\") " pod="openstack/manila-7368-account-create-update-qcqxb" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.154787 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-7368-account-create-update-qcqxb" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.158733 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.205950 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/656cb04b-f831-4671-a102-d01c6d01714f-config-data\") pod \"horizon-7f7f64877f-4sskz\" (UID: \"656cb04b-f831-4671-a102-d01c6d01714f\") " pod="openstack/horizon-7f7f64877f-4sskz" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.206000 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/656cb04b-f831-4671-a102-d01c6d01714f-logs\") pod \"horizon-7f7f64877f-4sskz\" (UID: \"656cb04b-f831-4671-a102-d01c6d01714f\") " pod="openstack/horizon-7f7f64877f-4sskz" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.206071 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c577b8bb-a3c5-405d-8da8-0289e8483ba5-scripts\") pod \"horizon-7f5b87f5c-z9b2b\" (UID: \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\") " pod="openstack/horizon-7f5b87f5c-z9b2b" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.206109 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c577b8bb-a3c5-405d-8da8-0289e8483ba5-horizon-secret-key\") pod \"horizon-7f5b87f5c-z9b2b\" (UID: \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\") " pod="openstack/horizon-7f5b87f5c-z9b2b" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.206135 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qf7b\" (UniqueName: \"kubernetes.io/projected/656cb04b-f831-4671-a102-d01c6d01714f-kube-api-access-4qf7b\") pod \"horizon-7f7f64877f-4sskz\" (UID: \"656cb04b-f831-4671-a102-d01c6d01714f\") " pod="openstack/horizon-7f7f64877f-4sskz" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.206167 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/656cb04b-f831-4671-a102-d01c6d01714f-horizon-secret-key\") pod \"horizon-7f7f64877f-4sskz\" (UID: \"656cb04b-f831-4671-a102-d01c6d01714f\") " pod="openstack/horizon-7f7f64877f-4sskz" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.206187 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c577b8bb-a3c5-405d-8da8-0289e8483ba5-config-data\") pod \"horizon-7f5b87f5c-z9b2b\" (UID: \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\") " pod="openstack/horizon-7f5b87f5c-z9b2b" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.206209 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/656cb04b-f831-4671-a102-d01c6d01714f-scripts\") pod \"horizon-7f7f64877f-4sskz\" (UID: \"656cb04b-f831-4671-a102-d01c6d01714f\") " pod="openstack/horizon-7f7f64877f-4sskz" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.206245 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnlf9\" (UniqueName: \"kubernetes.io/projected/c577b8bb-a3c5-405d-8da8-0289e8483ba5-kube-api-access-fnlf9\") pod \"horizon-7f5b87f5c-z9b2b\" (UID: \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\") " pod="openstack/horizon-7f5b87f5c-z9b2b" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.206319 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c577b8bb-a3c5-405d-8da8-0289e8483ba5-logs\") pod \"horizon-7f5b87f5c-z9b2b\" (UID: \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\") " pod="openstack/horizon-7f5b87f5c-z9b2b" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.206954 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c577b8bb-a3c5-405d-8da8-0289e8483ba5-scripts\") pod \"horizon-7f5b87f5c-z9b2b\" (UID: \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\") " pod="openstack/horizon-7f5b87f5c-z9b2b" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.206982 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c577b8bb-a3c5-405d-8da8-0289e8483ba5-logs\") pod \"horizon-7f5b87f5c-z9b2b\" (UID: \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\") " pod="openstack/horizon-7f5b87f5c-z9b2b" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.207742 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c577b8bb-a3c5-405d-8da8-0289e8483ba5-config-data\") pod \"horizon-7f5b87f5c-z9b2b\" (UID: \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\") " pod="openstack/horizon-7f5b87f5c-z9b2b" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.224074 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c577b8bb-a3c5-405d-8da8-0289e8483ba5-horizon-secret-key\") pod \"horizon-7f5b87f5c-z9b2b\" (UID: \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\") " pod="openstack/horizon-7f5b87f5c-z9b2b" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.227413 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnlf9\" (UniqueName: \"kubernetes.io/projected/c577b8bb-a3c5-405d-8da8-0289e8483ba5-kube-api-access-fnlf9\") pod \"horizon-7f5b87f5c-z9b2b\" (UID: \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\") " pod="openstack/horizon-7f5b87f5c-z9b2b" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.308688 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/656cb04b-f831-4671-a102-d01c6d01714f-config-data\") pod \"horizon-7f7f64877f-4sskz\" (UID: \"656cb04b-f831-4671-a102-d01c6d01714f\") " pod="openstack/horizon-7f7f64877f-4sskz" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.308739 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/656cb04b-f831-4671-a102-d01c6d01714f-logs\") pod \"horizon-7f7f64877f-4sskz\" (UID: \"656cb04b-f831-4671-a102-d01c6d01714f\") " pod="openstack/horizon-7f7f64877f-4sskz" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.308814 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qf7b\" (UniqueName: \"kubernetes.io/projected/656cb04b-f831-4671-a102-d01c6d01714f-kube-api-access-4qf7b\") pod \"horizon-7f7f64877f-4sskz\" (UID: \"656cb04b-f831-4671-a102-d01c6d01714f\") " pod="openstack/horizon-7f7f64877f-4sskz" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.308838 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/656cb04b-f831-4671-a102-d01c6d01714f-horizon-secret-key\") pod \"horizon-7f7f64877f-4sskz\" (UID: \"656cb04b-f831-4671-a102-d01c6d01714f\") " pod="openstack/horizon-7f7f64877f-4sskz" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.308872 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/656cb04b-f831-4671-a102-d01c6d01714f-scripts\") pod \"horizon-7f7f64877f-4sskz\" (UID: \"656cb04b-f831-4671-a102-d01c6d01714f\") " pod="openstack/horizon-7f7f64877f-4sskz" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.309536 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/656cb04b-f831-4671-a102-d01c6d01714f-scripts\") pod \"horizon-7f7f64877f-4sskz\" (UID: \"656cb04b-f831-4671-a102-d01c6d01714f\") " pod="openstack/horizon-7f7f64877f-4sskz" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.309891 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/656cb04b-f831-4671-a102-d01c6d01714f-config-data\") pod \"horizon-7f7f64877f-4sskz\" (UID: \"656cb04b-f831-4671-a102-d01c6d01714f\") " pod="openstack/horizon-7f7f64877f-4sskz" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.313317 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/656cb04b-f831-4671-a102-d01c6d01714f-horizon-secret-key\") pod \"horizon-7f7f64877f-4sskz\" (UID: \"656cb04b-f831-4671-a102-d01c6d01714f\") " pod="openstack/horizon-7f7f64877f-4sskz" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.319262 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/656cb04b-f831-4671-a102-d01c6d01714f-logs\") pod \"horizon-7f7f64877f-4sskz\" (UID: \"656cb04b-f831-4671-a102-d01c6d01714f\") " pod="openstack/horizon-7f7f64877f-4sskz" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.327407 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qf7b\" (UniqueName: \"kubernetes.io/projected/656cb04b-f831-4671-a102-d01c6d01714f-kube-api-access-4qf7b\") pod \"horizon-7f7f64877f-4sskz\" (UID: \"656cb04b-f831-4671-a102-d01c6d01714f\") " pod="openstack/horizon-7f7f64877f-4sskz" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.334448 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"2c89caf0-3bca-4172-9a9b-27657d46a1c5","Type":"ContainerStarted","Data":"ec76d97e065efed40aaf41a81c129ed16bfc0d42de06b9e5ed3e21cdef21e8fd"} Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.336677 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"5687a3ac-6850-4533-9c40-b0954d98c628","Type":"ContainerStarted","Data":"c1f6e3a1be3011958fdbf939274c1a327362b639020eaa5eda1fcbb8bdf26b8a"} Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.500482 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f5b87f5c-z9b2b" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.519722 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f7f64877f-4sskz" Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.773443 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.794856 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-8lxnr"] Nov 26 01:44:24 crc kubenswrapper[4766]: W1126 01:44:24.801105 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod672183ef_ddc2_4b72_b127_e1953eaa6549.slice/crio-b38a9527fe0612b1cbfe7f17022bb919c83f3527ebfc7502de763f2f180c73d9 WatchSource:0}: Error finding container b38a9527fe0612b1cbfe7f17022bb919c83f3527ebfc7502de763f2f180c73d9: Status 404 returned error can't find the container with id b38a9527fe0612b1cbfe7f17022bb919c83f3527ebfc7502de763f2f180c73d9 Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.885826 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 01:44:24 crc kubenswrapper[4766]: I1126 01:44:24.920917 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-7368-account-create-update-qcqxb"] Nov 26 01:44:25 crc kubenswrapper[4766]: I1126 01:44:25.394716 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-8lxnr" event={"ID":"aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e","Type":"ContainerStarted","Data":"0968be81c771f8ac5a4f846ffab730aa1e8eb3abf04354265c8586d7d2b46dc0"} Nov 26 01:44:25 crc kubenswrapper[4766]: I1126 01:44:25.395196 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-8lxnr" event={"ID":"aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e","Type":"ContainerStarted","Data":"ab0b95667e6017eac76fdc778c5384704a81ddc5a8452c341cc4702ab0053d49"} Nov 26 01:44:25 crc kubenswrapper[4766]: I1126 01:44:25.400887 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-7368-account-create-update-qcqxb" event={"ID":"dcf71c2f-342b-4253-b16e-da87cebb754f","Type":"ContainerStarted","Data":"847a8a0d32abee0675da8f8c432d7927cbe03d20210154d0f8861601f94e93f5"} Nov 26 01:44:25 crc kubenswrapper[4766]: I1126 01:44:25.423098 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"672183ef-ddc2-4b72-b127-e1953eaa6549","Type":"ContainerStarted","Data":"b38a9527fe0612b1cbfe7f17022bb919c83f3527ebfc7502de763f2f180c73d9"} Nov 26 01:44:25 crc kubenswrapper[4766]: I1126 01:44:25.425919 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-create-8lxnr" podStartSLOduration=2.425904887 podStartE2EDuration="2.425904887s" podCreationTimestamp="2025-11-26 01:44:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 01:44:25.423456897 +0000 UTC m=+4846.272227327" watchObservedRunningTime="2025-11-26 01:44:25.425904887 +0000 UTC m=+4846.274675317" Nov 26 01:44:25 crc kubenswrapper[4766]: I1126 01:44:25.448948 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f7f64877f-4sskz"] Nov 26 01:44:25 crc kubenswrapper[4766]: I1126 01:44:25.458278 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"99765d7e-a086-4d6a-9f06-3847a50c658e","Type":"ContainerStarted","Data":"b71d7625053683c4bdf259cc6dc4f940babcc83eb34989dec47427d93a19b352"} Nov 26 01:44:25 crc kubenswrapper[4766]: I1126 01:44:25.556362 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f5b87f5c-z9b2b"] Nov 26 01:44:25 crc kubenswrapper[4766]: W1126 01:44:25.564545 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc577b8bb_a3c5_405d_8da8_0289e8483ba5.slice/crio-975a79f0346bca6295480b6b00c88473a7568423bd71eb337e3e068bede48f83 WatchSource:0}: Error finding container 975a79f0346bca6295480b6b00c88473a7568423bd71eb337e3e068bede48f83: Status 404 returned error can't find the container with id 975a79f0346bca6295480b6b00c88473a7568423bd71eb337e3e068bede48f83 Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.509870 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"2c89caf0-3bca-4172-9a9b-27657d46a1c5","Type":"ContainerStarted","Data":"0f19b883936567eb6497fc2af00306a948529e36a9b8917cce5056215370b0e2"} Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.510784 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"2c89caf0-3bca-4172-9a9b-27657d46a1c5","Type":"ContainerStarted","Data":"ab6686eeab9b0ed66197b7026fc7ed3cfe4c0791ba9c8b719f00928effb23add"} Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.521578 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f7f64877f-4sskz" event={"ID":"656cb04b-f831-4671-a102-d01c6d01714f","Type":"ContainerStarted","Data":"d542487007485be94fc05162b14b1d011807e022670da76db1cc6656b1176848"} Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.544039 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"5687a3ac-6850-4533-9c40-b0954d98c628","Type":"ContainerStarted","Data":"547dd9669ef973c62b6d14aa5a65edb4b84c8dc1b8fdfc584ef802083e33b650"} Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.544084 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"5687a3ac-6850-4533-9c40-b0954d98c628","Type":"ContainerStarted","Data":"b1080470ef7a4afb9168eb46c87aa7a4d5a51ba3ba64e712cccaaad59bb15827"} Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.546340 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=3.20224681 podStartE2EDuration="4.546316171s" podCreationTimestamp="2025-11-26 01:44:22 +0000 UTC" firstStartedPulling="2025-11-26 01:44:24.109972466 +0000 UTC m=+4844.958742886" lastFinishedPulling="2025-11-26 01:44:25.454041817 +0000 UTC m=+4846.302812247" observedRunningTime="2025-11-26 01:44:26.537669379 +0000 UTC m=+4847.386439809" watchObservedRunningTime="2025-11-26 01:44:26.546316171 +0000 UTC m=+4847.395086601" Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.554608 4766 generic.go:334] "Generic (PLEG): container finished" podID="aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e" containerID="0968be81c771f8ac5a4f846ffab730aa1e8eb3abf04354265c8586d7d2b46dc0" exitCode=0 Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.554721 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-8lxnr" event={"ID":"aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e","Type":"ContainerDied","Data":"0968be81c771f8ac5a4f846ffab730aa1e8eb3abf04354265c8586d7d2b46dc0"} Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.561214 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f5b87f5c-z9b2b" event={"ID":"c577b8bb-a3c5-405d-8da8-0289e8483ba5","Type":"ContainerStarted","Data":"975a79f0346bca6295480b6b00c88473a7568423bd71eb337e3e068bede48f83"} Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.563615 4766 generic.go:334] "Generic (PLEG): container finished" podID="dcf71c2f-342b-4253-b16e-da87cebb754f" containerID="b944d21a05b42f79189879092c9064e4cbcb7f134a249ddc8945aec1c7805db2" exitCode=0 Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.563644 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-7368-account-create-update-qcqxb" event={"ID":"dcf71c2f-342b-4253-b16e-da87cebb754f","Type":"ContainerDied","Data":"b944d21a05b42f79189879092c9064e4cbcb7f134a249ddc8945aec1c7805db2"} Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.572427 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=3.508982944 podStartE2EDuration="4.572406281s" podCreationTimestamp="2025-11-26 01:44:22 +0000 UTC" firstStartedPulling="2025-11-26 01:44:23.76896146 +0000 UTC m=+4844.617731890" lastFinishedPulling="2025-11-26 01:44:24.832384797 +0000 UTC m=+4845.681155227" observedRunningTime="2025-11-26 01:44:26.570560506 +0000 UTC m=+4847.419330936" watchObservedRunningTime="2025-11-26 01:44:26.572406281 +0000 UTC m=+4847.421176711" Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.820726 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f5b87f5c-z9b2b"] Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.862381 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-65bd4d46f6-blndr"] Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.868313 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.874820 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.884581 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-65bd4d46f6-blndr"] Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.949138 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e2311f3-2074-4336-ba7c-dc676f60d452-scripts\") pod \"horizon-65bd4d46f6-blndr\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.949478 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e2311f3-2074-4336-ba7c-dc676f60d452-horizon-tls-certs\") pod \"horizon-65bd4d46f6-blndr\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.949536 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e2311f3-2074-4336-ba7c-dc676f60d452-combined-ca-bundle\") pod \"horizon-65bd4d46f6-blndr\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.949564 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e2311f3-2074-4336-ba7c-dc676f60d452-logs\") pod \"horizon-65bd4d46f6-blndr\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.949613 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pbrs\" (UniqueName: \"kubernetes.io/projected/9e2311f3-2074-4336-ba7c-dc676f60d452-kube-api-access-9pbrs\") pod \"horizon-65bd4d46f6-blndr\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.949681 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e2311f3-2074-4336-ba7c-dc676f60d452-config-data\") pod \"horizon-65bd4d46f6-blndr\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:26 crc kubenswrapper[4766]: I1126 01:44:26.949797 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9e2311f3-2074-4336-ba7c-dc676f60d452-horizon-secret-key\") pod \"horizon-65bd4d46f6-blndr\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.009256 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f7f64877f-4sskz"] Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.042183 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-cf46f664-8k6s8"] Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.044503 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.053414 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e2311f3-2074-4336-ba7c-dc676f60d452-logs\") pod \"horizon-65bd4d46f6-blndr\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.053483 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pbrs\" (UniqueName: \"kubernetes.io/projected/9e2311f3-2074-4336-ba7c-dc676f60d452-kube-api-access-9pbrs\") pod \"horizon-65bd4d46f6-blndr\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.053527 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e2311f3-2074-4336-ba7c-dc676f60d452-config-data\") pod \"horizon-65bd4d46f6-blndr\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.053618 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9e2311f3-2074-4336-ba7c-dc676f60d452-horizon-secret-key\") pod \"horizon-65bd4d46f6-blndr\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.053651 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e2311f3-2074-4336-ba7c-dc676f60d452-scripts\") pod \"horizon-65bd4d46f6-blndr\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.053737 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e2311f3-2074-4336-ba7c-dc676f60d452-horizon-tls-certs\") pod \"horizon-65bd4d46f6-blndr\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.060992 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e2311f3-2074-4336-ba7c-dc676f60d452-logs\") pod \"horizon-65bd4d46f6-blndr\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.061631 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e2311f3-2074-4336-ba7c-dc676f60d452-scripts\") pod \"horizon-65bd4d46f6-blndr\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.064288 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e2311f3-2074-4336-ba7c-dc676f60d452-config-data\") pod \"horizon-65bd4d46f6-blndr\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.066181 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e2311f3-2074-4336-ba7c-dc676f60d452-combined-ca-bundle\") pod \"horizon-65bd4d46f6-blndr\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.069846 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e2311f3-2074-4336-ba7c-dc676f60d452-horizon-tls-certs\") pod \"horizon-65bd4d46f6-blndr\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.070260 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9e2311f3-2074-4336-ba7c-dc676f60d452-horizon-secret-key\") pod \"horizon-65bd4d46f6-blndr\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.079334 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-cf46f664-8k6s8"] Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.080112 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e2311f3-2074-4336-ba7c-dc676f60d452-combined-ca-bundle\") pod \"horizon-65bd4d46f6-blndr\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.107958 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pbrs\" (UniqueName: \"kubernetes.io/projected/9e2311f3-2074-4336-ba7c-dc676f60d452-kube-api-access-9pbrs\") pod \"horizon-65bd4d46f6-blndr\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.172254 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/778430ca-2e98-4156-b1d9-fdf05b76ef7e-config-data\") pod \"horizon-cf46f664-8k6s8\" (UID: \"778430ca-2e98-4156-b1d9-fdf05b76ef7e\") " pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.213872 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/778430ca-2e98-4156-b1d9-fdf05b76ef7e-horizon-tls-certs\") pod \"horizon-cf46f664-8k6s8\" (UID: \"778430ca-2e98-4156-b1d9-fdf05b76ef7e\") " pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.214077 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/778430ca-2e98-4156-b1d9-fdf05b76ef7e-scripts\") pod \"horizon-cf46f664-8k6s8\" (UID: \"778430ca-2e98-4156-b1d9-fdf05b76ef7e\") " pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.214234 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztvdn\" (UniqueName: \"kubernetes.io/projected/778430ca-2e98-4156-b1d9-fdf05b76ef7e-kube-api-access-ztvdn\") pod \"horizon-cf46f664-8k6s8\" (UID: \"778430ca-2e98-4156-b1d9-fdf05b76ef7e\") " pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.214265 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/778430ca-2e98-4156-b1d9-fdf05b76ef7e-horizon-secret-key\") pod \"horizon-cf46f664-8k6s8\" (UID: \"778430ca-2e98-4156-b1d9-fdf05b76ef7e\") " pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.214316 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778430ca-2e98-4156-b1d9-fdf05b76ef7e-combined-ca-bundle\") pod \"horizon-cf46f664-8k6s8\" (UID: \"778430ca-2e98-4156-b1d9-fdf05b76ef7e\") " pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.214441 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/778430ca-2e98-4156-b1d9-fdf05b76ef7e-logs\") pod \"horizon-cf46f664-8k6s8\" (UID: \"778430ca-2e98-4156-b1d9-fdf05b76ef7e\") " pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.306203 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.316043 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/778430ca-2e98-4156-b1d9-fdf05b76ef7e-scripts\") pod \"horizon-cf46f664-8k6s8\" (UID: \"778430ca-2e98-4156-b1d9-fdf05b76ef7e\") " pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.316141 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztvdn\" (UniqueName: \"kubernetes.io/projected/778430ca-2e98-4156-b1d9-fdf05b76ef7e-kube-api-access-ztvdn\") pod \"horizon-cf46f664-8k6s8\" (UID: \"778430ca-2e98-4156-b1d9-fdf05b76ef7e\") " pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.316167 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/778430ca-2e98-4156-b1d9-fdf05b76ef7e-horizon-secret-key\") pod \"horizon-cf46f664-8k6s8\" (UID: \"778430ca-2e98-4156-b1d9-fdf05b76ef7e\") " pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.316192 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778430ca-2e98-4156-b1d9-fdf05b76ef7e-combined-ca-bundle\") pod \"horizon-cf46f664-8k6s8\" (UID: \"778430ca-2e98-4156-b1d9-fdf05b76ef7e\") " pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.316245 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/778430ca-2e98-4156-b1d9-fdf05b76ef7e-logs\") pod \"horizon-cf46f664-8k6s8\" (UID: \"778430ca-2e98-4156-b1d9-fdf05b76ef7e\") " pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.316296 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/778430ca-2e98-4156-b1d9-fdf05b76ef7e-config-data\") pod \"horizon-cf46f664-8k6s8\" (UID: \"778430ca-2e98-4156-b1d9-fdf05b76ef7e\") " pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.316340 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/778430ca-2e98-4156-b1d9-fdf05b76ef7e-horizon-tls-certs\") pod \"horizon-cf46f664-8k6s8\" (UID: \"778430ca-2e98-4156-b1d9-fdf05b76ef7e\") " pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.316825 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/778430ca-2e98-4156-b1d9-fdf05b76ef7e-scripts\") pod \"horizon-cf46f664-8k6s8\" (UID: \"778430ca-2e98-4156-b1d9-fdf05b76ef7e\") " pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.317420 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/778430ca-2e98-4156-b1d9-fdf05b76ef7e-logs\") pod \"horizon-cf46f664-8k6s8\" (UID: \"778430ca-2e98-4156-b1d9-fdf05b76ef7e\") " pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.322341 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/778430ca-2e98-4156-b1d9-fdf05b76ef7e-horizon-tls-certs\") pod \"horizon-cf46f664-8k6s8\" (UID: \"778430ca-2e98-4156-b1d9-fdf05b76ef7e\") " pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.322350 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/778430ca-2e98-4156-b1d9-fdf05b76ef7e-config-data\") pod \"horizon-cf46f664-8k6s8\" (UID: \"778430ca-2e98-4156-b1d9-fdf05b76ef7e\") " pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.340373 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778430ca-2e98-4156-b1d9-fdf05b76ef7e-combined-ca-bundle\") pod \"horizon-cf46f664-8k6s8\" (UID: \"778430ca-2e98-4156-b1d9-fdf05b76ef7e\") " pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.340948 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/778430ca-2e98-4156-b1d9-fdf05b76ef7e-horizon-secret-key\") pod \"horizon-cf46f664-8k6s8\" (UID: \"778430ca-2e98-4156-b1d9-fdf05b76ef7e\") " pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.374291 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztvdn\" (UniqueName: \"kubernetes.io/projected/778430ca-2e98-4156-b1d9-fdf05b76ef7e-kube-api-access-ztvdn\") pod \"horizon-cf46f664-8k6s8\" (UID: \"778430ca-2e98-4156-b1d9-fdf05b76ef7e\") " pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.489899 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.603104 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"672183ef-ddc2-4b72-b127-e1953eaa6549","Type":"ContainerStarted","Data":"97b2c92430ca8e7a38801980d6adb163f28d4a1554c0a568f8d613709cfcd2ed"} Nov 26 01:44:27 crc kubenswrapper[4766]: I1126 01:44:27.614922 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"99765d7e-a086-4d6a-9f06-3847a50c658e","Type":"ContainerStarted","Data":"975a0e3584508fbf4de378af6006000d3cf8d683877e14d8552379ea41a8837d"} Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.032927 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.139261 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.221938 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-7368-account-create-update-qcqxb" Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.345876 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dcf71c2f-342b-4253-b16e-da87cebb754f-operator-scripts\") pod \"dcf71c2f-342b-4253-b16e-da87cebb754f\" (UID: \"dcf71c2f-342b-4253-b16e-da87cebb754f\") " Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.346565 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kq72v\" (UniqueName: \"kubernetes.io/projected/dcf71c2f-342b-4253-b16e-da87cebb754f-kube-api-access-kq72v\") pod \"dcf71c2f-342b-4253-b16e-da87cebb754f\" (UID: \"dcf71c2f-342b-4253-b16e-da87cebb754f\") " Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.348109 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcf71c2f-342b-4253-b16e-da87cebb754f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dcf71c2f-342b-4253-b16e-da87cebb754f" (UID: "dcf71c2f-342b-4253-b16e-da87cebb754f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.362419 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcf71c2f-342b-4253-b16e-da87cebb754f-kube-api-access-kq72v" (OuterVolumeSpecName: "kube-api-access-kq72v") pod "dcf71c2f-342b-4253-b16e-da87cebb754f" (UID: "dcf71c2f-342b-4253-b16e-da87cebb754f"). InnerVolumeSpecName "kube-api-access-kq72v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.449907 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kq72v\" (UniqueName: \"kubernetes.io/projected/dcf71c2f-342b-4253-b16e-da87cebb754f-kube-api-access-kq72v\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.450267 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dcf71c2f-342b-4253-b16e-da87cebb754f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.485225 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-65bd4d46f6-blndr"] Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.491007 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-8lxnr" Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.552199 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e-operator-scripts\") pod \"aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e\" (UID: \"aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e\") " Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.552474 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwn9k\" (UniqueName: \"kubernetes.io/projected/aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e-kube-api-access-hwn9k\") pod \"aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e\" (UID: \"aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e\") " Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.553120 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e" (UID: "aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.557685 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e-kube-api-access-hwn9k" (OuterVolumeSpecName: "kube-api-access-hwn9k") pod "aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e" (UID: "aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e"). InnerVolumeSpecName "kube-api-access-hwn9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.628519 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-7368-account-create-update-qcqxb" event={"ID":"dcf71c2f-342b-4253-b16e-da87cebb754f","Type":"ContainerDied","Data":"847a8a0d32abee0675da8f8c432d7927cbe03d20210154d0f8861601f94e93f5"} Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.628759 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="847a8a0d32abee0675da8f8c432d7927cbe03d20210154d0f8861601f94e93f5" Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.628529 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-7368-account-create-update-qcqxb" Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.674369 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwn9k\" (UniqueName: \"kubernetes.io/projected/aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e-kube-api-access-hwn9k\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.674406 4766 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.677753 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"672183ef-ddc2-4b72-b127-e1953eaa6549","Type":"ContainerStarted","Data":"53d176eb2027fa77ca44f31ec3f00948dc06bf83e0f7b4cf348ae1405123bad4"} Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.677915 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="672183ef-ddc2-4b72-b127-e1953eaa6549" containerName="glance-log" containerID="cri-o://97b2c92430ca8e7a38801980d6adb163f28d4a1554c0a568f8d613709cfcd2ed" gracePeriod=30 Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.678436 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="672183ef-ddc2-4b72-b127-e1953eaa6549" containerName="glance-httpd" containerID="cri-o://53d176eb2027fa77ca44f31ec3f00948dc06bf83e0f7b4cf348ae1405123bad4" gracePeriod=30 Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.700393 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"99765d7e-a086-4d6a-9f06-3847a50c658e","Type":"ContainerStarted","Data":"c308dba6e1cf38ff5124e2db1f0217049f358f1cab773c1c42f49fb8da2efe5d"} Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.700639 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="99765d7e-a086-4d6a-9f06-3847a50c658e" containerName="glance-log" containerID="cri-o://975a0e3584508fbf4de378af6006000d3cf8d683877e14d8552379ea41a8837d" gracePeriod=30 Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.701008 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="99765d7e-a086-4d6a-9f06-3847a50c658e" containerName="glance-httpd" containerID="cri-o://c308dba6e1cf38ff5124e2db1f0217049f358f1cab773c1c42f49fb8da2efe5d" gracePeriod=30 Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.714124 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65bd4d46f6-blndr" event={"ID":"9e2311f3-2074-4336-ba7c-dc676f60d452","Type":"ContainerStarted","Data":"bbd4e6d9f2a4fe590fe2863099e55e529e07a462d9b22eb61a56a22301cc3930"} Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.718777 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-8lxnr" event={"ID":"aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e","Type":"ContainerDied","Data":"ab0b95667e6017eac76fdc778c5384704a81ddc5a8452c341cc4702ab0053d49"} Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.718817 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab0b95667e6017eac76fdc778c5384704a81ddc5a8452c341cc4702ab0053d49" Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.718870 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-8lxnr" Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.737647 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.737601594 podStartE2EDuration="6.737601594s" podCreationTimestamp="2025-11-26 01:44:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 01:44:28.713664456 +0000 UTC m=+4849.562434886" watchObservedRunningTime="2025-11-26 01:44:28.737601594 +0000 UTC m=+4849.586372024" Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.763924 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-cf46f664-8k6s8"] Nov 26 01:44:28 crc kubenswrapper[4766]: I1126 01:44:28.769045 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.769027245 podStartE2EDuration="6.769027245s" podCreationTimestamp="2025-11-26 01:44:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 01:44:28.761204673 +0000 UTC m=+4849.609975103" watchObservedRunningTime="2025-11-26 01:44:28.769027245 +0000 UTC m=+4849.617797665" Nov 26 01:44:28 crc kubenswrapper[4766]: W1126 01:44:28.835925 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod778430ca_2e98_4156_b1d9_fdf05b76ef7e.slice/crio-9ca294933aaa2d05803c21d8aba1b12b8823a0ee3eabcff5f994fcb76967e310 WatchSource:0}: Error finding container 9ca294933aaa2d05803c21d8aba1b12b8823a0ee3eabcff5f994fcb76967e310: Status 404 returned error can't find the container with id 9ca294933aaa2d05803c21d8aba1b12b8823a0ee3eabcff5f994fcb76967e310 Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.467481 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.604493 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-combined-ca-bundle\") pod \"672183ef-ddc2-4b72-b127-e1953eaa6549\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.604755 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmvnv\" (UniqueName: \"kubernetes.io/projected/672183ef-ddc2-4b72-b127-e1953eaa6549-kube-api-access-qmvnv\") pod \"672183ef-ddc2-4b72-b127-e1953eaa6549\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.604821 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-config-data\") pod \"672183ef-ddc2-4b72-b127-e1953eaa6549\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.604854 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/672183ef-ddc2-4b72-b127-e1953eaa6549-httpd-run\") pod \"672183ef-ddc2-4b72-b127-e1953eaa6549\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.604968 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-public-tls-certs\") pod \"672183ef-ddc2-4b72-b127-e1953eaa6549\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.605004 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/672183ef-ddc2-4b72-b127-e1953eaa6549-ceph\") pod \"672183ef-ddc2-4b72-b127-e1953eaa6549\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.605064 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-scripts\") pod \"672183ef-ddc2-4b72-b127-e1953eaa6549\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.605088 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"672183ef-ddc2-4b72-b127-e1953eaa6549\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.605137 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/672183ef-ddc2-4b72-b127-e1953eaa6549-logs\") pod \"672183ef-ddc2-4b72-b127-e1953eaa6549\" (UID: \"672183ef-ddc2-4b72-b127-e1953eaa6549\") " Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.606030 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/672183ef-ddc2-4b72-b127-e1953eaa6549-logs" (OuterVolumeSpecName: "logs") pod "672183ef-ddc2-4b72-b127-e1953eaa6549" (UID: "672183ef-ddc2-4b72-b127-e1953eaa6549"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.606523 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/672183ef-ddc2-4b72-b127-e1953eaa6549-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "672183ef-ddc2-4b72-b127-e1953eaa6549" (UID: "672183ef-ddc2-4b72-b127-e1953eaa6549"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.612206 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/672183ef-ddc2-4b72-b127-e1953eaa6549-ceph" (OuterVolumeSpecName: "ceph") pod "672183ef-ddc2-4b72-b127-e1953eaa6549" (UID: "672183ef-ddc2-4b72-b127-e1953eaa6549"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.612765 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "672183ef-ddc2-4b72-b127-e1953eaa6549" (UID: "672183ef-ddc2-4b72-b127-e1953eaa6549"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.613869 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/672183ef-ddc2-4b72-b127-e1953eaa6549-kube-api-access-qmvnv" (OuterVolumeSpecName: "kube-api-access-qmvnv") pod "672183ef-ddc2-4b72-b127-e1953eaa6549" (UID: "672183ef-ddc2-4b72-b127-e1953eaa6549"). InnerVolumeSpecName "kube-api-access-qmvnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.622939 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-scripts" (OuterVolumeSpecName: "scripts") pod "672183ef-ddc2-4b72-b127-e1953eaa6549" (UID: "672183ef-ddc2-4b72-b127-e1953eaa6549"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.656938 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "672183ef-ddc2-4b72-b127-e1953eaa6549" (UID: "672183ef-ddc2-4b72-b127-e1953eaa6549"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.682258 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-config-data" (OuterVolumeSpecName: "config-data") pod "672183ef-ddc2-4b72-b127-e1953eaa6549" (UID: "672183ef-ddc2-4b72-b127-e1953eaa6549"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.687166 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.691594 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "672183ef-ddc2-4b72-b127-e1953eaa6549" (UID: "672183ef-ddc2-4b72-b127-e1953eaa6549"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.708056 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.708103 4766 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.708116 4766 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/672183ef-ddc2-4b72-b127-e1953eaa6549-logs\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.708125 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.708137 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmvnv\" (UniqueName: \"kubernetes.io/projected/672183ef-ddc2-4b72-b127-e1953eaa6549-kube-api-access-qmvnv\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.708146 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.708154 4766 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/672183ef-ddc2-4b72-b127-e1953eaa6549-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.708162 4766 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/672183ef-ddc2-4b72-b127-e1953eaa6549-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.708170 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/672183ef-ddc2-4b72-b127-e1953eaa6549-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.741969 4766 generic.go:334] "Generic (PLEG): container finished" podID="672183ef-ddc2-4b72-b127-e1953eaa6549" containerID="53d176eb2027fa77ca44f31ec3f00948dc06bf83e0f7b4cf348ae1405123bad4" exitCode=143 Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.742005 4766 generic.go:334] "Generic (PLEG): container finished" podID="672183ef-ddc2-4b72-b127-e1953eaa6549" containerID="97b2c92430ca8e7a38801980d6adb163f28d4a1554c0a568f8d613709cfcd2ed" exitCode=143 Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.742051 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"672183ef-ddc2-4b72-b127-e1953eaa6549","Type":"ContainerDied","Data":"53d176eb2027fa77ca44f31ec3f00948dc06bf83e0f7b4cf348ae1405123bad4"} Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.742080 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"672183ef-ddc2-4b72-b127-e1953eaa6549","Type":"ContainerDied","Data":"97b2c92430ca8e7a38801980d6adb163f28d4a1554c0a568f8d613709cfcd2ed"} Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.742089 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"672183ef-ddc2-4b72-b127-e1953eaa6549","Type":"ContainerDied","Data":"b38a9527fe0612b1cbfe7f17022bb919c83f3527ebfc7502de763f2f180c73d9"} Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.742105 4766 scope.go:117] "RemoveContainer" containerID="53d176eb2027fa77ca44f31ec3f00948dc06bf83e0f7b4cf348ae1405123bad4" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.742260 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.742754 4766 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.757342 4766 generic.go:334] "Generic (PLEG): container finished" podID="99765d7e-a086-4d6a-9f06-3847a50c658e" containerID="c308dba6e1cf38ff5124e2db1f0217049f358f1cab773c1c42f49fb8da2efe5d" exitCode=143 Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.757383 4766 generic.go:334] "Generic (PLEG): container finished" podID="99765d7e-a086-4d6a-9f06-3847a50c658e" containerID="975a0e3584508fbf4de378af6006000d3cf8d683877e14d8552379ea41a8837d" exitCode=143 Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.757443 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"99765d7e-a086-4d6a-9f06-3847a50c658e","Type":"ContainerDied","Data":"c308dba6e1cf38ff5124e2db1f0217049f358f1cab773c1c42f49fb8da2efe5d"} Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.757473 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.757497 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"99765d7e-a086-4d6a-9f06-3847a50c658e","Type":"ContainerDied","Data":"975a0e3584508fbf4de378af6006000d3cf8d683877e14d8552379ea41a8837d"} Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.757511 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"99765d7e-a086-4d6a-9f06-3847a50c658e","Type":"ContainerDied","Data":"b71d7625053683c4bdf259cc6dc4f940babcc83eb34989dec47427d93a19b352"} Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.761073 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cf46f664-8k6s8" event={"ID":"778430ca-2e98-4156-b1d9-fdf05b76ef7e","Type":"ContainerStarted","Data":"9ca294933aaa2d05803c21d8aba1b12b8823a0ee3eabcff5f994fcb76967e310"} Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.801898 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.803123 4766 scope.go:117] "RemoveContainer" containerID="97b2c92430ca8e7a38801980d6adb163f28d4a1554c0a568f8d613709cfcd2ed" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.810806 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"99765d7e-a086-4d6a-9f06-3847a50c658e\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.810886 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hs6rd\" (UniqueName: \"kubernetes.io/projected/99765d7e-a086-4d6a-9f06-3847a50c658e-kube-api-access-hs6rd\") pod \"99765d7e-a086-4d6a-9f06-3847a50c658e\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.811052 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-internal-tls-certs\") pod \"99765d7e-a086-4d6a-9f06-3847a50c658e\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.811248 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/99765d7e-a086-4d6a-9f06-3847a50c658e-ceph\") pod \"99765d7e-a086-4d6a-9f06-3847a50c658e\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.811338 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99765d7e-a086-4d6a-9f06-3847a50c658e-logs\") pod \"99765d7e-a086-4d6a-9f06-3847a50c658e\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.811515 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99765d7e-a086-4d6a-9f06-3847a50c658e-httpd-run\") pod \"99765d7e-a086-4d6a-9f06-3847a50c658e\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.811641 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-config-data\") pod \"99765d7e-a086-4d6a-9f06-3847a50c658e\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.811689 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-scripts\") pod \"99765d7e-a086-4d6a-9f06-3847a50c658e\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.811712 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-combined-ca-bundle\") pod \"99765d7e-a086-4d6a-9f06-3847a50c658e\" (UID: \"99765d7e-a086-4d6a-9f06-3847a50c658e\") " Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.812272 4766 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.816348 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99765d7e-a086-4d6a-9f06-3847a50c658e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "99765d7e-a086-4d6a-9f06-3847a50c658e" (UID: "99765d7e-a086-4d6a-9f06-3847a50c658e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.816742 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99765d7e-a086-4d6a-9f06-3847a50c658e-logs" (OuterVolumeSpecName: "logs") pod "99765d7e-a086-4d6a-9f06-3847a50c658e" (UID: "99765d7e-a086-4d6a-9f06-3847a50c658e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.816873 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99765d7e-a086-4d6a-9f06-3847a50c658e-kube-api-access-hs6rd" (OuterVolumeSpecName: "kube-api-access-hs6rd") pod "99765d7e-a086-4d6a-9f06-3847a50c658e" (UID: "99765d7e-a086-4d6a-9f06-3847a50c658e"). InnerVolumeSpecName "kube-api-access-hs6rd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.826335 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99765d7e-a086-4d6a-9f06-3847a50c658e-ceph" (OuterVolumeSpecName: "ceph") pod "99765d7e-a086-4d6a-9f06-3847a50c658e" (UID: "99765d7e-a086-4d6a-9f06-3847a50c658e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.827556 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "99765d7e-a086-4d6a-9f06-3847a50c658e" (UID: "99765d7e-a086-4d6a-9f06-3847a50c658e"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.853107 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-scripts" (OuterVolumeSpecName: "scripts") pod "99765d7e-a086-4d6a-9f06-3847a50c658e" (UID: "99765d7e-a086-4d6a-9f06-3847a50c658e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.879490 4766 scope.go:117] "RemoveContainer" containerID="53d176eb2027fa77ca44f31ec3f00948dc06bf83e0f7b4cf348ae1405123bad4" Nov 26 01:44:29 crc kubenswrapper[4766]: E1126 01:44:29.880266 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53d176eb2027fa77ca44f31ec3f00948dc06bf83e0f7b4cf348ae1405123bad4\": container with ID starting with 53d176eb2027fa77ca44f31ec3f00948dc06bf83e0f7b4cf348ae1405123bad4 not found: ID does not exist" containerID="53d176eb2027fa77ca44f31ec3f00948dc06bf83e0f7b4cf348ae1405123bad4" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.880362 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53d176eb2027fa77ca44f31ec3f00948dc06bf83e0f7b4cf348ae1405123bad4"} err="failed to get container status \"53d176eb2027fa77ca44f31ec3f00948dc06bf83e0f7b4cf348ae1405123bad4\": rpc error: code = NotFound desc = could not find container \"53d176eb2027fa77ca44f31ec3f00948dc06bf83e0f7b4cf348ae1405123bad4\": container with ID starting with 53d176eb2027fa77ca44f31ec3f00948dc06bf83e0f7b4cf348ae1405123bad4 not found: ID does not exist" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.880398 4766 scope.go:117] "RemoveContainer" containerID="97b2c92430ca8e7a38801980d6adb163f28d4a1554c0a568f8d613709cfcd2ed" Nov 26 01:44:29 crc kubenswrapper[4766]: E1126 01:44:29.880834 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97b2c92430ca8e7a38801980d6adb163f28d4a1554c0a568f8d613709cfcd2ed\": container with ID starting with 97b2c92430ca8e7a38801980d6adb163f28d4a1554c0a568f8d613709cfcd2ed not found: ID does not exist" containerID="97b2c92430ca8e7a38801980d6adb163f28d4a1554c0a568f8d613709cfcd2ed" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.880858 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97b2c92430ca8e7a38801980d6adb163f28d4a1554c0a568f8d613709cfcd2ed"} err="failed to get container status \"97b2c92430ca8e7a38801980d6adb163f28d4a1554c0a568f8d613709cfcd2ed\": rpc error: code = NotFound desc = could not find container \"97b2c92430ca8e7a38801980d6adb163f28d4a1554c0a568f8d613709cfcd2ed\": container with ID starting with 97b2c92430ca8e7a38801980d6adb163f28d4a1554c0a568f8d613709cfcd2ed not found: ID does not exist" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.880872 4766 scope.go:117] "RemoveContainer" containerID="53d176eb2027fa77ca44f31ec3f00948dc06bf83e0f7b4cf348ae1405123bad4" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.881295 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53d176eb2027fa77ca44f31ec3f00948dc06bf83e0f7b4cf348ae1405123bad4"} err="failed to get container status \"53d176eb2027fa77ca44f31ec3f00948dc06bf83e0f7b4cf348ae1405123bad4\": rpc error: code = NotFound desc = could not find container \"53d176eb2027fa77ca44f31ec3f00948dc06bf83e0f7b4cf348ae1405123bad4\": container with ID starting with 53d176eb2027fa77ca44f31ec3f00948dc06bf83e0f7b4cf348ae1405123bad4 not found: ID does not exist" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.881310 4766 scope.go:117] "RemoveContainer" containerID="97b2c92430ca8e7a38801980d6adb163f28d4a1554c0a568f8d613709cfcd2ed" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.881877 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97b2c92430ca8e7a38801980d6adb163f28d4a1554c0a568f8d613709cfcd2ed"} err="failed to get container status \"97b2c92430ca8e7a38801980d6adb163f28d4a1554c0a568f8d613709cfcd2ed\": rpc error: code = NotFound desc = could not find container \"97b2c92430ca8e7a38801980d6adb163f28d4a1554c0a568f8d613709cfcd2ed\": container with ID starting with 97b2c92430ca8e7a38801980d6adb163f28d4a1554c0a568f8d613709cfcd2ed not found: ID does not exist" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.881961 4766 scope.go:117] "RemoveContainer" containerID="c308dba6e1cf38ff5124e2db1f0217049f358f1cab773c1c42f49fb8da2efe5d" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.930571 4766 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99765d7e-a086-4d6a-9f06-3847a50c658e-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.930626 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.942021 4766 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.942104 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hs6rd\" (UniqueName: \"kubernetes.io/projected/99765d7e-a086-4d6a-9f06-3847a50c658e-kube-api-access-hs6rd\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.942128 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/99765d7e-a086-4d6a-9f06-3847a50c658e-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.942140 4766 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99765d7e-a086-4d6a-9f06-3847a50c658e-logs\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.992025 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-config-data" (OuterVolumeSpecName: "config-data") pod "99765d7e-a086-4d6a-9f06-3847a50c658e" (UID: "99765d7e-a086-4d6a-9f06-3847a50c658e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:44:29 crc kubenswrapper[4766]: I1126 01:44:29.998390 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "99765d7e-a086-4d6a-9f06-3847a50c658e" (UID: "99765d7e-a086-4d6a-9f06-3847a50c658e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.048924 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.049192 4766 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.081820 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99765d7e-a086-4d6a-9f06-3847a50c658e" (UID: "99765d7e-a086-4d6a-9f06-3847a50c658e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.132254 4766 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.152083 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99765d7e-a086-4d6a-9f06-3847a50c658e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.152122 4766 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.155256 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.155290 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 01:44:30 crc kubenswrapper[4766]: E1126 01:44:30.155773 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcf71c2f-342b-4253-b16e-da87cebb754f" containerName="mariadb-account-create-update" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.155795 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcf71c2f-342b-4253-b16e-da87cebb754f" containerName="mariadb-account-create-update" Nov 26 01:44:30 crc kubenswrapper[4766]: E1126 01:44:30.155820 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e" containerName="mariadb-database-create" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.155827 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e" containerName="mariadb-database-create" Nov 26 01:44:30 crc kubenswrapper[4766]: E1126 01:44:30.155849 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99765d7e-a086-4d6a-9f06-3847a50c658e" containerName="glance-httpd" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.155855 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="99765d7e-a086-4d6a-9f06-3847a50c658e" containerName="glance-httpd" Nov 26 01:44:30 crc kubenswrapper[4766]: E1126 01:44:30.155869 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="672183ef-ddc2-4b72-b127-e1953eaa6549" containerName="glance-log" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.155875 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="672183ef-ddc2-4b72-b127-e1953eaa6549" containerName="glance-log" Nov 26 01:44:30 crc kubenswrapper[4766]: E1126 01:44:30.155894 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="672183ef-ddc2-4b72-b127-e1953eaa6549" containerName="glance-httpd" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.155900 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="672183ef-ddc2-4b72-b127-e1953eaa6549" containerName="glance-httpd" Nov 26 01:44:30 crc kubenswrapper[4766]: E1126 01:44:30.155912 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99765d7e-a086-4d6a-9f06-3847a50c658e" containerName="glance-log" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.155918 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="99765d7e-a086-4d6a-9f06-3847a50c658e" containerName="glance-log" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.156116 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e" containerName="mariadb-database-create" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.156139 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="99765d7e-a086-4d6a-9f06-3847a50c658e" containerName="glance-log" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.156147 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="672183ef-ddc2-4b72-b127-e1953eaa6549" containerName="glance-httpd" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.156160 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="672183ef-ddc2-4b72-b127-e1953eaa6549" containerName="glance-log" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.156168 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcf71c2f-342b-4253-b16e-da87cebb754f" containerName="mariadb-account-create-update" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.156185 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="99765d7e-a086-4d6a-9f06-3847a50c658e" containerName="glance-httpd" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.157239 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.157336 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.164296 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.164458 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.246750 4766 scope.go:117] "RemoveContainer" containerID="975a0e3584508fbf4de378af6006000d3cf8d683877e14d8552379ea41a8837d" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.253901 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5ea31542-fdc7-481a-853a-5935e39fca92-ceph\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.253945 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ea31542-fdc7-481a-853a-5935e39fca92-config-data\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.253974 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ea31542-fdc7-481a-853a-5935e39fca92-scripts\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.254325 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ea31542-fdc7-481a-853a-5935e39fca92-logs\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.254509 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.254536 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ea31542-fdc7-481a-853a-5935e39fca92-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.254585 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ea31542-fdc7-481a-853a-5935e39fca92-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.254613 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5ea31542-fdc7-481a-853a-5935e39fca92-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.254702 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mr559\" (UniqueName: \"kubernetes.io/projected/5ea31542-fdc7-481a-853a-5935e39fca92-kube-api-access-mr559\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.279423 4766 scope.go:117] "RemoveContainer" containerID="c308dba6e1cf38ff5124e2db1f0217049f358f1cab773c1c42f49fb8da2efe5d" Nov 26 01:44:30 crc kubenswrapper[4766]: E1126 01:44:30.279976 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c308dba6e1cf38ff5124e2db1f0217049f358f1cab773c1c42f49fb8da2efe5d\": container with ID starting with c308dba6e1cf38ff5124e2db1f0217049f358f1cab773c1c42f49fb8da2efe5d not found: ID does not exist" containerID="c308dba6e1cf38ff5124e2db1f0217049f358f1cab773c1c42f49fb8da2efe5d" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.280032 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c308dba6e1cf38ff5124e2db1f0217049f358f1cab773c1c42f49fb8da2efe5d"} err="failed to get container status \"c308dba6e1cf38ff5124e2db1f0217049f358f1cab773c1c42f49fb8da2efe5d\": rpc error: code = NotFound desc = could not find container \"c308dba6e1cf38ff5124e2db1f0217049f358f1cab773c1c42f49fb8da2efe5d\": container with ID starting with c308dba6e1cf38ff5124e2db1f0217049f358f1cab773c1c42f49fb8da2efe5d not found: ID does not exist" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.280076 4766 scope.go:117] "RemoveContainer" containerID="975a0e3584508fbf4de378af6006000d3cf8d683877e14d8552379ea41a8837d" Nov 26 01:44:30 crc kubenswrapper[4766]: E1126 01:44:30.280573 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"975a0e3584508fbf4de378af6006000d3cf8d683877e14d8552379ea41a8837d\": container with ID starting with 975a0e3584508fbf4de378af6006000d3cf8d683877e14d8552379ea41a8837d not found: ID does not exist" containerID="975a0e3584508fbf4de378af6006000d3cf8d683877e14d8552379ea41a8837d" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.280613 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"975a0e3584508fbf4de378af6006000d3cf8d683877e14d8552379ea41a8837d"} err="failed to get container status \"975a0e3584508fbf4de378af6006000d3cf8d683877e14d8552379ea41a8837d\": rpc error: code = NotFound desc = could not find container \"975a0e3584508fbf4de378af6006000d3cf8d683877e14d8552379ea41a8837d\": container with ID starting with 975a0e3584508fbf4de378af6006000d3cf8d683877e14d8552379ea41a8837d not found: ID does not exist" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.280642 4766 scope.go:117] "RemoveContainer" containerID="c308dba6e1cf38ff5124e2db1f0217049f358f1cab773c1c42f49fb8da2efe5d" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.281380 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c308dba6e1cf38ff5124e2db1f0217049f358f1cab773c1c42f49fb8da2efe5d"} err="failed to get container status \"c308dba6e1cf38ff5124e2db1f0217049f358f1cab773c1c42f49fb8da2efe5d\": rpc error: code = NotFound desc = could not find container \"c308dba6e1cf38ff5124e2db1f0217049f358f1cab773c1c42f49fb8da2efe5d\": container with ID starting with c308dba6e1cf38ff5124e2db1f0217049f358f1cab773c1c42f49fb8da2efe5d not found: ID does not exist" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.281403 4766 scope.go:117] "RemoveContainer" containerID="975a0e3584508fbf4de378af6006000d3cf8d683877e14d8552379ea41a8837d" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.281849 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"975a0e3584508fbf4de378af6006000d3cf8d683877e14d8552379ea41a8837d"} err="failed to get container status \"975a0e3584508fbf4de378af6006000d3cf8d683877e14d8552379ea41a8837d\": rpc error: code = NotFound desc = could not find container \"975a0e3584508fbf4de378af6006000d3cf8d683877e14d8552379ea41a8837d\": container with ID starting with 975a0e3584508fbf4de378af6006000d3cf8d683877e14d8552379ea41a8837d not found: ID does not exist" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.356686 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ea31542-fdc7-481a-853a-5935e39fca92-logs\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.356779 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.356798 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ea31542-fdc7-481a-853a-5935e39fca92-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.356826 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ea31542-fdc7-481a-853a-5935e39fca92-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.356846 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5ea31542-fdc7-481a-853a-5935e39fca92-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.356882 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mr559\" (UniqueName: \"kubernetes.io/projected/5ea31542-fdc7-481a-853a-5935e39fca92-kube-api-access-mr559\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.356925 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5ea31542-fdc7-481a-853a-5935e39fca92-ceph\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.356942 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ea31542-fdc7-481a-853a-5935e39fca92-config-data\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.356961 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ea31542-fdc7-481a-853a-5935e39fca92-scripts\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.357161 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ea31542-fdc7-481a-853a-5935e39fca92-logs\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.357493 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5ea31542-fdc7-481a-853a-5935e39fca92-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.357862 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.364926 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ea31542-fdc7-481a-853a-5935e39fca92-config-data\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.365452 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5ea31542-fdc7-481a-853a-5935e39fca92-ceph\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.366217 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ea31542-fdc7-481a-853a-5935e39fca92-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.366997 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ea31542-fdc7-481a-853a-5935e39fca92-scripts\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.372224 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ea31542-fdc7-481a-853a-5935e39fca92-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.374594 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mr559\" (UniqueName: \"kubernetes.io/projected/5ea31542-fdc7-481a-853a-5935e39fca92-kube-api-access-mr559\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.397502 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"5ea31542-fdc7-481a-853a-5935e39fca92\") " pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.446895 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.484227 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.484793 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.515625 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.517696 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.521039 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.521309 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.527918 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.665982 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npggz\" (UniqueName: \"kubernetes.io/projected/f5f6c81c-04d9-4542-9ab1-b932737f388c-kube-api-access-npggz\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.666039 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f5f6c81c-04d9-4542-9ab1-b932737f388c-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.666073 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5f6c81c-04d9-4542-9ab1-b932737f388c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.666105 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.666126 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5f6c81c-04d9-4542-9ab1-b932737f388c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.666168 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5f6c81c-04d9-4542-9ab1-b932737f388c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.666208 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5f6c81c-04d9-4542-9ab1-b932737f388c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.666233 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5f6c81c-04d9-4542-9ab1-b932737f388c-logs\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.666260 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5f6c81c-04d9-4542-9ab1-b932737f388c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.768394 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f5f6c81c-04d9-4542-9ab1-b932737f388c-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.768448 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5f6c81c-04d9-4542-9ab1-b932737f388c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.768483 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.768506 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5f6c81c-04d9-4542-9ab1-b932737f388c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.768560 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5f6c81c-04d9-4542-9ab1-b932737f388c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.768603 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5f6c81c-04d9-4542-9ab1-b932737f388c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.768627 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5f6c81c-04d9-4542-9ab1-b932737f388c-logs\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.768673 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5f6c81c-04d9-4542-9ab1-b932737f388c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.768791 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npggz\" (UniqueName: \"kubernetes.io/projected/f5f6c81c-04d9-4542-9ab1-b932737f388c-kube-api-access-npggz\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.769984 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.775971 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f5f6c81c-04d9-4542-9ab1-b932737f388c-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.776321 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5f6c81c-04d9-4542-9ab1-b932737f388c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.776557 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5f6c81c-04d9-4542-9ab1-b932737f388c-logs\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.776987 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5f6c81c-04d9-4542-9ab1-b932737f388c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.780734 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5f6c81c-04d9-4542-9ab1-b932737f388c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.790769 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5f6c81c-04d9-4542-9ab1-b932737f388c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.796393 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npggz\" (UniqueName: \"kubernetes.io/projected/f5f6c81c-04d9-4542-9ab1-b932737f388c-kube-api-access-npggz\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.822054 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5f6c81c-04d9-4542-9ab1-b932737f388c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:30 crc kubenswrapper[4766]: I1126 01:44:30.847840 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f5f6c81c-04d9-4542-9ab1-b932737f388c\") " pod="openstack/glance-default-internal-api-0" Nov 26 01:44:31 crc kubenswrapper[4766]: I1126 01:44:31.138490 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 01:44:31 crc kubenswrapper[4766]: I1126 01:44:31.142176 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 26 01:44:31 crc kubenswrapper[4766]: I1126 01:44:31.777563 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 01:44:31 crc kubenswrapper[4766]: I1126 01:44:31.846117 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="672183ef-ddc2-4b72-b127-e1953eaa6549" path="/var/lib/kubelet/pods/672183ef-ddc2-4b72-b127-e1953eaa6549/volumes" Nov 26 01:44:31 crc kubenswrapper[4766]: I1126 01:44:31.847020 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99765d7e-a086-4d6a-9f06-3847a50c658e" path="/var/lib/kubelet/pods/99765d7e-a086-4d6a-9f06-3847a50c658e/volumes" Nov 26 01:44:31 crc kubenswrapper[4766]: I1126 01:44:31.847886 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5ea31542-fdc7-481a-853a-5935e39fca92","Type":"ContainerStarted","Data":"7729ac67beda6d44f287457bbc95074ecff1ea046c99bb6c5b182fd9d41c0e10"} Nov 26 01:44:31 crc kubenswrapper[4766]: I1126 01:44:31.847914 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f5f6c81c-04d9-4542-9ab1-b932737f388c","Type":"ContainerStarted","Data":"287074223ae074a5cc2f61ac6d1559c2a2cd25860e3d035845e14e199fdc511f"} Nov 26 01:44:32 crc kubenswrapper[4766]: I1126 01:44:32.864425 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f5f6c81c-04d9-4542-9ab1-b932737f388c","Type":"ContainerStarted","Data":"23d64be2e7c5cfefc6380cfa32f5b9bbef2ecd71202218de93927ea33df10c1f"} Nov 26 01:44:32 crc kubenswrapper[4766]: I1126 01:44:32.874523 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5ea31542-fdc7-481a-853a-5935e39fca92","Type":"ContainerStarted","Data":"966a0b4c6e1307f792734a109be5e726a66847a72e54414f6326235f7d378743"} Nov 26 01:44:33 crc kubenswrapper[4766]: I1126 01:44:33.332097 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Nov 26 01:44:33 crc kubenswrapper[4766]: I1126 01:44:33.408177 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Nov 26 01:44:33 crc kubenswrapper[4766]: I1126 01:44:33.888135 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5ea31542-fdc7-481a-853a-5935e39fca92","Type":"ContainerStarted","Data":"3c8ab2ec034913a6016f684f66670941c0d58af38941dab44fa65220e69c3c7d"} Nov 26 01:44:33 crc kubenswrapper[4766]: I1126 01:44:33.902858 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f5f6c81c-04d9-4542-9ab1-b932737f388c","Type":"ContainerStarted","Data":"d5871fe569c102a0a36ff9758ba3165996a5cce67e364740ca883c5a34ff893a"} Nov 26 01:44:33 crc kubenswrapper[4766]: I1126 01:44:33.930448 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.9304292069999995 podStartE2EDuration="4.930429207s" podCreationTimestamp="2025-11-26 01:44:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 01:44:33.92196389 +0000 UTC m=+4854.770734310" watchObservedRunningTime="2025-11-26 01:44:33.930429207 +0000 UTC m=+4854.779199637" Nov 26 01:44:33 crc kubenswrapper[4766]: I1126 01:44:33.950617 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.950594972 podStartE2EDuration="3.950594972s" podCreationTimestamp="2025-11-26 01:44:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 01:44:33.945172419 +0000 UTC m=+4854.793942849" watchObservedRunningTime="2025-11-26 01:44:33.950594972 +0000 UTC m=+4854.799365402" Nov 26 01:44:34 crc kubenswrapper[4766]: I1126 01:44:34.026976 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-dcdnc"] Nov 26 01:44:34 crc kubenswrapper[4766]: I1126 01:44:34.028497 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-dcdnc" Nov 26 01:44:34 crc kubenswrapper[4766]: I1126 01:44:34.031919 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Nov 26 01:44:34 crc kubenswrapper[4766]: I1126 01:44:34.031922 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-k9p8p" Nov 26 01:44:34 crc kubenswrapper[4766]: I1126 01:44:34.042979 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9ljq\" (UniqueName: \"kubernetes.io/projected/18eb3d3c-9197-4730-b085-31e77d7741f1-kube-api-access-d9ljq\") pod \"manila-db-sync-dcdnc\" (UID: \"18eb3d3c-9197-4730-b085-31e77d7741f1\") " pod="openstack/manila-db-sync-dcdnc" Nov 26 01:44:34 crc kubenswrapper[4766]: I1126 01:44:34.043029 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/18eb3d3c-9197-4730-b085-31e77d7741f1-job-config-data\") pod \"manila-db-sync-dcdnc\" (UID: \"18eb3d3c-9197-4730-b085-31e77d7741f1\") " pod="openstack/manila-db-sync-dcdnc" Nov 26 01:44:34 crc kubenswrapper[4766]: I1126 01:44:34.043098 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18eb3d3c-9197-4730-b085-31e77d7741f1-combined-ca-bundle\") pod \"manila-db-sync-dcdnc\" (UID: \"18eb3d3c-9197-4730-b085-31e77d7741f1\") " pod="openstack/manila-db-sync-dcdnc" Nov 26 01:44:34 crc kubenswrapper[4766]: I1126 01:44:34.043137 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18eb3d3c-9197-4730-b085-31e77d7741f1-config-data\") pod \"manila-db-sync-dcdnc\" (UID: \"18eb3d3c-9197-4730-b085-31e77d7741f1\") " pod="openstack/manila-db-sync-dcdnc" Nov 26 01:44:34 crc kubenswrapper[4766]: I1126 01:44:34.055238 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-dcdnc"] Nov 26 01:44:34 crc kubenswrapper[4766]: I1126 01:44:34.145127 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18eb3d3c-9197-4730-b085-31e77d7741f1-combined-ca-bundle\") pod \"manila-db-sync-dcdnc\" (UID: \"18eb3d3c-9197-4730-b085-31e77d7741f1\") " pod="openstack/manila-db-sync-dcdnc" Nov 26 01:44:34 crc kubenswrapper[4766]: I1126 01:44:34.145204 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18eb3d3c-9197-4730-b085-31e77d7741f1-config-data\") pod \"manila-db-sync-dcdnc\" (UID: \"18eb3d3c-9197-4730-b085-31e77d7741f1\") " pod="openstack/manila-db-sync-dcdnc" Nov 26 01:44:34 crc kubenswrapper[4766]: I1126 01:44:34.145346 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9ljq\" (UniqueName: \"kubernetes.io/projected/18eb3d3c-9197-4730-b085-31e77d7741f1-kube-api-access-d9ljq\") pod \"manila-db-sync-dcdnc\" (UID: \"18eb3d3c-9197-4730-b085-31e77d7741f1\") " pod="openstack/manila-db-sync-dcdnc" Nov 26 01:44:34 crc kubenswrapper[4766]: I1126 01:44:34.145384 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/18eb3d3c-9197-4730-b085-31e77d7741f1-job-config-data\") pod \"manila-db-sync-dcdnc\" (UID: \"18eb3d3c-9197-4730-b085-31e77d7741f1\") " pod="openstack/manila-db-sync-dcdnc" Nov 26 01:44:34 crc kubenswrapper[4766]: I1126 01:44:34.153768 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/18eb3d3c-9197-4730-b085-31e77d7741f1-job-config-data\") pod \"manila-db-sync-dcdnc\" (UID: \"18eb3d3c-9197-4730-b085-31e77d7741f1\") " pod="openstack/manila-db-sync-dcdnc" Nov 26 01:44:34 crc kubenswrapper[4766]: I1126 01:44:34.154682 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18eb3d3c-9197-4730-b085-31e77d7741f1-combined-ca-bundle\") pod \"manila-db-sync-dcdnc\" (UID: \"18eb3d3c-9197-4730-b085-31e77d7741f1\") " pod="openstack/manila-db-sync-dcdnc" Nov 26 01:44:34 crc kubenswrapper[4766]: I1126 01:44:34.156748 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18eb3d3c-9197-4730-b085-31e77d7741f1-config-data\") pod \"manila-db-sync-dcdnc\" (UID: \"18eb3d3c-9197-4730-b085-31e77d7741f1\") " pod="openstack/manila-db-sync-dcdnc" Nov 26 01:44:34 crc kubenswrapper[4766]: I1126 01:44:34.171228 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9ljq\" (UniqueName: \"kubernetes.io/projected/18eb3d3c-9197-4730-b085-31e77d7741f1-kube-api-access-d9ljq\") pod \"manila-db-sync-dcdnc\" (UID: \"18eb3d3c-9197-4730-b085-31e77d7741f1\") " pod="openstack/manila-db-sync-dcdnc" Nov 26 01:44:34 crc kubenswrapper[4766]: I1126 01:44:34.363372 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-dcdnc" Nov 26 01:44:39 crc kubenswrapper[4766]: I1126 01:44:39.992094 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65bd4d46f6-blndr" event={"ID":"9e2311f3-2074-4336-ba7c-dc676f60d452","Type":"ContainerStarted","Data":"331dfc0a6e5abb0e07ca1ea14a78e5f63c07bd07abbf90a2b2dc0c61ad816a52"} Nov 26 01:44:39 crc kubenswrapper[4766]: I1126 01:44:39.995370 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f7f64877f-4sskz" event={"ID":"656cb04b-f831-4671-a102-d01c6d01714f","Type":"ContainerStarted","Data":"fe8bea5f9be1857687215d30d22402fc1211456a7511bb72ccd1c779511c43e5"} Nov 26 01:44:40 crc kubenswrapper[4766]: I1126 01:44:40.002012 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f5b87f5c-z9b2b" event={"ID":"c577b8bb-a3c5-405d-8da8-0289e8483ba5","Type":"ContainerStarted","Data":"1c619971b704f1525550b524ab022265c6c1aaa8fd615b299de74ec0f58aaaae"} Nov 26 01:44:40 crc kubenswrapper[4766]: I1126 01:44:40.006736 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cf46f664-8k6s8" event={"ID":"778430ca-2e98-4156-b1d9-fdf05b76ef7e","Type":"ContainerStarted","Data":"6a17153678bdca8d87750a40ae53c8ac811ca23557fc544d590bc1473cabf870"} Nov 26 01:44:40 crc kubenswrapper[4766]: I1126 01:44:40.113231 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-dcdnc"] Nov 26 01:44:40 crc kubenswrapper[4766]: W1126 01:44:40.130000 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18eb3d3c_9197_4730_b085_31e77d7741f1.slice/crio-b16f80a8f88190eecd3a32ef1210ffd950921333a40c9f3e69d405bef845358b WatchSource:0}: Error finding container b16f80a8f88190eecd3a32ef1210ffd950921333a40c9f3e69d405bef845358b: Status 404 returned error can't find the container with id b16f80a8f88190eecd3a32ef1210ffd950921333a40c9f3e69d405bef845358b Nov 26 01:44:40 crc kubenswrapper[4766]: I1126 01:44:40.486421 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 26 01:44:40 crc kubenswrapper[4766]: I1126 01:44:40.486763 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 26 01:44:40 crc kubenswrapper[4766]: I1126 01:44:40.528389 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 26 01:44:40 crc kubenswrapper[4766]: I1126 01:44:40.541205 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 26 01:44:41 crc kubenswrapper[4766]: I1126 01:44:41.021119 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-dcdnc" event={"ID":"18eb3d3c-9197-4730-b085-31e77d7741f1","Type":"ContainerStarted","Data":"b16f80a8f88190eecd3a32ef1210ffd950921333a40c9f3e69d405bef845358b"} Nov 26 01:44:41 crc kubenswrapper[4766]: I1126 01:44:41.022971 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cf46f664-8k6s8" event={"ID":"778430ca-2e98-4156-b1d9-fdf05b76ef7e","Type":"ContainerStarted","Data":"ba22a7ddc587ab6739dd2919be421ec51db1c90797e17bed5171685e91992a57"} Nov 26 01:44:41 crc kubenswrapper[4766]: I1126 01:44:41.026352 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65bd4d46f6-blndr" event={"ID":"9e2311f3-2074-4336-ba7c-dc676f60d452","Type":"ContainerStarted","Data":"425ff6f4884d714d97e857b8513f474a1d431613a4dc5063d13ebf02dcb8adad"} Nov 26 01:44:41 crc kubenswrapper[4766]: I1126 01:44:41.029039 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f7f64877f-4sskz" event={"ID":"656cb04b-f831-4671-a102-d01c6d01714f","Type":"ContainerStarted","Data":"dcfd1bcc023c09f3d1070f702e5f6368c5444f8c30ec70ebff460066679e4f44"} Nov 26 01:44:41 crc kubenswrapper[4766]: I1126 01:44:41.029153 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f7f64877f-4sskz" podUID="656cb04b-f831-4671-a102-d01c6d01714f" containerName="horizon-log" containerID="cri-o://fe8bea5f9be1857687215d30d22402fc1211456a7511bb72ccd1c779511c43e5" gracePeriod=30 Nov 26 01:44:41 crc kubenswrapper[4766]: I1126 01:44:41.029383 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f7f64877f-4sskz" podUID="656cb04b-f831-4671-a102-d01c6d01714f" containerName="horizon" containerID="cri-o://dcfd1bcc023c09f3d1070f702e5f6368c5444f8c30ec70ebff460066679e4f44" gracePeriod=30 Nov 26 01:44:41 crc kubenswrapper[4766]: I1126 01:44:41.036353 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f5b87f5c-z9b2b" event={"ID":"c577b8bb-a3c5-405d-8da8-0289e8483ba5","Type":"ContainerStarted","Data":"797e1e95d5a6b9de52f72bb3fef2fafe5c0c40666f3b784bc4621f22407d05f2"} Nov 26 01:44:41 crc kubenswrapper[4766]: I1126 01:44:41.036337 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f5b87f5c-z9b2b" podUID="c577b8bb-a3c5-405d-8da8-0289e8483ba5" containerName="horizon-log" containerID="cri-o://1c619971b704f1525550b524ab022265c6c1aaa8fd615b299de74ec0f58aaaae" gracePeriod=30 Nov 26 01:44:41 crc kubenswrapper[4766]: I1126 01:44:41.036597 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 26 01:44:41 crc kubenswrapper[4766]: I1126 01:44:41.037031 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 26 01:44:41 crc kubenswrapper[4766]: I1126 01:44:41.037978 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f5b87f5c-z9b2b" podUID="c577b8bb-a3c5-405d-8da8-0289e8483ba5" containerName="horizon" containerID="cri-o://797e1e95d5a6b9de52f72bb3fef2fafe5c0c40666f3b784bc4621f22407d05f2" gracePeriod=30 Nov 26 01:44:41 crc kubenswrapper[4766]: I1126 01:44:41.060067 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-cf46f664-8k6s8" podStartSLOduration=4.260652765 podStartE2EDuration="15.060040231s" podCreationTimestamp="2025-11-26 01:44:26 +0000 UTC" firstStartedPulling="2025-11-26 01:44:28.837989216 +0000 UTC m=+4849.686759646" lastFinishedPulling="2025-11-26 01:44:39.637376682 +0000 UTC m=+4860.486147112" observedRunningTime="2025-11-26 01:44:41.045832192 +0000 UTC m=+4861.894602632" watchObservedRunningTime="2025-11-26 01:44:41.060040231 +0000 UTC m=+4861.908810701" Nov 26 01:44:41 crc kubenswrapper[4766]: I1126 01:44:41.083820 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7f7f64877f-4sskz" podStartSLOduration=3.96045015 podStartE2EDuration="18.083799244s" podCreationTimestamp="2025-11-26 01:44:23 +0000 UTC" firstStartedPulling="2025-11-26 01:44:25.489026375 +0000 UTC m=+4846.337796805" lastFinishedPulling="2025-11-26 01:44:39.612375469 +0000 UTC m=+4860.461145899" observedRunningTime="2025-11-26 01:44:41.070218991 +0000 UTC m=+4861.918989421" watchObservedRunningTime="2025-11-26 01:44:41.083799244 +0000 UTC m=+4861.932569684" Nov 26 01:44:41 crc kubenswrapper[4766]: I1126 01:44:41.100252 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7f5b87f5c-z9b2b" podStartSLOduration=4.107278693 podStartE2EDuration="18.100230407s" podCreationTimestamp="2025-11-26 01:44:23 +0000 UTC" firstStartedPulling="2025-11-26 01:44:25.620808548 +0000 UTC m=+4846.469578978" lastFinishedPulling="2025-11-26 01:44:39.613760262 +0000 UTC m=+4860.462530692" observedRunningTime="2025-11-26 01:44:41.09382514 +0000 UTC m=+4861.942595580" watchObservedRunningTime="2025-11-26 01:44:41.100230407 +0000 UTC m=+4861.949000847" Nov 26 01:44:41 crc kubenswrapper[4766]: I1126 01:44:41.143086 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 26 01:44:41 crc kubenswrapper[4766]: I1126 01:44:41.143132 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 26 01:44:41 crc kubenswrapper[4766]: I1126 01:44:41.152636 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-65bd4d46f6-blndr" podStartSLOduration=4.16785741 podStartE2EDuration="15.152612762s" podCreationTimestamp="2025-11-26 01:44:26 +0000 UTC" firstStartedPulling="2025-11-26 01:44:28.531085918 +0000 UTC m=+4849.379856348" lastFinishedPulling="2025-11-26 01:44:39.51584127 +0000 UTC m=+4860.364611700" observedRunningTime="2025-11-26 01:44:41.141919679 +0000 UTC m=+4861.990690099" watchObservedRunningTime="2025-11-26 01:44:41.152612762 +0000 UTC m=+4862.001383192" Nov 26 01:44:41 crc kubenswrapper[4766]: I1126 01:44:41.211286 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 26 01:44:41 crc kubenswrapper[4766]: I1126 01:44:41.218247 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 26 01:44:42 crc kubenswrapper[4766]: I1126 01:44:42.052355 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 26 01:44:42 crc kubenswrapper[4766]: I1126 01:44:42.052672 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 26 01:44:43 crc kubenswrapper[4766]: I1126 01:44:43.066805 4766 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 01:44:43 crc kubenswrapper[4766]: I1126 01:44:43.066841 4766 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 01:44:44 crc kubenswrapper[4766]: I1126 01:44:44.511796 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7f5b87f5c-z9b2b" Nov 26 01:44:44 crc kubenswrapper[4766]: I1126 01:44:44.524752 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7f7f64877f-4sskz" Nov 26 01:44:45 crc kubenswrapper[4766]: I1126 01:44:45.197450 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 26 01:44:45 crc kubenswrapper[4766]: I1126 01:44:45.197784 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 26 01:44:45 crc kubenswrapper[4766]: I1126 01:44:45.208631 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 26 01:44:45 crc kubenswrapper[4766]: I1126 01:44:45.208767 4766 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 01:44:45 crc kubenswrapper[4766]: I1126 01:44:45.209349 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 26 01:44:47 crc kubenswrapper[4766]: I1126 01:44:47.307189 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:47 crc kubenswrapper[4766]: I1126 01:44:47.307510 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:44:47 crc kubenswrapper[4766]: I1126 01:44:47.490503 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:47 crc kubenswrapper[4766]: I1126 01:44:47.490548 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:44:49 crc kubenswrapper[4766]: I1126 01:44:49.149891 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-dcdnc" event={"ID":"18eb3d3c-9197-4730-b085-31e77d7741f1","Type":"ContainerStarted","Data":"67dfd0eeeda51abb33036c92ddbe133254089a8baec53a765fef53db80e54c18"} Nov 26 01:44:49 crc kubenswrapper[4766]: I1126 01:44:49.177449 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-dcdnc" podStartSLOduration=8.817793468 podStartE2EDuration="16.177434935s" podCreationTimestamp="2025-11-26 01:44:33 +0000 UTC" firstStartedPulling="2025-11-26 01:44:40.132267742 +0000 UTC m=+4860.981038182" lastFinishedPulling="2025-11-26 01:44:47.491909219 +0000 UTC m=+4868.340679649" observedRunningTime="2025-11-26 01:44:49.174676167 +0000 UTC m=+4870.023446597" watchObservedRunningTime="2025-11-26 01:44:49.177434935 +0000 UTC m=+4870.026205365" Nov 26 01:44:57 crc kubenswrapper[4766]: I1126 01:44:57.308462 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-65bd4d46f6-blndr" podUID="9e2311f3-2074-4336-ba7c-dc676f60d452" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.71:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.71:8443: connect: connection refused" Nov 26 01:44:57 crc kubenswrapper[4766]: I1126 01:44:57.492620 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-cf46f664-8k6s8" podUID="778430ca-2e98-4156-b1d9-fdf05b76ef7e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.72:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.72:8443: connect: connection refused" Nov 26 01:44:59 crc kubenswrapper[4766]: I1126 01:44:59.281407 4766 generic.go:334] "Generic (PLEG): container finished" podID="18eb3d3c-9197-4730-b085-31e77d7741f1" containerID="67dfd0eeeda51abb33036c92ddbe133254089a8baec53a765fef53db80e54c18" exitCode=0 Nov 26 01:44:59 crc kubenswrapper[4766]: I1126 01:44:59.281583 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-dcdnc" event={"ID":"18eb3d3c-9197-4730-b085-31e77d7741f1","Type":"ContainerDied","Data":"67dfd0eeeda51abb33036c92ddbe133254089a8baec53a765fef53db80e54c18"} Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.160328 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402025-vsmwp"] Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.163426 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402025-vsmwp" Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.165941 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.170217 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.174064 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402025-vsmwp"] Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.249538 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb60fd86-c072-40b9-a2e3-047f3b83c5bd-config-volume\") pod \"collect-profiles-29402025-vsmwp\" (UID: \"bb60fd86-c072-40b9-a2e3-047f3b83c5bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402025-vsmwp" Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.250148 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw647\" (UniqueName: \"kubernetes.io/projected/bb60fd86-c072-40b9-a2e3-047f3b83c5bd-kube-api-access-kw647\") pod \"collect-profiles-29402025-vsmwp\" (UID: \"bb60fd86-c072-40b9-a2e3-047f3b83c5bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402025-vsmwp" Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.250198 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb60fd86-c072-40b9-a2e3-047f3b83c5bd-secret-volume\") pod \"collect-profiles-29402025-vsmwp\" (UID: \"bb60fd86-c072-40b9-a2e3-047f3b83c5bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402025-vsmwp" Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.352329 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw647\" (UniqueName: \"kubernetes.io/projected/bb60fd86-c072-40b9-a2e3-047f3b83c5bd-kube-api-access-kw647\") pod \"collect-profiles-29402025-vsmwp\" (UID: \"bb60fd86-c072-40b9-a2e3-047f3b83c5bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402025-vsmwp" Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.352383 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb60fd86-c072-40b9-a2e3-047f3b83c5bd-secret-volume\") pod \"collect-profiles-29402025-vsmwp\" (UID: \"bb60fd86-c072-40b9-a2e3-047f3b83c5bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402025-vsmwp" Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.352433 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb60fd86-c072-40b9-a2e3-047f3b83c5bd-config-volume\") pod \"collect-profiles-29402025-vsmwp\" (UID: \"bb60fd86-c072-40b9-a2e3-047f3b83c5bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402025-vsmwp" Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.353523 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb60fd86-c072-40b9-a2e3-047f3b83c5bd-config-volume\") pod \"collect-profiles-29402025-vsmwp\" (UID: \"bb60fd86-c072-40b9-a2e3-047f3b83c5bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402025-vsmwp" Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.359399 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb60fd86-c072-40b9-a2e3-047f3b83c5bd-secret-volume\") pod \"collect-profiles-29402025-vsmwp\" (UID: \"bb60fd86-c072-40b9-a2e3-047f3b83c5bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402025-vsmwp" Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.372128 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw647\" (UniqueName: \"kubernetes.io/projected/bb60fd86-c072-40b9-a2e3-047f3b83c5bd-kube-api-access-kw647\") pod \"collect-profiles-29402025-vsmwp\" (UID: \"bb60fd86-c072-40b9-a2e3-047f3b83c5bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402025-vsmwp" Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.505838 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402025-vsmwp" Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.824728 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-dcdnc" Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.964765 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18eb3d3c-9197-4730-b085-31e77d7741f1-combined-ca-bundle\") pod \"18eb3d3c-9197-4730-b085-31e77d7741f1\" (UID: \"18eb3d3c-9197-4730-b085-31e77d7741f1\") " Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.965101 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/18eb3d3c-9197-4730-b085-31e77d7741f1-job-config-data\") pod \"18eb3d3c-9197-4730-b085-31e77d7741f1\" (UID: \"18eb3d3c-9197-4730-b085-31e77d7741f1\") " Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.965663 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18eb3d3c-9197-4730-b085-31e77d7741f1-config-data\") pod \"18eb3d3c-9197-4730-b085-31e77d7741f1\" (UID: \"18eb3d3c-9197-4730-b085-31e77d7741f1\") " Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.965688 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9ljq\" (UniqueName: \"kubernetes.io/projected/18eb3d3c-9197-4730-b085-31e77d7741f1-kube-api-access-d9ljq\") pod \"18eb3d3c-9197-4730-b085-31e77d7741f1\" (UID: \"18eb3d3c-9197-4730-b085-31e77d7741f1\") " Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.972213 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18eb3d3c-9197-4730-b085-31e77d7741f1-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "18eb3d3c-9197-4730-b085-31e77d7741f1" (UID: "18eb3d3c-9197-4730-b085-31e77d7741f1"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.973184 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18eb3d3c-9197-4730-b085-31e77d7741f1-kube-api-access-d9ljq" (OuterVolumeSpecName: "kube-api-access-d9ljq") pod "18eb3d3c-9197-4730-b085-31e77d7741f1" (UID: "18eb3d3c-9197-4730-b085-31e77d7741f1"). InnerVolumeSpecName "kube-api-access-d9ljq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.976708 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18eb3d3c-9197-4730-b085-31e77d7741f1-config-data" (OuterVolumeSpecName: "config-data") pod "18eb3d3c-9197-4730-b085-31e77d7741f1" (UID: "18eb3d3c-9197-4730-b085-31e77d7741f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:00 crc kubenswrapper[4766]: I1126 01:45:00.999578 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18eb3d3c-9197-4730-b085-31e77d7741f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18eb3d3c-9197-4730-b085-31e77d7741f1" (UID: "18eb3d3c-9197-4730-b085-31e77d7741f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.068301 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18eb3d3c-9197-4730-b085-31e77d7741f1-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.068361 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9ljq\" (UniqueName: \"kubernetes.io/projected/18eb3d3c-9197-4730-b085-31e77d7741f1-kube-api-access-d9ljq\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.068384 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18eb3d3c-9197-4730-b085-31e77d7741f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.068405 4766 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/18eb3d3c-9197-4730-b085-31e77d7741f1-job-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.154217 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402025-vsmwp"] Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.307181 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-dcdnc" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.307207 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-dcdnc" event={"ID":"18eb3d3c-9197-4730-b085-31e77d7741f1","Type":"ContainerDied","Data":"b16f80a8f88190eecd3a32ef1210ffd950921333a40c9f3e69d405bef845358b"} Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.307254 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b16f80a8f88190eecd3a32ef1210ffd950921333a40c9f3e69d405bef845358b" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.309001 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402025-vsmwp" event={"ID":"bb60fd86-c072-40b9-a2e3-047f3b83c5bd","Type":"ContainerStarted","Data":"cc4948058faa6d5844e952845e3016705f456f2072abca13738079c557c27232"} Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.616708 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Nov 26 01:45:01 crc kubenswrapper[4766]: E1126 01:45:01.617377 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18eb3d3c-9197-4730-b085-31e77d7741f1" containerName="manila-db-sync" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.617389 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="18eb3d3c-9197-4730-b085-31e77d7741f1" containerName="manila-db-sync" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.617612 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="18eb3d3c-9197-4730-b085-31e77d7741f1" containerName="manila-db-sync" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.618815 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.641190 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-k9p8p" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.641258 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.641378 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.641490 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.652822 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.675493 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.678083 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.681723 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.681805 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-scripts\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.681829 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7z8h\" (UniqueName: \"kubernetes.io/projected/1118ff67-6cfc-487c-8887-071283f2cc18-kube-api-access-v7z8h\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.681858 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.681900 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1118ff67-6cfc-487c-8887-071283f2cc18-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.681915 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1c921728-52e2-4b04-8202-ac19c632833c-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.681942 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-config-data\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.681967 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1118ff67-6cfc-487c-8887-071283f2cc18-ceph\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.681995 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnq4r\" (UniqueName: \"kubernetes.io/projected/1c921728-52e2-4b04-8202-ac19c632833c-kube-api-access-mnq4r\") pod \"manila-scheduler-0\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.682016 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/1118ff67-6cfc-487c-8887-071283f2cc18-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.682058 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.682076 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-config-data\") pod \"manila-scheduler-0\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.682106 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.682121 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-scripts\") pod \"manila-scheduler-0\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.685064 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.759746 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.784087 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-config-data\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.784145 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1118ff67-6cfc-487c-8887-071283f2cc18-ceph\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.784182 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnq4r\" (UniqueName: \"kubernetes.io/projected/1c921728-52e2-4b04-8202-ac19c632833c-kube-api-access-mnq4r\") pod \"manila-scheduler-0\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.784209 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/1118ff67-6cfc-487c-8887-071283f2cc18-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.784258 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.784280 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-config-data\") pod \"manila-scheduler-0\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.784313 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.784330 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-scripts\") pod \"manila-scheduler-0\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.784353 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.784405 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-scripts\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.784429 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7z8h\" (UniqueName: \"kubernetes.io/projected/1118ff67-6cfc-487c-8887-071283f2cc18-kube-api-access-v7z8h\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.784461 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.784508 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1118ff67-6cfc-487c-8887-071283f2cc18-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.784524 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1c921728-52e2-4b04-8202-ac19c632833c-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.784614 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1c921728-52e2-4b04-8202-ac19c632833c-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.785779 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/1118ff67-6cfc-487c-8887-071283f2cc18-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.790624 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-c8d8d886c-9lgvr"] Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.792883 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.794131 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1118ff67-6cfc-487c-8887-071283f2cc18-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.803881 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.804803 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-scripts\") pod \"manila-scheduler-0\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.804878 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.805241 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1118ff67-6cfc-487c-8887-071283f2cc18-ceph\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.805943 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-config-data\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.808861 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-scripts\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.811664 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c8d8d886c-9lgvr"] Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.814777 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.824078 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.841273 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnq4r\" (UniqueName: \"kubernetes.io/projected/1c921728-52e2-4b04-8202-ac19c632833c-kube-api-access-mnq4r\") pod \"manila-scheduler-0\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.851694 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-config-data\") pod \"manila-scheduler-0\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.860308 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7z8h\" (UniqueName: \"kubernetes.io/projected/1118ff67-6cfc-487c-8887-071283f2cc18-kube-api-access-v7z8h\") pod \"manila-share-share1-0\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.946766 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.948684 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.952575 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.953283 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.978350 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.991213 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/316b23d8-2140-4ac5-8e26-d63171814293-dns-swift-storage-0\") pod \"dnsmasq-dns-c8d8d886c-9lgvr\" (UID: \"316b23d8-2140-4ac5-8e26-d63171814293\") " pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.991307 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/316b23d8-2140-4ac5-8e26-d63171814293-config\") pod \"dnsmasq-dns-c8d8d886c-9lgvr\" (UID: \"316b23d8-2140-4ac5-8e26-d63171814293\") " pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.991415 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/316b23d8-2140-4ac5-8e26-d63171814293-dns-svc\") pod \"dnsmasq-dns-c8d8d886c-9lgvr\" (UID: \"316b23d8-2140-4ac5-8e26-d63171814293\") " pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.991472 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/316b23d8-2140-4ac5-8e26-d63171814293-ovsdbserver-sb\") pod \"dnsmasq-dns-c8d8d886c-9lgvr\" (UID: \"316b23d8-2140-4ac5-8e26-d63171814293\") " pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.991531 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pmwn\" (UniqueName: \"kubernetes.io/projected/316b23d8-2140-4ac5-8e26-d63171814293-kube-api-access-5pmwn\") pod \"dnsmasq-dns-c8d8d886c-9lgvr\" (UID: \"316b23d8-2140-4ac5-8e26-d63171814293\") " pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.991571 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/316b23d8-2140-4ac5-8e26-d63171814293-openstack-edpm-ipam\") pod \"dnsmasq-dns-c8d8d886c-9lgvr\" (UID: \"316b23d8-2140-4ac5-8e26-d63171814293\") " pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:01 crc kubenswrapper[4766]: I1126 01:45:01.991638 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/316b23d8-2140-4ac5-8e26-d63171814293-ovsdbserver-nb\") pod \"dnsmasq-dns-c8d8d886c-9lgvr\" (UID: \"316b23d8-2140-4ac5-8e26-d63171814293\") " pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.027182 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.101202 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-config-data-custom\") pod \"manila-api-0\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.101308 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/316b23d8-2140-4ac5-8e26-d63171814293-openstack-edpm-ipam\") pod \"dnsmasq-dns-c8d8d886c-9lgvr\" (UID: \"316b23d8-2140-4ac5-8e26-d63171814293\") " pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.101390 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.101446 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/316b23d8-2140-4ac5-8e26-d63171814293-ovsdbserver-nb\") pod \"dnsmasq-dns-c8d8d886c-9lgvr\" (UID: \"316b23d8-2140-4ac5-8e26-d63171814293\") " pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.101539 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-config-data\") pod \"manila-api-0\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.101583 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-logs\") pod \"manila-api-0\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.101642 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/316b23d8-2140-4ac5-8e26-d63171814293-dns-swift-storage-0\") pod \"dnsmasq-dns-c8d8d886c-9lgvr\" (UID: \"316b23d8-2140-4ac5-8e26-d63171814293\") " pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.101735 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-scripts\") pod \"manila-api-0\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.101763 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/316b23d8-2140-4ac5-8e26-d63171814293-config\") pod \"dnsmasq-dns-c8d8d886c-9lgvr\" (UID: \"316b23d8-2140-4ac5-8e26-d63171814293\") " pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.101878 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-etc-machine-id\") pod \"manila-api-0\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.101951 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/316b23d8-2140-4ac5-8e26-d63171814293-dns-svc\") pod \"dnsmasq-dns-c8d8d886c-9lgvr\" (UID: \"316b23d8-2140-4ac5-8e26-d63171814293\") " pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.102014 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f7hv\" (UniqueName: \"kubernetes.io/projected/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-kube-api-access-4f7hv\") pod \"manila-api-0\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.102056 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/316b23d8-2140-4ac5-8e26-d63171814293-ovsdbserver-sb\") pod \"dnsmasq-dns-c8d8d886c-9lgvr\" (UID: \"316b23d8-2140-4ac5-8e26-d63171814293\") " pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.102118 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pmwn\" (UniqueName: \"kubernetes.io/projected/316b23d8-2140-4ac5-8e26-d63171814293-kube-api-access-5pmwn\") pod \"dnsmasq-dns-c8d8d886c-9lgvr\" (UID: \"316b23d8-2140-4ac5-8e26-d63171814293\") " pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.103379 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/316b23d8-2140-4ac5-8e26-d63171814293-openstack-edpm-ipam\") pod \"dnsmasq-dns-c8d8d886c-9lgvr\" (UID: \"316b23d8-2140-4ac5-8e26-d63171814293\") " pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.103898 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/316b23d8-2140-4ac5-8e26-d63171814293-ovsdbserver-nb\") pod \"dnsmasq-dns-c8d8d886c-9lgvr\" (UID: \"316b23d8-2140-4ac5-8e26-d63171814293\") " pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.104415 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/316b23d8-2140-4ac5-8e26-d63171814293-dns-swift-storage-0\") pod \"dnsmasq-dns-c8d8d886c-9lgvr\" (UID: \"316b23d8-2140-4ac5-8e26-d63171814293\") " pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.105441 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/316b23d8-2140-4ac5-8e26-d63171814293-config\") pod \"dnsmasq-dns-c8d8d886c-9lgvr\" (UID: \"316b23d8-2140-4ac5-8e26-d63171814293\") " pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.105986 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/316b23d8-2140-4ac5-8e26-d63171814293-dns-svc\") pod \"dnsmasq-dns-c8d8d886c-9lgvr\" (UID: \"316b23d8-2140-4ac5-8e26-d63171814293\") " pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.106458 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/316b23d8-2140-4ac5-8e26-d63171814293-ovsdbserver-sb\") pod \"dnsmasq-dns-c8d8d886c-9lgvr\" (UID: \"316b23d8-2140-4ac5-8e26-d63171814293\") " pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.145000 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pmwn\" (UniqueName: \"kubernetes.io/projected/316b23d8-2140-4ac5-8e26-d63171814293-kube-api-access-5pmwn\") pod \"dnsmasq-dns-c8d8d886c-9lgvr\" (UID: \"316b23d8-2140-4ac5-8e26-d63171814293\") " pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.215172 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-config-data\") pod \"manila-api-0\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.215226 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-logs\") pod \"manila-api-0\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.215267 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-scripts\") pod \"manila-api-0\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.215317 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-etc-machine-id\") pod \"manila-api-0\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.215365 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4f7hv\" (UniqueName: \"kubernetes.io/projected/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-kube-api-access-4f7hv\") pod \"manila-api-0\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.215415 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-config-data-custom\") pod \"manila-api-0\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.215458 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.218922 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-etc-machine-id\") pod \"manila-api-0\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.219715 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-logs\") pod \"manila-api-0\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.223779 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-scripts\") pod \"manila-api-0\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.225856 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.227740 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-config-data-custom\") pod \"manila-api-0\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.229094 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-config-data\") pod \"manila-api-0\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.247581 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4f7hv\" (UniqueName: \"kubernetes.io/projected/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-kube-api-access-4f7hv\") pod \"manila-api-0\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.274686 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.282971 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.332431 4766 generic.go:334] "Generic (PLEG): container finished" podID="bb60fd86-c072-40b9-a2e3-047f3b83c5bd" containerID="eef8ee593ead07a9e091d4276577af4eca5229e4b3eab352eb6c8e78f8cbd6ce" exitCode=0 Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.332482 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402025-vsmwp" event={"ID":"bb60fd86-c072-40b9-a2e3-047f3b83c5bd","Type":"ContainerDied","Data":"eef8ee593ead07a9e091d4276577af4eca5229e4b3eab352eb6c8e78f8cbd6ce"} Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.587670 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 26 01:45:02 crc kubenswrapper[4766]: I1126 01:45:02.714584 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 26 01:45:03 crc kubenswrapper[4766]: I1126 01:45:03.045430 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 26 01:45:03 crc kubenswrapper[4766]: I1126 01:45:03.132642 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c8d8d886c-9lgvr"] Nov 26 01:45:03 crc kubenswrapper[4766]: I1126 01:45:03.350369 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"1c921728-52e2-4b04-8202-ac19c632833c","Type":"ContainerStarted","Data":"5fb34dd9c6155825e7d3e6c9c54a74e165886ad0d4c964fce3d80c53f7251bdb"} Nov 26 01:45:03 crc kubenswrapper[4766]: I1126 01:45:03.351613 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94","Type":"ContainerStarted","Data":"eb3014f04525d7ea6c4d725841e8909bd4deceda6bef0ec6dac0991c0c9f73c2"} Nov 26 01:45:03 crc kubenswrapper[4766]: I1126 01:45:03.381466 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1118ff67-6cfc-487c-8887-071283f2cc18","Type":"ContainerStarted","Data":"fcb98840cac5e4bdee2fa0111c4aa99dd4e411325a8c8829f0042f0cc4cc32ca"} Nov 26 01:45:03 crc kubenswrapper[4766]: I1126 01:45:03.458095 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" event={"ID":"316b23d8-2140-4ac5-8e26-d63171814293","Type":"ContainerStarted","Data":"d805aae5c1486cecb3800e5fae39c02ef5f962c78ae3095aa9016c68a9b414b2"} Nov 26 01:45:04 crc kubenswrapper[4766]: I1126 01:45:04.472274 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94","Type":"ContainerStarted","Data":"9f625e8e71d8c4f9ec9bc354e8b9da271a4a408d8829f968c86a626aa171d3b0"} Nov 26 01:45:04 crc kubenswrapper[4766]: I1126 01:45:04.473950 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402025-vsmwp" event={"ID":"bb60fd86-c072-40b9-a2e3-047f3b83c5bd","Type":"ContainerDied","Data":"cc4948058faa6d5844e952845e3016705f456f2072abca13738079c557c27232"} Nov 26 01:45:04 crc kubenswrapper[4766]: I1126 01:45:04.473975 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc4948058faa6d5844e952845e3016705f456f2072abca13738079c557c27232" Nov 26 01:45:04 crc kubenswrapper[4766]: I1126 01:45:04.825157 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402025-vsmwp" Nov 26 01:45:04 crc kubenswrapper[4766]: I1126 01:45:04.917323 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb60fd86-c072-40b9-a2e3-047f3b83c5bd-secret-volume\") pod \"bb60fd86-c072-40b9-a2e3-047f3b83c5bd\" (UID: \"bb60fd86-c072-40b9-a2e3-047f3b83c5bd\") " Nov 26 01:45:04 crc kubenswrapper[4766]: I1126 01:45:04.917597 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kw647\" (UniqueName: \"kubernetes.io/projected/bb60fd86-c072-40b9-a2e3-047f3b83c5bd-kube-api-access-kw647\") pod \"bb60fd86-c072-40b9-a2e3-047f3b83c5bd\" (UID: \"bb60fd86-c072-40b9-a2e3-047f3b83c5bd\") " Nov 26 01:45:04 crc kubenswrapper[4766]: I1126 01:45:04.917741 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb60fd86-c072-40b9-a2e3-047f3b83c5bd-config-volume\") pod \"bb60fd86-c072-40b9-a2e3-047f3b83c5bd\" (UID: \"bb60fd86-c072-40b9-a2e3-047f3b83c5bd\") " Nov 26 01:45:04 crc kubenswrapper[4766]: I1126 01:45:04.918710 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb60fd86-c072-40b9-a2e3-047f3b83c5bd-config-volume" (OuterVolumeSpecName: "config-volume") pod "bb60fd86-c072-40b9-a2e3-047f3b83c5bd" (UID: "bb60fd86-c072-40b9-a2e3-047f3b83c5bd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:45:04 crc kubenswrapper[4766]: I1126 01:45:04.921067 4766 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb60fd86-c072-40b9-a2e3-047f3b83c5bd-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:04 crc kubenswrapper[4766]: I1126 01:45:04.924179 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb60fd86-c072-40b9-a2e3-047f3b83c5bd-kube-api-access-kw647" (OuterVolumeSpecName: "kube-api-access-kw647") pod "bb60fd86-c072-40b9-a2e3-047f3b83c5bd" (UID: "bb60fd86-c072-40b9-a2e3-047f3b83c5bd"). InnerVolumeSpecName "kube-api-access-kw647". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:45:04 crc kubenswrapper[4766]: I1126 01:45:04.924538 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb60fd86-c072-40b9-a2e3-047f3b83c5bd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bb60fd86-c072-40b9-a2e3-047f3b83c5bd" (UID: "bb60fd86-c072-40b9-a2e3-047f3b83c5bd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:05 crc kubenswrapper[4766]: I1126 01:45:05.024713 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kw647\" (UniqueName: \"kubernetes.io/projected/bb60fd86-c072-40b9-a2e3-047f3b83c5bd-kube-api-access-kw647\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:05 crc kubenswrapper[4766]: I1126 01:45:05.025063 4766 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb60fd86-c072-40b9-a2e3-047f3b83c5bd-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:05 crc kubenswrapper[4766]: I1126 01:45:05.378362 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Nov 26 01:45:05 crc kubenswrapper[4766]: I1126 01:45:05.486351 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"1c921728-52e2-4b04-8202-ac19c632833c","Type":"ContainerStarted","Data":"6fe68e9a872e3c7c16975868bb45f96814c70638cbfce6ec0e77ad8f83406632"} Nov 26 01:45:05 crc kubenswrapper[4766]: I1126 01:45:05.486421 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"1c921728-52e2-4b04-8202-ac19c632833c","Type":"ContainerStarted","Data":"d7357d8b357a23f6567bce95ef9894289ea485a98616ce1f81463ec94202a8b7"} Nov 26 01:45:05 crc kubenswrapper[4766]: I1126 01:45:05.507598 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.3539921059999998 podStartE2EDuration="4.507581244s" podCreationTimestamp="2025-11-26 01:45:01 +0000 UTC" firstStartedPulling="2025-11-26 01:45:02.609957644 +0000 UTC m=+4883.458728074" lastFinishedPulling="2025-11-26 01:45:03.763546782 +0000 UTC m=+4884.612317212" observedRunningTime="2025-11-26 01:45:05.503067714 +0000 UTC m=+4886.351838144" watchObservedRunningTime="2025-11-26 01:45:05.507581244 +0000 UTC m=+4886.356351674" Nov 26 01:45:05 crc kubenswrapper[4766]: I1126 01:45:05.522110 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94","Type":"ContainerStarted","Data":"c434f422814daa203ea5aa2ea47c9440464b5825ef47696da96e513186ea72a9"} Nov 26 01:45:05 crc kubenswrapper[4766]: I1126 01:45:05.522222 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Nov 26 01:45:05 crc kubenswrapper[4766]: I1126 01:45:05.531539 4766 generic.go:334] "Generic (PLEG): container finished" podID="316b23d8-2140-4ac5-8e26-d63171814293" containerID="b4c35713a0cb6561769467a0160726fb0ae9765ee8ab59c6888469d2b0c760e8" exitCode=0 Nov 26 01:45:05 crc kubenswrapper[4766]: I1126 01:45:05.531619 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402025-vsmwp" Nov 26 01:45:05 crc kubenswrapper[4766]: I1126 01:45:05.536839 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" event={"ID":"316b23d8-2140-4ac5-8e26-d63171814293","Type":"ContainerDied","Data":"b4c35713a0cb6561769467a0160726fb0ae9765ee8ab59c6888469d2b0c760e8"} Nov 26 01:45:05 crc kubenswrapper[4766]: I1126 01:45:05.559283 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=4.559260072 podStartE2EDuration="4.559260072s" podCreationTimestamp="2025-11-26 01:45:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 01:45:05.539020546 +0000 UTC m=+4886.387790976" watchObservedRunningTime="2025-11-26 01:45:05.559260072 +0000 UTC m=+4886.408030502" Nov 26 01:45:05 crc kubenswrapper[4766]: I1126 01:45:05.908667 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w"] Nov 26 01:45:05 crc kubenswrapper[4766]: I1126 01:45:05.919521 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401980-n2x8w"] Nov 26 01:45:06 crc kubenswrapper[4766]: I1126 01:45:06.548149 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" event={"ID":"316b23d8-2140-4ac5-8e26-d63171814293","Type":"ContainerStarted","Data":"e81e568fe8bcbbd913dd7782989da25911570f059d18986bbaa19691c3c873ca"} Nov 26 01:45:06 crc kubenswrapper[4766]: I1126 01:45:06.548503 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94" containerName="manila-api-log" containerID="cri-o://9f625e8e71d8c4f9ec9bc354e8b9da271a4a408d8829f968c86a626aa171d3b0" gracePeriod=30 Nov 26 01:45:06 crc kubenswrapper[4766]: I1126 01:45:06.548584 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:06 crc kubenswrapper[4766]: I1126 01:45:06.548619 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94" containerName="manila-api" containerID="cri-o://c434f422814daa203ea5aa2ea47c9440464b5825ef47696da96e513186ea72a9" gracePeriod=30 Nov 26 01:45:06 crc kubenswrapper[4766]: I1126 01:45:06.576793 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" podStartSLOduration=5.576772072 podStartE2EDuration="5.576772072s" podCreationTimestamp="2025-11-26 01:45:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 01:45:06.565989987 +0000 UTC m=+4887.414760427" watchObservedRunningTime="2025-11-26 01:45:06.576772072 +0000 UTC m=+4887.425542492" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.328202 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.385799 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-config-data\") pod \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.385870 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-logs\") pod \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.386048 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-config-data-custom\") pod \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.386092 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-combined-ca-bundle\") pod \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.386267 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-etc-machine-id\") pod \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.386336 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4f7hv\" (UniqueName: \"kubernetes.io/projected/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-kube-api-access-4f7hv\") pod \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.386360 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-logs" (OuterVolumeSpecName: "logs") pod "7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94" (UID: "7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.386386 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-scripts\") pod \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\" (UID: \"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94\") " Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.386417 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94" (UID: "7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.387017 4766 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.387035 4766 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-logs\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.394402 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-kube-api-access-4f7hv" (OuterVolumeSpecName: "kube-api-access-4f7hv") pod "7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94" (UID: "7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94"). InnerVolumeSpecName "kube-api-access-4f7hv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.401544 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-scripts" (OuterVolumeSpecName: "scripts") pod "7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94" (UID: "7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.410930 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94" (UID: "7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.438427 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94" (UID: "7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.486870 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-config-data" (OuterVolumeSpecName: "config-data") pod "7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94" (UID: "7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.488493 4766 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.488520 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.488531 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4f7hv\" (UniqueName: \"kubernetes.io/projected/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-kube-api-access-4f7hv\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.488541 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.488551 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.556973 4766 generic.go:334] "Generic (PLEG): container finished" podID="7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94" containerID="c434f422814daa203ea5aa2ea47c9440464b5825ef47696da96e513186ea72a9" exitCode=0 Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.557007 4766 generic.go:334] "Generic (PLEG): container finished" podID="7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94" containerID="9f625e8e71d8c4f9ec9bc354e8b9da271a4a408d8829f968c86a626aa171d3b0" exitCode=143 Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.557071 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94","Type":"ContainerDied","Data":"c434f422814daa203ea5aa2ea47c9440464b5825ef47696da96e513186ea72a9"} Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.557114 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94","Type":"ContainerDied","Data":"9f625e8e71d8c4f9ec9bc354e8b9da271a4a408d8829f968c86a626aa171d3b0"} Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.557126 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94","Type":"ContainerDied","Data":"eb3014f04525d7ea6c4d725841e8909bd4deceda6bef0ec6dac0991c0c9f73c2"} Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.557143 4766 scope.go:117] "RemoveContainer" containerID="c434f422814daa203ea5aa2ea47c9440464b5825ef47696da96e513186ea72a9" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.557319 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.579128 4766 scope.go:117] "RemoveContainer" containerID="9f625e8e71d8c4f9ec9bc354e8b9da271a4a408d8829f968c86a626aa171d3b0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.625453 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.639801 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.644002 4766 scope.go:117] "RemoveContainer" containerID="c434f422814daa203ea5aa2ea47c9440464b5825ef47696da96e513186ea72a9" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.653138 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Nov 26 01:45:07 crc kubenswrapper[4766]: E1126 01:45:07.653596 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb60fd86-c072-40b9-a2e3-047f3b83c5bd" containerName="collect-profiles" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.653613 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb60fd86-c072-40b9-a2e3-047f3b83c5bd" containerName="collect-profiles" Nov 26 01:45:07 crc kubenswrapper[4766]: E1126 01:45:07.653630 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94" containerName="manila-api-log" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.653637 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94" containerName="manila-api-log" Nov 26 01:45:07 crc kubenswrapper[4766]: E1126 01:45:07.653668 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94" containerName="manila-api" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.653674 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94" containerName="manila-api" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.653895 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb60fd86-c072-40b9-a2e3-047f3b83c5bd" containerName="collect-profiles" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.653908 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94" containerName="manila-api" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.653928 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94" containerName="manila-api-log" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.655062 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: E1126 01:45:07.659834 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c434f422814daa203ea5aa2ea47c9440464b5825ef47696da96e513186ea72a9\": container with ID starting with c434f422814daa203ea5aa2ea47c9440464b5825ef47696da96e513186ea72a9 not found: ID does not exist" containerID="c434f422814daa203ea5aa2ea47c9440464b5825ef47696da96e513186ea72a9" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.659870 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c434f422814daa203ea5aa2ea47c9440464b5825ef47696da96e513186ea72a9"} err="failed to get container status \"c434f422814daa203ea5aa2ea47c9440464b5825ef47696da96e513186ea72a9\": rpc error: code = NotFound desc = could not find container \"c434f422814daa203ea5aa2ea47c9440464b5825ef47696da96e513186ea72a9\": container with ID starting with c434f422814daa203ea5aa2ea47c9440464b5825ef47696da96e513186ea72a9 not found: ID does not exist" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.659892 4766 scope.go:117] "RemoveContainer" containerID="9f625e8e71d8c4f9ec9bc354e8b9da271a4a408d8829f968c86a626aa171d3b0" Nov 26 01:45:07 crc kubenswrapper[4766]: E1126 01:45:07.660245 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f625e8e71d8c4f9ec9bc354e8b9da271a4a408d8829f968c86a626aa171d3b0\": container with ID starting with 9f625e8e71d8c4f9ec9bc354e8b9da271a4a408d8829f968c86a626aa171d3b0 not found: ID does not exist" containerID="9f625e8e71d8c4f9ec9bc354e8b9da271a4a408d8829f968c86a626aa171d3b0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.660266 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f625e8e71d8c4f9ec9bc354e8b9da271a4a408d8829f968c86a626aa171d3b0"} err="failed to get container status \"9f625e8e71d8c4f9ec9bc354e8b9da271a4a408d8829f968c86a626aa171d3b0\": rpc error: code = NotFound desc = could not find container \"9f625e8e71d8c4f9ec9bc354e8b9da271a4a408d8829f968c86a626aa171d3b0\": container with ID starting with 9f625e8e71d8c4f9ec9bc354e8b9da271a4a408d8829f968c86a626aa171d3b0 not found: ID does not exist" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.660280 4766 scope.go:117] "RemoveContainer" containerID="c434f422814daa203ea5aa2ea47c9440464b5825ef47696da96e513186ea72a9" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.662070 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.662845 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c434f422814daa203ea5aa2ea47c9440464b5825ef47696da96e513186ea72a9"} err="failed to get container status \"c434f422814daa203ea5aa2ea47c9440464b5825ef47696da96e513186ea72a9\": rpc error: code = NotFound desc = could not find container \"c434f422814daa203ea5aa2ea47c9440464b5825ef47696da96e513186ea72a9\": container with ID starting with c434f422814daa203ea5aa2ea47c9440464b5825ef47696da96e513186ea72a9 not found: ID does not exist" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.662867 4766 scope.go:117] "RemoveContainer" containerID="9f625e8e71d8c4f9ec9bc354e8b9da271a4a408d8829f968c86a626aa171d3b0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.663280 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.663379 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f625e8e71d8c4f9ec9bc354e8b9da271a4a408d8829f968c86a626aa171d3b0"} err="failed to get container status \"9f625e8e71d8c4f9ec9bc354e8b9da271a4a408d8829f968c86a626aa171d3b0\": rpc error: code = NotFound desc = could not find container \"9f625e8e71d8c4f9ec9bc354e8b9da271a4a408d8829f968c86a626aa171d3b0\": container with ID starting with 9f625e8e71d8c4f9ec9bc354e8b9da271a4a408d8829f968c86a626aa171d3b0 not found: ID does not exist" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.663451 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.664733 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.679212 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.679468 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a22c5084-e53e-4529-ba7f-31468f99fe79" containerName="ceilometer-central-agent" containerID="cri-o://98755bd910987eb515252f2f2e9abe4a387ffd403a0971db75e38ef549222632" gracePeriod=30 Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.680279 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a22c5084-e53e-4529-ba7f-31468f99fe79" containerName="ceilometer-notification-agent" containerID="cri-o://13db0c5081c86540d776fae42cf51d65f73dfc346caeeb8e9af07a15742c3a49" gracePeriod=30 Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.680293 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a22c5084-e53e-4529-ba7f-31468f99fe79" containerName="proxy-httpd" containerID="cri-o://c35f2a49b9d1b4c98a9e3c26b6539ac9cb5060fa828c620187f1bfd114411294" gracePeriod=30 Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.680263 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a22c5084-e53e-4529-ba7f-31468f99fe79" containerName="sg-core" containerID="cri-o://3fd7408fd83a3c1fe26d8c321a6fb0dcf192d0be1a723c7318ac2243a5451575" gracePeriod=30 Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.693032 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cb6838c-a086-4990-82f0-a3c36675be67-internal-tls-certs\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.693104 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8cb6838c-a086-4990-82f0-a3c36675be67-etc-machine-id\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.693132 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x85t7\" (UniqueName: \"kubernetes.io/projected/8cb6838c-a086-4990-82f0-a3c36675be67-kube-api-access-x85t7\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.693249 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cb6838c-a086-4990-82f0-a3c36675be67-config-data\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.693398 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8cb6838c-a086-4990-82f0-a3c36675be67-config-data-custom\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.693520 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cb6838c-a086-4990-82f0-a3c36675be67-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.696197 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cb6838c-a086-4990-82f0-a3c36675be67-public-tls-certs\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.696279 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cb6838c-a086-4990-82f0-a3c36675be67-scripts\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.696501 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cb6838c-a086-4990-82f0-a3c36675be67-logs\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.798360 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8cb6838c-a086-4990-82f0-a3c36675be67-config-data-custom\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.798441 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cb6838c-a086-4990-82f0-a3c36675be67-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.798484 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cb6838c-a086-4990-82f0-a3c36675be67-public-tls-certs\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.798563 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cb6838c-a086-4990-82f0-a3c36675be67-scripts\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.798593 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cb6838c-a086-4990-82f0-a3c36675be67-logs\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.799307 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cb6838c-a086-4990-82f0-a3c36675be67-logs\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.799346 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cb6838c-a086-4990-82f0-a3c36675be67-internal-tls-certs\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.799388 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8cb6838c-a086-4990-82f0-a3c36675be67-etc-machine-id\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.799418 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x85t7\" (UniqueName: \"kubernetes.io/projected/8cb6838c-a086-4990-82f0-a3c36675be67-kube-api-access-x85t7\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.799445 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cb6838c-a086-4990-82f0-a3c36675be67-config-data\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.799644 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8cb6838c-a086-4990-82f0-a3c36675be67-etc-machine-id\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.803397 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cb6838c-a086-4990-82f0-a3c36675be67-public-tls-certs\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.803563 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cb6838c-a086-4990-82f0-a3c36675be67-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.804689 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cb6838c-a086-4990-82f0-a3c36675be67-config-data\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.805139 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cb6838c-a086-4990-82f0-a3c36675be67-scripts\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.806495 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cb6838c-a086-4990-82f0-a3c36675be67-internal-tls-certs\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.807231 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8cb6838c-a086-4990-82f0-a3c36675be67-config-data-custom\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.821133 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x85t7\" (UniqueName: \"kubernetes.io/projected/8cb6838c-a086-4990-82f0-a3c36675be67-kube-api-access-x85t7\") pod \"manila-api-0\" (UID: \"8cb6838c-a086-4990-82f0-a3c36675be67\") " pod="openstack/manila-api-0" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.843560 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94" path="/var/lib/kubelet/pods/7f7ceb46-1493-4adb-a0bc-f3bfbd89eb94/volumes" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.844524 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a13a895b-28a8-46d1-895d-f9f6cce17490" path="/var/lib/kubelet/pods/a13a895b-28a8-46d1-895d-f9f6cce17490/volumes" Nov 26 01:45:07 crc kubenswrapper[4766]: I1126 01:45:07.980927 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 26 01:45:08 crc kubenswrapper[4766]: I1126 01:45:08.574203 4766 generic.go:334] "Generic (PLEG): container finished" podID="a22c5084-e53e-4529-ba7f-31468f99fe79" containerID="c35f2a49b9d1b4c98a9e3c26b6539ac9cb5060fa828c620187f1bfd114411294" exitCode=0 Nov 26 01:45:08 crc kubenswrapper[4766]: I1126 01:45:08.574526 4766 generic.go:334] "Generic (PLEG): container finished" podID="a22c5084-e53e-4529-ba7f-31468f99fe79" containerID="3fd7408fd83a3c1fe26d8c321a6fb0dcf192d0be1a723c7318ac2243a5451575" exitCode=2 Nov 26 01:45:08 crc kubenswrapper[4766]: I1126 01:45:08.574535 4766 generic.go:334] "Generic (PLEG): container finished" podID="a22c5084-e53e-4529-ba7f-31468f99fe79" containerID="98755bd910987eb515252f2f2e9abe4a387ffd403a0971db75e38ef549222632" exitCode=0 Nov 26 01:45:08 crc kubenswrapper[4766]: I1126 01:45:08.574623 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a22c5084-e53e-4529-ba7f-31468f99fe79","Type":"ContainerDied","Data":"c35f2a49b9d1b4c98a9e3c26b6539ac9cb5060fa828c620187f1bfd114411294"} Nov 26 01:45:08 crc kubenswrapper[4766]: I1126 01:45:08.574665 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a22c5084-e53e-4529-ba7f-31468f99fe79","Type":"ContainerDied","Data":"3fd7408fd83a3c1fe26d8c321a6fb0dcf192d0be1a723c7318ac2243a5451575"} Nov 26 01:45:08 crc kubenswrapper[4766]: I1126 01:45:08.574678 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a22c5084-e53e-4529-ba7f-31468f99fe79","Type":"ContainerDied","Data":"98755bd910987eb515252f2f2e9abe4a387ffd403a0971db75e38ef549222632"} Nov 26 01:45:08 crc kubenswrapper[4766]: I1126 01:45:08.794496 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.239930 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.351468 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-combined-ca-bundle\") pod \"a22c5084-e53e-4529-ba7f-31468f99fe79\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.351829 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-ceilometer-tls-certs\") pod \"a22c5084-e53e-4529-ba7f-31468f99fe79\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.351860 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a22c5084-e53e-4529-ba7f-31468f99fe79-run-httpd\") pod \"a22c5084-e53e-4529-ba7f-31468f99fe79\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.351951 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-config-data\") pod \"a22c5084-e53e-4529-ba7f-31468f99fe79\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.352011 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9rfm\" (UniqueName: \"kubernetes.io/projected/a22c5084-e53e-4529-ba7f-31468f99fe79-kube-api-access-g9rfm\") pod \"a22c5084-e53e-4529-ba7f-31468f99fe79\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.352054 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-scripts\") pod \"a22c5084-e53e-4529-ba7f-31468f99fe79\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.352101 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a22c5084-e53e-4529-ba7f-31468f99fe79-log-httpd\") pod \"a22c5084-e53e-4529-ba7f-31468f99fe79\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.352185 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-sg-core-conf-yaml\") pod \"a22c5084-e53e-4529-ba7f-31468f99fe79\" (UID: \"a22c5084-e53e-4529-ba7f-31468f99fe79\") " Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.352625 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a22c5084-e53e-4529-ba7f-31468f99fe79-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a22c5084-e53e-4529-ba7f-31468f99fe79" (UID: "a22c5084-e53e-4529-ba7f-31468f99fe79"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.352970 4766 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a22c5084-e53e-4529-ba7f-31468f99fe79-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.353929 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a22c5084-e53e-4529-ba7f-31468f99fe79-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a22c5084-e53e-4529-ba7f-31468f99fe79" (UID: "a22c5084-e53e-4529-ba7f-31468f99fe79"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.360807 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a22c5084-e53e-4529-ba7f-31468f99fe79-kube-api-access-g9rfm" (OuterVolumeSpecName: "kube-api-access-g9rfm") pod "a22c5084-e53e-4529-ba7f-31468f99fe79" (UID: "a22c5084-e53e-4529-ba7f-31468f99fe79"). InnerVolumeSpecName "kube-api-access-g9rfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.363700 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-scripts" (OuterVolumeSpecName: "scripts") pod "a22c5084-e53e-4529-ba7f-31468f99fe79" (UID: "a22c5084-e53e-4529-ba7f-31468f99fe79"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.446050 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a22c5084-e53e-4529-ba7f-31468f99fe79" (UID: "a22c5084-e53e-4529-ba7f-31468f99fe79"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.455354 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9rfm\" (UniqueName: \"kubernetes.io/projected/a22c5084-e53e-4529-ba7f-31468f99fe79-kube-api-access-g9rfm\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.455381 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.455389 4766 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a22c5084-e53e-4529-ba7f-31468f99fe79-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.455398 4766 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.489082 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "a22c5084-e53e-4529-ba7f-31468f99fe79" (UID: "a22c5084-e53e-4529-ba7f-31468f99fe79"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.532925 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a22c5084-e53e-4529-ba7f-31468f99fe79" (UID: "a22c5084-e53e-4529-ba7f-31468f99fe79"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.558184 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.558211 4766 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.590980 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-config-data" (OuterVolumeSpecName: "config-data") pod "a22c5084-e53e-4529-ba7f-31468f99fe79" (UID: "a22c5084-e53e-4529-ba7f-31468f99fe79"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.598616 4766 generic.go:334] "Generic (PLEG): container finished" podID="a22c5084-e53e-4529-ba7f-31468f99fe79" containerID="13db0c5081c86540d776fae42cf51d65f73dfc346caeeb8e9af07a15742c3a49" exitCode=0 Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.598690 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a22c5084-e53e-4529-ba7f-31468f99fe79","Type":"ContainerDied","Data":"13db0c5081c86540d776fae42cf51d65f73dfc346caeeb8e9af07a15742c3a49"} Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.598725 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.598747 4766 scope.go:117] "RemoveContainer" containerID="c35f2a49b9d1b4c98a9e3c26b6539ac9cb5060fa828c620187f1bfd114411294" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.598735 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a22c5084-e53e-4529-ba7f-31468f99fe79","Type":"ContainerDied","Data":"dbc81db846693815c88a8535746e7cfb37e6bfd385b685d2151dc4dab4d704be"} Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.605406 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"8cb6838c-a086-4990-82f0-a3c36675be67","Type":"ContainerStarted","Data":"8fd02e3210389c0b6ee5a17e7a4dc5096433e00d2bc061af0a607962abe3869b"} Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.605673 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"8cb6838c-a086-4990-82f0-a3c36675be67","Type":"ContainerStarted","Data":"d9240987d2e5218fc3a987a21cbef9f59aa6f492b831c6b5e97ffe8876d1f569"} Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.626674 4766 scope.go:117] "RemoveContainer" containerID="3fd7408fd83a3c1fe26d8c321a6fb0dcf192d0be1a723c7318ac2243a5451575" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.653126 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.653273 4766 scope.go:117] "RemoveContainer" containerID="13db0c5081c86540d776fae42cf51d65f73dfc346caeeb8e9af07a15742c3a49" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.661250 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a22c5084-e53e-4529-ba7f-31468f99fe79-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.675957 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.681439 4766 scope.go:117] "RemoveContainer" containerID="98755bd910987eb515252f2f2e9abe4a387ffd403a0971db75e38ef549222632" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.687737 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 26 01:45:09 crc kubenswrapper[4766]: E1126 01:45:09.688202 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a22c5084-e53e-4529-ba7f-31468f99fe79" containerName="sg-core" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.688216 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="a22c5084-e53e-4529-ba7f-31468f99fe79" containerName="sg-core" Nov 26 01:45:09 crc kubenswrapper[4766]: E1126 01:45:09.688238 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a22c5084-e53e-4529-ba7f-31468f99fe79" containerName="proxy-httpd" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.688244 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="a22c5084-e53e-4529-ba7f-31468f99fe79" containerName="proxy-httpd" Nov 26 01:45:09 crc kubenswrapper[4766]: E1126 01:45:09.688277 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a22c5084-e53e-4529-ba7f-31468f99fe79" containerName="ceilometer-notification-agent" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.688283 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="a22c5084-e53e-4529-ba7f-31468f99fe79" containerName="ceilometer-notification-agent" Nov 26 01:45:09 crc kubenswrapper[4766]: E1126 01:45:09.688300 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a22c5084-e53e-4529-ba7f-31468f99fe79" containerName="ceilometer-central-agent" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.688306 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="a22c5084-e53e-4529-ba7f-31468f99fe79" containerName="ceilometer-central-agent" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.688490 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="a22c5084-e53e-4529-ba7f-31468f99fe79" containerName="sg-core" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.688507 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="a22c5084-e53e-4529-ba7f-31468f99fe79" containerName="ceilometer-central-agent" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.688528 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="a22c5084-e53e-4529-ba7f-31468f99fe79" containerName="proxy-httpd" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.688540 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="a22c5084-e53e-4529-ba7f-31468f99fe79" containerName="ceilometer-notification-agent" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.690577 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.694329 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.694750 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.700007 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.702490 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.749292 4766 scope.go:117] "RemoveContainer" containerID="c35f2a49b9d1b4c98a9e3c26b6539ac9cb5060fa828c620187f1bfd114411294" Nov 26 01:45:09 crc kubenswrapper[4766]: E1126 01:45:09.750184 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c35f2a49b9d1b4c98a9e3c26b6539ac9cb5060fa828c620187f1bfd114411294\": container with ID starting with c35f2a49b9d1b4c98a9e3c26b6539ac9cb5060fa828c620187f1bfd114411294 not found: ID does not exist" containerID="c35f2a49b9d1b4c98a9e3c26b6539ac9cb5060fa828c620187f1bfd114411294" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.750249 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c35f2a49b9d1b4c98a9e3c26b6539ac9cb5060fa828c620187f1bfd114411294"} err="failed to get container status \"c35f2a49b9d1b4c98a9e3c26b6539ac9cb5060fa828c620187f1bfd114411294\": rpc error: code = NotFound desc = could not find container \"c35f2a49b9d1b4c98a9e3c26b6539ac9cb5060fa828c620187f1bfd114411294\": container with ID starting with c35f2a49b9d1b4c98a9e3c26b6539ac9cb5060fa828c620187f1bfd114411294 not found: ID does not exist" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.750295 4766 scope.go:117] "RemoveContainer" containerID="3fd7408fd83a3c1fe26d8c321a6fb0dcf192d0be1a723c7318ac2243a5451575" Nov 26 01:45:09 crc kubenswrapper[4766]: E1126 01:45:09.750815 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fd7408fd83a3c1fe26d8c321a6fb0dcf192d0be1a723c7318ac2243a5451575\": container with ID starting with 3fd7408fd83a3c1fe26d8c321a6fb0dcf192d0be1a723c7318ac2243a5451575 not found: ID does not exist" containerID="3fd7408fd83a3c1fe26d8c321a6fb0dcf192d0be1a723c7318ac2243a5451575" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.750839 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fd7408fd83a3c1fe26d8c321a6fb0dcf192d0be1a723c7318ac2243a5451575"} err="failed to get container status \"3fd7408fd83a3c1fe26d8c321a6fb0dcf192d0be1a723c7318ac2243a5451575\": rpc error: code = NotFound desc = could not find container \"3fd7408fd83a3c1fe26d8c321a6fb0dcf192d0be1a723c7318ac2243a5451575\": container with ID starting with 3fd7408fd83a3c1fe26d8c321a6fb0dcf192d0be1a723c7318ac2243a5451575 not found: ID does not exist" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.750854 4766 scope.go:117] "RemoveContainer" containerID="13db0c5081c86540d776fae42cf51d65f73dfc346caeeb8e9af07a15742c3a49" Nov 26 01:45:09 crc kubenswrapper[4766]: E1126 01:45:09.751068 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13db0c5081c86540d776fae42cf51d65f73dfc346caeeb8e9af07a15742c3a49\": container with ID starting with 13db0c5081c86540d776fae42cf51d65f73dfc346caeeb8e9af07a15742c3a49 not found: ID does not exist" containerID="13db0c5081c86540d776fae42cf51d65f73dfc346caeeb8e9af07a15742c3a49" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.751087 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13db0c5081c86540d776fae42cf51d65f73dfc346caeeb8e9af07a15742c3a49"} err="failed to get container status \"13db0c5081c86540d776fae42cf51d65f73dfc346caeeb8e9af07a15742c3a49\": rpc error: code = NotFound desc = could not find container \"13db0c5081c86540d776fae42cf51d65f73dfc346caeeb8e9af07a15742c3a49\": container with ID starting with 13db0c5081c86540d776fae42cf51d65f73dfc346caeeb8e9af07a15742c3a49 not found: ID does not exist" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.751099 4766 scope.go:117] "RemoveContainer" containerID="98755bd910987eb515252f2f2e9abe4a387ffd403a0971db75e38ef549222632" Nov 26 01:45:09 crc kubenswrapper[4766]: E1126 01:45:09.751841 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98755bd910987eb515252f2f2e9abe4a387ffd403a0971db75e38ef549222632\": container with ID starting with 98755bd910987eb515252f2f2e9abe4a387ffd403a0971db75e38ef549222632 not found: ID does not exist" containerID="98755bd910987eb515252f2f2e9abe4a387ffd403a0971db75e38ef549222632" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.751863 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98755bd910987eb515252f2f2e9abe4a387ffd403a0971db75e38ef549222632"} err="failed to get container status \"98755bd910987eb515252f2f2e9abe4a387ffd403a0971db75e38ef549222632\": rpc error: code = NotFound desc = could not find container \"98755bd910987eb515252f2f2e9abe4a387ffd403a0971db75e38ef549222632\": container with ID starting with 98755bd910987eb515252f2f2e9abe4a387ffd403a0971db75e38ef549222632 not found: ID does not exist" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.762590 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c561040f-fa3f-4b22-b85a-cd50fbcd4020-log-httpd\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.762695 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c561040f-fa3f-4b22-b85a-cd50fbcd4020-run-httpd\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.762728 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.762770 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-scripts\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.762839 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.762891 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.762927 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-config-data\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.762972 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcwkh\" (UniqueName: \"kubernetes.io/projected/c561040f-fa3f-4b22-b85a-cd50fbcd4020-kube-api-access-kcwkh\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.872393 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c561040f-fa3f-4b22-b85a-cd50fbcd4020-run-httpd\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.872509 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.872623 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-scripts\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.872714 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.872806 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.872830 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-config-data\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.872929 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcwkh\" (UniqueName: \"kubernetes.io/projected/c561040f-fa3f-4b22-b85a-cd50fbcd4020-kube-api-access-kcwkh\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.873028 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c561040f-fa3f-4b22-b85a-cd50fbcd4020-log-httpd\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.874692 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c561040f-fa3f-4b22-b85a-cd50fbcd4020-run-httpd\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.876026 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c561040f-fa3f-4b22-b85a-cd50fbcd4020-log-httpd\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.878734 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.880210 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.881012 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-config-data\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.883328 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.898465 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-scripts\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.916060 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcwkh\" (UniqueName: \"kubernetes.io/projected/c561040f-fa3f-4b22-b85a-cd50fbcd4020-kube-api-access-kcwkh\") pod \"ceilometer-0\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " pod="openstack/ceilometer-0" Nov 26 01:45:09 crc kubenswrapper[4766]: I1126 01:45:09.944756 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a22c5084-e53e-4529-ba7f-31468f99fe79" path="/var/lib/kubelet/pods/a22c5084-e53e-4529-ba7f-31468f99fe79/volumes" Nov 26 01:45:10 crc kubenswrapper[4766]: I1126 01:45:10.014047 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 01:45:10 crc kubenswrapper[4766]: I1126 01:45:10.238578 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:45:10 crc kubenswrapper[4766]: I1126 01:45:10.253471 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:45:10 crc kubenswrapper[4766]: I1126 01:45:10.622634 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"8cb6838c-a086-4990-82f0-a3c36675be67","Type":"ContainerStarted","Data":"d45ce657a0e66f1260b8a17225784f2eb4183441a001e1a52ee3b9c6d0e12579"} Nov 26 01:45:10 crc kubenswrapper[4766]: I1126 01:45:10.622791 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Nov 26 01:45:10 crc kubenswrapper[4766]: I1126 01:45:10.647763 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.647742986 podStartE2EDuration="3.647742986s" podCreationTimestamp="2025-11-26 01:45:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 01:45:10.643562393 +0000 UTC m=+4891.492332823" watchObservedRunningTime="2025-11-26 01:45:10.647742986 +0000 UTC m=+4891.496513416" Nov 26 01:45:11 crc kubenswrapper[4766]: I1126 01:45:11.635057 4766 generic.go:334] "Generic (PLEG): container finished" podID="656cb04b-f831-4671-a102-d01c6d01714f" containerID="dcfd1bcc023c09f3d1070f702e5f6368c5444f8c30ec70ebff460066679e4f44" exitCode=137 Nov 26 01:45:11 crc kubenswrapper[4766]: I1126 01:45:11.635295 4766 generic.go:334] "Generic (PLEG): container finished" podID="656cb04b-f831-4671-a102-d01c6d01714f" containerID="fe8bea5f9be1857687215d30d22402fc1211456a7511bb72ccd1c779511c43e5" exitCode=137 Nov 26 01:45:11 crc kubenswrapper[4766]: I1126 01:45:11.635349 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f7f64877f-4sskz" event={"ID":"656cb04b-f831-4671-a102-d01c6d01714f","Type":"ContainerDied","Data":"dcfd1bcc023c09f3d1070f702e5f6368c5444f8c30ec70ebff460066679e4f44"} Nov 26 01:45:11 crc kubenswrapper[4766]: I1126 01:45:11.635375 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f7f64877f-4sskz" event={"ID":"656cb04b-f831-4671-a102-d01c6d01714f","Type":"ContainerDied","Data":"fe8bea5f9be1857687215d30d22402fc1211456a7511bb72ccd1c779511c43e5"} Nov 26 01:45:11 crc kubenswrapper[4766]: I1126 01:45:11.638717 4766 generic.go:334] "Generic (PLEG): container finished" podID="c577b8bb-a3c5-405d-8da8-0289e8483ba5" containerID="797e1e95d5a6b9de52f72bb3fef2fafe5c0c40666f3b784bc4621f22407d05f2" exitCode=137 Nov 26 01:45:11 crc kubenswrapper[4766]: I1126 01:45:11.638750 4766 generic.go:334] "Generic (PLEG): container finished" podID="c577b8bb-a3c5-405d-8da8-0289e8483ba5" containerID="1c619971b704f1525550b524ab022265c6c1aaa8fd615b299de74ec0f58aaaae" exitCode=137 Nov 26 01:45:11 crc kubenswrapper[4766]: I1126 01:45:11.638832 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f5b87f5c-z9b2b" event={"ID":"c577b8bb-a3c5-405d-8da8-0289e8483ba5","Type":"ContainerDied","Data":"797e1e95d5a6b9de52f72bb3fef2fafe5c0c40666f3b784bc4621f22407d05f2"} Nov 26 01:45:11 crc kubenswrapper[4766]: I1126 01:45:11.638899 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f5b87f5c-z9b2b" event={"ID":"c577b8bb-a3c5-405d-8da8-0289e8483ba5","Type":"ContainerDied","Data":"1c619971b704f1525550b524ab022265c6c1aaa8fd615b299de74ec0f58aaaae"} Nov 26 01:45:12 crc kubenswrapper[4766]: I1126 01:45:12.028674 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Nov 26 01:45:12 crc kubenswrapper[4766]: I1126 01:45:12.199072 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:45:12 crc kubenswrapper[4766]: I1126 01:45:12.234347 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-cf46f664-8k6s8" Nov 26 01:45:12 crc kubenswrapper[4766]: I1126 01:45:12.280803 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-c8d8d886c-9lgvr" Nov 26 01:45:12 crc kubenswrapper[4766]: I1126 01:45:12.337988 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-65bd4d46f6-blndr"] Nov 26 01:45:12 crc kubenswrapper[4766]: I1126 01:45:12.349386 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9"] Nov 26 01:45:12 crc kubenswrapper[4766]: I1126 01:45:12.349772 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" podUID="f586d76c-85fe-40dc-8158-52a1747703d7" containerName="dnsmasq-dns" containerID="cri-o://38c1d8e99a74fa7e315f5cf2480050f5585758a433f3aec3b382768482a5866c" gracePeriod=10 Nov 26 01:45:12 crc kubenswrapper[4766]: I1126 01:45:12.663643 4766 generic.go:334] "Generic (PLEG): container finished" podID="f586d76c-85fe-40dc-8158-52a1747703d7" containerID="38c1d8e99a74fa7e315f5cf2480050f5585758a433f3aec3b382768482a5866c" exitCode=0 Nov 26 01:45:12 crc kubenswrapper[4766]: I1126 01:45:12.663685 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" event={"ID":"f586d76c-85fe-40dc-8158-52a1747703d7","Type":"ContainerDied","Data":"38c1d8e99a74fa7e315f5cf2480050f5585758a433f3aec3b382768482a5866c"} Nov 26 01:45:12 crc kubenswrapper[4766]: I1126 01:45:12.663906 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-65bd4d46f6-blndr" podUID="9e2311f3-2074-4336-ba7c-dc676f60d452" containerName="horizon-log" containerID="cri-o://331dfc0a6e5abb0e07ca1ea14a78e5f63c07bd07abbf90a2b2dc0c61ad816a52" gracePeriod=30 Nov 26 01:45:12 crc kubenswrapper[4766]: I1126 01:45:12.664010 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-65bd4d46f6-blndr" podUID="9e2311f3-2074-4336-ba7c-dc676f60d452" containerName="horizon" containerID="cri-o://425ff6f4884d714d97e857b8513f474a1d431613a4dc5063d13ebf02dcb8adad" gracePeriod=30 Nov 26 01:45:13 crc kubenswrapper[4766]: I1126 01:45:13.632972 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f7f64877f-4sskz" Nov 26 01:45:13 crc kubenswrapper[4766]: I1126 01:45:13.683286 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f7f64877f-4sskz" event={"ID":"656cb04b-f831-4671-a102-d01c6d01714f","Type":"ContainerDied","Data":"d542487007485be94fc05162b14b1d011807e022670da76db1cc6656b1176848"} Nov 26 01:45:13 crc kubenswrapper[4766]: I1126 01:45:13.683377 4766 scope.go:117] "RemoveContainer" containerID="dcfd1bcc023c09f3d1070f702e5f6368c5444f8c30ec70ebff460066679e4f44" Nov 26 01:45:13 crc kubenswrapper[4766]: I1126 01:45:13.683689 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f7f64877f-4sskz" Nov 26 01:45:13 crc kubenswrapper[4766]: I1126 01:45:13.686139 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qf7b\" (UniqueName: \"kubernetes.io/projected/656cb04b-f831-4671-a102-d01c6d01714f-kube-api-access-4qf7b\") pod \"656cb04b-f831-4671-a102-d01c6d01714f\" (UID: \"656cb04b-f831-4671-a102-d01c6d01714f\") " Nov 26 01:45:13 crc kubenswrapper[4766]: I1126 01:45:13.686343 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/656cb04b-f831-4671-a102-d01c6d01714f-scripts\") pod \"656cb04b-f831-4671-a102-d01c6d01714f\" (UID: \"656cb04b-f831-4671-a102-d01c6d01714f\") " Nov 26 01:45:13 crc kubenswrapper[4766]: I1126 01:45:13.686448 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/656cb04b-f831-4671-a102-d01c6d01714f-logs\") pod \"656cb04b-f831-4671-a102-d01c6d01714f\" (UID: \"656cb04b-f831-4671-a102-d01c6d01714f\") " Nov 26 01:45:13 crc kubenswrapper[4766]: I1126 01:45:13.686687 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/656cb04b-f831-4671-a102-d01c6d01714f-horizon-secret-key\") pod \"656cb04b-f831-4671-a102-d01c6d01714f\" (UID: \"656cb04b-f831-4671-a102-d01c6d01714f\") " Nov 26 01:45:13 crc kubenswrapper[4766]: I1126 01:45:13.686731 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/656cb04b-f831-4671-a102-d01c6d01714f-config-data\") pod \"656cb04b-f831-4671-a102-d01c6d01714f\" (UID: \"656cb04b-f831-4671-a102-d01c6d01714f\") " Nov 26 01:45:13 crc kubenswrapper[4766]: I1126 01:45:13.687304 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/656cb04b-f831-4671-a102-d01c6d01714f-logs" (OuterVolumeSpecName: "logs") pod "656cb04b-f831-4671-a102-d01c6d01714f" (UID: "656cb04b-f831-4671-a102-d01c6d01714f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:45:13 crc kubenswrapper[4766]: I1126 01:45:13.688923 4766 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/656cb04b-f831-4671-a102-d01c6d01714f-logs\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:13 crc kubenswrapper[4766]: I1126 01:45:13.698104 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/656cb04b-f831-4671-a102-d01c6d01714f-kube-api-access-4qf7b" (OuterVolumeSpecName: "kube-api-access-4qf7b") pod "656cb04b-f831-4671-a102-d01c6d01714f" (UID: "656cb04b-f831-4671-a102-d01c6d01714f"). InnerVolumeSpecName "kube-api-access-4qf7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:45:13 crc kubenswrapper[4766]: I1126 01:45:13.698222 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/656cb04b-f831-4671-a102-d01c6d01714f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "656cb04b-f831-4671-a102-d01c6d01714f" (UID: "656cb04b-f831-4671-a102-d01c6d01714f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:13 crc kubenswrapper[4766]: I1126 01:45:13.748613 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/656cb04b-f831-4671-a102-d01c6d01714f-config-data" (OuterVolumeSpecName: "config-data") pod "656cb04b-f831-4671-a102-d01c6d01714f" (UID: "656cb04b-f831-4671-a102-d01c6d01714f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:45:13 crc kubenswrapper[4766]: I1126 01:45:13.766047 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/656cb04b-f831-4671-a102-d01c6d01714f-scripts" (OuterVolumeSpecName: "scripts") pod "656cb04b-f831-4671-a102-d01c6d01714f" (UID: "656cb04b-f831-4671-a102-d01c6d01714f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:45:13 crc kubenswrapper[4766]: I1126 01:45:13.791468 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/656cb04b-f831-4671-a102-d01c6d01714f-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:13 crc kubenswrapper[4766]: I1126 01:45:13.791514 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qf7b\" (UniqueName: \"kubernetes.io/projected/656cb04b-f831-4671-a102-d01c6d01714f-kube-api-access-4qf7b\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:13 crc kubenswrapper[4766]: I1126 01:45:13.791526 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/656cb04b-f831-4671-a102-d01c6d01714f-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:13 crc kubenswrapper[4766]: I1126 01:45:13.791534 4766 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/656cb04b-f831-4671-a102-d01c6d01714f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:13 crc kubenswrapper[4766]: I1126 01:45:13.858563 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 01:45:13 crc kubenswrapper[4766]: I1126 01:45:13.905697 4766 scope.go:117] "RemoveContainer" containerID="fe8bea5f9be1857687215d30d22402fc1211456a7511bb72ccd1c779511c43e5" Nov 26 01:45:14 crc kubenswrapper[4766]: I1126 01:45:14.009233 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f7f64877f-4sskz"] Nov 26 01:45:14 crc kubenswrapper[4766]: I1126 01:45:14.020632 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7f7f64877f-4sskz"] Nov 26 01:45:14 crc kubenswrapper[4766]: I1126 01:45:14.703948 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1118ff67-6cfc-487c-8887-071283f2cc18","Type":"ContainerStarted","Data":"7686ca0421b6dd799426d8a1e3aad0450a845c4e0d9ee7d3056a14e88cee7c54"} Nov 26 01:45:14 crc kubenswrapper[4766]: W1126 01:45:14.814182 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc561040f_fa3f_4b22_b85a_cd50fbcd4020.slice/crio-e66a500dd654a7b245a0f11d862da5fb8d96dbe9f8c99a2551101f26662c1c2e WatchSource:0}: Error finding container e66a500dd654a7b245a0f11d862da5fb8d96dbe9f8c99a2551101f26662c1c2e: Status 404 returned error can't find the container with id e66a500dd654a7b245a0f11d862da5fb8d96dbe9f8c99a2551101f26662c1c2e Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.115452 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.116461 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f5b87f5c-z9b2b" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.223136 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c577b8bb-a3c5-405d-8da8-0289e8483ba5-horizon-secret-key\") pod \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\" (UID: \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\") " Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.223329 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c577b8bb-a3c5-405d-8da8-0289e8483ba5-config-data\") pod \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\" (UID: \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\") " Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.223357 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-ovsdbserver-sb\") pod \"f586d76c-85fe-40dc-8158-52a1747703d7\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.223400 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-dns-swift-storage-0\") pod \"f586d76c-85fe-40dc-8158-52a1747703d7\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.223428 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c577b8bb-a3c5-405d-8da8-0289e8483ba5-logs\") pod \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\" (UID: \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\") " Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.223479 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnlf9\" (UniqueName: \"kubernetes.io/projected/c577b8bb-a3c5-405d-8da8-0289e8483ba5-kube-api-access-fnlf9\") pod \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\" (UID: \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\") " Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.223545 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-dns-svc\") pod \"f586d76c-85fe-40dc-8158-52a1747703d7\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.223586 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpxqs\" (UniqueName: \"kubernetes.io/projected/f586d76c-85fe-40dc-8158-52a1747703d7-kube-api-access-jpxqs\") pod \"f586d76c-85fe-40dc-8158-52a1747703d7\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.223674 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-ovsdbserver-nb\") pod \"f586d76c-85fe-40dc-8158-52a1747703d7\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.223704 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c577b8bb-a3c5-405d-8da8-0289e8483ba5-scripts\") pod \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\" (UID: \"c577b8bb-a3c5-405d-8da8-0289e8483ba5\") " Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.223757 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-config\") pod \"f586d76c-85fe-40dc-8158-52a1747703d7\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.223793 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-openstack-edpm-ipam\") pod \"f586d76c-85fe-40dc-8158-52a1747703d7\" (UID: \"f586d76c-85fe-40dc-8158-52a1747703d7\") " Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.224313 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c577b8bb-a3c5-405d-8da8-0289e8483ba5-logs" (OuterVolumeSpecName: "logs") pod "c577b8bb-a3c5-405d-8da8-0289e8483ba5" (UID: "c577b8bb-a3c5-405d-8da8-0289e8483ba5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.266401 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c577b8bb-a3c5-405d-8da8-0289e8483ba5-kube-api-access-fnlf9" (OuterVolumeSpecName: "kube-api-access-fnlf9") pod "c577b8bb-a3c5-405d-8da8-0289e8483ba5" (UID: "c577b8bb-a3c5-405d-8da8-0289e8483ba5"). InnerVolumeSpecName "kube-api-access-fnlf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.268385 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f586d76c-85fe-40dc-8158-52a1747703d7-kube-api-access-jpxqs" (OuterVolumeSpecName: "kube-api-access-jpxqs") pod "f586d76c-85fe-40dc-8158-52a1747703d7" (UID: "f586d76c-85fe-40dc-8158-52a1747703d7"). InnerVolumeSpecName "kube-api-access-jpxqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.273773 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c577b8bb-a3c5-405d-8da8-0289e8483ba5-scripts" (OuterVolumeSpecName: "scripts") pod "c577b8bb-a3c5-405d-8da8-0289e8483ba5" (UID: "c577b8bb-a3c5-405d-8da8-0289e8483ba5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.273845 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c577b8bb-a3c5-405d-8da8-0289e8483ba5-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "c577b8bb-a3c5-405d-8da8-0289e8483ba5" (UID: "c577b8bb-a3c5-405d-8da8-0289e8483ba5"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.294983 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c577b8bb-a3c5-405d-8da8-0289e8483ba5-config-data" (OuterVolumeSpecName: "config-data") pod "c577b8bb-a3c5-405d-8da8-0289e8483ba5" (UID: "c577b8bb-a3c5-405d-8da8-0289e8483ba5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.318008 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f586d76c-85fe-40dc-8158-52a1747703d7" (UID: "f586d76c-85fe-40dc-8158-52a1747703d7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.328017 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c577b8bb-a3c5-405d-8da8-0289e8483ba5-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.328057 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.328116 4766 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c577b8bb-a3c5-405d-8da8-0289e8483ba5-logs\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.328129 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnlf9\" (UniqueName: \"kubernetes.io/projected/c577b8bb-a3c5-405d-8da8-0289e8483ba5-kube-api-access-fnlf9\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.328144 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpxqs\" (UniqueName: \"kubernetes.io/projected/f586d76c-85fe-40dc-8158-52a1747703d7-kube-api-access-jpxqs\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.328157 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c577b8bb-a3c5-405d-8da8-0289e8483ba5-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.328170 4766 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c577b8bb-a3c5-405d-8da8-0289e8483ba5-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.333055 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "f586d76c-85fe-40dc-8158-52a1747703d7" (UID: "f586d76c-85fe-40dc-8158-52a1747703d7"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.340276 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-config" (OuterVolumeSpecName: "config") pod "f586d76c-85fe-40dc-8158-52a1747703d7" (UID: "f586d76c-85fe-40dc-8158-52a1747703d7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.355075 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f586d76c-85fe-40dc-8158-52a1747703d7" (UID: "f586d76c-85fe-40dc-8158-52a1747703d7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.362589 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f586d76c-85fe-40dc-8158-52a1747703d7" (UID: "f586d76c-85fe-40dc-8158-52a1747703d7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.376386 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f586d76c-85fe-40dc-8158-52a1747703d7" (UID: "f586d76c-85fe-40dc-8158-52a1747703d7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.430645 4766 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.430687 4766 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.430698 4766 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.430707 4766 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-config\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.430718 4766 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f586d76c-85fe-40dc-8158-52a1747703d7-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.734994 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" event={"ID":"f586d76c-85fe-40dc-8158-52a1747703d7","Type":"ContainerDied","Data":"904092226a87448a7c622cf5243115d77ef33901904e7b56b2ead915230935d5"} Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.735031 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.735043 4766 scope.go:117] "RemoveContainer" containerID="38c1d8e99a74fa7e315f5cf2480050f5585758a433f3aec3b382768482a5866c" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.747121 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1118ff67-6cfc-487c-8887-071283f2cc18","Type":"ContainerStarted","Data":"e7810811e77fd0d22b61f946b9119b8302e45e6df86a3ca99f2301baa3361aea"} Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.752633 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f5b87f5c-z9b2b" event={"ID":"c577b8bb-a3c5-405d-8da8-0289e8483ba5","Type":"ContainerDied","Data":"975a79f0346bca6295480b6b00c88473a7568423bd71eb337e3e068bede48f83"} Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.752704 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f5b87f5c-z9b2b" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.778726 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c561040f-fa3f-4b22-b85a-cd50fbcd4020","Type":"ContainerStarted","Data":"804a79dc3edf3bad1d0d8f777c44cb0df6dabfbff35e3f379d3a18dcaafbdad8"} Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.778768 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c561040f-fa3f-4b22-b85a-cd50fbcd4020","Type":"ContainerStarted","Data":"e66a500dd654a7b245a0f11d862da5fb8d96dbe9f8c99a2551101f26662c1c2e"} Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.808966 4766 scope.go:117] "RemoveContainer" containerID="394dc6058d6729712a339c628479c02ab45716008c538eead220353c77dcf99b" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.835629 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=4.27543561 podStartE2EDuration="14.835592378s" podCreationTimestamp="2025-11-26 01:45:01 +0000 UTC" firstStartedPulling="2025-11-26 01:45:02.735721149 +0000 UTC m=+4883.584491579" lastFinishedPulling="2025-11-26 01:45:13.295877917 +0000 UTC m=+4894.144648347" observedRunningTime="2025-11-26 01:45:15.824530626 +0000 UTC m=+4896.673301076" watchObservedRunningTime="2025-11-26 01:45:15.835592378 +0000 UTC m=+4896.684362818" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.850146 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="656cb04b-f831-4671-a102-d01c6d01714f" path="/var/lib/kubelet/pods/656cb04b-f831-4671-a102-d01c6d01714f/volumes" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.892797 4766 scope.go:117] "RemoveContainer" containerID="797e1e95d5a6b9de52f72bb3fef2fafe5c0c40666f3b784bc4621f22407d05f2" Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.900742 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f5b87f5c-z9b2b"] Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.920725 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7f5b87f5c-z9b2b"] Nov 26 01:45:15 crc kubenswrapper[4766]: I1126 01:45:15.991249 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9"] Nov 26 01:45:16 crc kubenswrapper[4766]: I1126 01:45:16.006391 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cf7b6cbf7-jxdn9"] Nov 26 01:45:16 crc kubenswrapper[4766]: I1126 01:45:16.114540 4766 scope.go:117] "RemoveContainer" containerID="1c619971b704f1525550b524ab022265c6c1aaa8fd615b299de74ec0f58aaaae" Nov 26 01:45:16 crc kubenswrapper[4766]: I1126 01:45:16.791560 4766 generic.go:334] "Generic (PLEG): container finished" podID="9e2311f3-2074-4336-ba7c-dc676f60d452" containerID="425ff6f4884d714d97e857b8513f474a1d431613a4dc5063d13ebf02dcb8adad" exitCode=0 Nov 26 01:45:16 crc kubenswrapper[4766]: I1126 01:45:16.791643 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65bd4d46f6-blndr" event={"ID":"9e2311f3-2074-4336-ba7c-dc676f60d452","Type":"ContainerDied","Data":"425ff6f4884d714d97e857b8513f474a1d431613a4dc5063d13ebf02dcb8adad"} Nov 26 01:45:17 crc kubenswrapper[4766]: I1126 01:45:17.042124 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 01:45:17 crc kubenswrapper[4766]: I1126 01:45:17.307144 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-65bd4d46f6-blndr" podUID="9e2311f3-2074-4336-ba7c-dc676f60d452" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.71:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.71:8443: connect: connection refused" Nov 26 01:45:17 crc kubenswrapper[4766]: I1126 01:45:17.808638 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c561040f-fa3f-4b22-b85a-cd50fbcd4020","Type":"ContainerStarted","Data":"304a9ad816704f348b8775d5c035d25428bd7a1f3031c7fdd4615b798e3fd175"} Nov 26 01:45:17 crc kubenswrapper[4766]: I1126 01:45:17.843235 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c577b8bb-a3c5-405d-8da8-0289e8483ba5" path="/var/lib/kubelet/pods/c577b8bb-a3c5-405d-8da8-0289e8483ba5/volumes" Nov 26 01:45:17 crc kubenswrapper[4766]: I1126 01:45:17.844058 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f586d76c-85fe-40dc-8158-52a1747703d7" path="/var/lib/kubelet/pods/f586d76c-85fe-40dc-8158-52a1747703d7/volumes" Nov 26 01:45:18 crc kubenswrapper[4766]: I1126 01:45:18.824188 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c561040f-fa3f-4b22-b85a-cd50fbcd4020","Type":"ContainerStarted","Data":"ba517bafa142e7856741cb2767af497c86eb90b806f0bd6d8b696e4ba5fd9178"} Nov 26 01:45:20 crc kubenswrapper[4766]: I1126 01:45:20.857812 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c561040f-fa3f-4b22-b85a-cd50fbcd4020","Type":"ContainerStarted","Data":"ec47c2b1e5898d5099a1c792d9bb69f2c6e146c4b4afb81e6c0017b8d8c28663"} Nov 26 01:45:20 crc kubenswrapper[4766]: I1126 01:45:20.858571 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 26 01:45:20 crc kubenswrapper[4766]: I1126 01:45:20.858173 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c561040f-fa3f-4b22-b85a-cd50fbcd4020" containerName="sg-core" containerID="cri-o://ba517bafa142e7856741cb2767af497c86eb90b806f0bd6d8b696e4ba5fd9178" gracePeriod=30 Nov 26 01:45:20 crc kubenswrapper[4766]: I1126 01:45:20.858140 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c561040f-fa3f-4b22-b85a-cd50fbcd4020" containerName="proxy-httpd" containerID="cri-o://ec47c2b1e5898d5099a1c792d9bb69f2c6e146c4b4afb81e6c0017b8d8c28663" gracePeriod=30 Nov 26 01:45:20 crc kubenswrapper[4766]: I1126 01:45:20.858110 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c561040f-fa3f-4b22-b85a-cd50fbcd4020" containerName="ceilometer-central-agent" containerID="cri-o://804a79dc3edf3bad1d0d8f777c44cb0df6dabfbff35e3f379d3a18dcaafbdad8" gracePeriod=30 Nov 26 01:45:20 crc kubenswrapper[4766]: I1126 01:45:20.858262 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c561040f-fa3f-4b22-b85a-cd50fbcd4020" containerName="ceilometer-notification-agent" containerID="cri-o://304a9ad816704f348b8775d5c035d25428bd7a1f3031c7fdd4615b798e3fd175" gracePeriod=30 Nov 26 01:45:21 crc kubenswrapper[4766]: I1126 01:45:21.873706 4766 generic.go:334] "Generic (PLEG): container finished" podID="c561040f-fa3f-4b22-b85a-cd50fbcd4020" containerID="ec47c2b1e5898d5099a1c792d9bb69f2c6e146c4b4afb81e6c0017b8d8c28663" exitCode=0 Nov 26 01:45:21 crc kubenswrapper[4766]: I1126 01:45:21.874061 4766 generic.go:334] "Generic (PLEG): container finished" podID="c561040f-fa3f-4b22-b85a-cd50fbcd4020" containerID="ba517bafa142e7856741cb2767af497c86eb90b806f0bd6d8b696e4ba5fd9178" exitCode=2 Nov 26 01:45:21 crc kubenswrapper[4766]: I1126 01:45:21.874074 4766 generic.go:334] "Generic (PLEG): container finished" podID="c561040f-fa3f-4b22-b85a-cd50fbcd4020" containerID="304a9ad816704f348b8775d5c035d25428bd7a1f3031c7fdd4615b798e3fd175" exitCode=0 Nov 26 01:45:21 crc kubenswrapper[4766]: I1126 01:45:21.874094 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c561040f-fa3f-4b22-b85a-cd50fbcd4020","Type":"ContainerDied","Data":"ec47c2b1e5898d5099a1c792d9bb69f2c6e146c4b4afb81e6c0017b8d8c28663"} Nov 26 01:45:21 crc kubenswrapper[4766]: I1126 01:45:21.874118 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c561040f-fa3f-4b22-b85a-cd50fbcd4020","Type":"ContainerDied","Data":"ba517bafa142e7856741cb2767af497c86eb90b806f0bd6d8b696e4ba5fd9178"} Nov 26 01:45:21 crc kubenswrapper[4766]: I1126 01:45:21.874130 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c561040f-fa3f-4b22-b85a-cd50fbcd4020","Type":"ContainerDied","Data":"304a9ad816704f348b8775d5c035d25428bd7a1f3031c7fdd4615b798e3fd175"} Nov 26 01:45:21 crc kubenswrapper[4766]: I1126 01:45:21.954021 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Nov 26 01:45:23 crc kubenswrapper[4766]: I1126 01:45:23.897874 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Nov 26 01:45:23 crc kubenswrapper[4766]: I1126 01:45:23.922805 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=10.065569782 podStartE2EDuration="14.922774342s" podCreationTimestamp="2025-11-26 01:45:09 +0000 UTC" firstStartedPulling="2025-11-26 01:45:14.887203463 +0000 UTC m=+4895.735973893" lastFinishedPulling="2025-11-26 01:45:19.744408023 +0000 UTC m=+4900.593178453" observedRunningTime="2025-11-26 01:45:20.893356367 +0000 UTC m=+4901.742126807" watchObservedRunningTime="2025-11-26 01:45:23.922774342 +0000 UTC m=+4904.771544812" Nov 26 01:45:23 crc kubenswrapper[4766]: I1126 01:45:23.972871 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Nov 26 01:45:23 crc kubenswrapper[4766]: I1126 01:45:23.973115 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="1c921728-52e2-4b04-8202-ac19c632833c" containerName="manila-scheduler" containerID="cri-o://d7357d8b357a23f6567bce95ef9894289ea485a98616ce1f81463ec94202a8b7" gracePeriod=30 Nov 26 01:45:23 crc kubenswrapper[4766]: I1126 01:45:23.973601 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="1c921728-52e2-4b04-8202-ac19c632833c" containerName="probe" containerID="cri-o://6fe68e9a872e3c7c16975868bb45f96814c70638cbfce6ec0e77ad8f83406632" gracePeriod=30 Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.593424 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.752611 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c561040f-fa3f-4b22-b85a-cd50fbcd4020-log-httpd\") pod \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.752917 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c561040f-fa3f-4b22-b85a-cd50fbcd4020-run-httpd\") pod \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.753076 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-sg-core-conf-yaml\") pod \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.753118 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c561040f-fa3f-4b22-b85a-cd50fbcd4020-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c561040f-fa3f-4b22-b85a-cd50fbcd4020" (UID: "c561040f-fa3f-4b22-b85a-cd50fbcd4020"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.753235 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c561040f-fa3f-4b22-b85a-cd50fbcd4020-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c561040f-fa3f-4b22-b85a-cd50fbcd4020" (UID: "c561040f-fa3f-4b22-b85a-cd50fbcd4020"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.753177 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-scripts\") pod \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.753375 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-combined-ca-bundle\") pod \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.753508 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-ceilometer-tls-certs\") pod \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.753612 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcwkh\" (UniqueName: \"kubernetes.io/projected/c561040f-fa3f-4b22-b85a-cd50fbcd4020-kube-api-access-kcwkh\") pod \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.753698 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-config-data\") pod \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\" (UID: \"c561040f-fa3f-4b22-b85a-cd50fbcd4020\") " Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.755430 4766 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c561040f-fa3f-4b22-b85a-cd50fbcd4020-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.755457 4766 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c561040f-fa3f-4b22-b85a-cd50fbcd4020-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.760578 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-scripts" (OuterVolumeSpecName: "scripts") pod "c561040f-fa3f-4b22-b85a-cd50fbcd4020" (UID: "c561040f-fa3f-4b22-b85a-cd50fbcd4020"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.764685 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c561040f-fa3f-4b22-b85a-cd50fbcd4020-kube-api-access-kcwkh" (OuterVolumeSpecName: "kube-api-access-kcwkh") pod "c561040f-fa3f-4b22-b85a-cd50fbcd4020" (UID: "c561040f-fa3f-4b22-b85a-cd50fbcd4020"). InnerVolumeSpecName "kube-api-access-kcwkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.796103 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c561040f-fa3f-4b22-b85a-cd50fbcd4020" (UID: "c561040f-fa3f-4b22-b85a-cd50fbcd4020"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.830185 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "c561040f-fa3f-4b22-b85a-cd50fbcd4020" (UID: "c561040f-fa3f-4b22-b85a-cd50fbcd4020"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.853996 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c561040f-fa3f-4b22-b85a-cd50fbcd4020" (UID: "c561040f-fa3f-4b22-b85a-cd50fbcd4020"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.857374 4766 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.857406 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.857416 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.857424 4766 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.857433 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcwkh\" (UniqueName: \"kubernetes.io/projected/c561040f-fa3f-4b22-b85a-cd50fbcd4020-kube-api-access-kcwkh\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.926538 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-config-data" (OuterVolumeSpecName: "config-data") pod "c561040f-fa3f-4b22-b85a-cd50fbcd4020" (UID: "c561040f-fa3f-4b22-b85a-cd50fbcd4020"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.931420 4766 generic.go:334] "Generic (PLEG): container finished" podID="1c921728-52e2-4b04-8202-ac19c632833c" containerID="6fe68e9a872e3c7c16975868bb45f96814c70638cbfce6ec0e77ad8f83406632" exitCode=0 Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.931480 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"1c921728-52e2-4b04-8202-ac19c632833c","Type":"ContainerDied","Data":"6fe68e9a872e3c7c16975868bb45f96814c70638cbfce6ec0e77ad8f83406632"} Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.935315 4766 generic.go:334] "Generic (PLEG): container finished" podID="c561040f-fa3f-4b22-b85a-cd50fbcd4020" containerID="804a79dc3edf3bad1d0d8f777c44cb0df6dabfbff35e3f379d3a18dcaafbdad8" exitCode=0 Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.935359 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c561040f-fa3f-4b22-b85a-cd50fbcd4020","Type":"ContainerDied","Data":"804a79dc3edf3bad1d0d8f777c44cb0df6dabfbff35e3f379d3a18dcaafbdad8"} Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.935568 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c561040f-fa3f-4b22-b85a-cd50fbcd4020","Type":"ContainerDied","Data":"e66a500dd654a7b245a0f11d862da5fb8d96dbe9f8c99a2551101f26662c1c2e"} Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.935386 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.935620 4766 scope.go:117] "RemoveContainer" containerID="ec47c2b1e5898d5099a1c792d9bb69f2c6e146c4b4afb81e6c0017b8d8c28663" Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.958977 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c561040f-fa3f-4b22-b85a-cd50fbcd4020-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.963321 4766 scope.go:117] "RemoveContainer" containerID="ba517bafa142e7856741cb2767af497c86eb90b806f0bd6d8b696e4ba5fd9178" Nov 26 01:45:24 crc kubenswrapper[4766]: I1126 01:45:24.989437 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.002381 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.012360 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 26 01:45:25 crc kubenswrapper[4766]: E1126 01:45:25.012815 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f586d76c-85fe-40dc-8158-52a1747703d7" containerName="dnsmasq-dns" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.012832 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="f586d76c-85fe-40dc-8158-52a1747703d7" containerName="dnsmasq-dns" Nov 26 01:45:25 crc kubenswrapper[4766]: E1126 01:45:25.012855 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f586d76c-85fe-40dc-8158-52a1747703d7" containerName="init" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.012862 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="f586d76c-85fe-40dc-8158-52a1747703d7" containerName="init" Nov 26 01:45:25 crc kubenswrapper[4766]: E1126 01:45:25.012881 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c561040f-fa3f-4b22-b85a-cd50fbcd4020" containerName="ceilometer-notification-agent" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.012887 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c561040f-fa3f-4b22-b85a-cd50fbcd4020" containerName="ceilometer-notification-agent" Nov 26 01:45:25 crc kubenswrapper[4766]: E1126 01:45:25.012899 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c561040f-fa3f-4b22-b85a-cd50fbcd4020" containerName="sg-core" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.012904 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c561040f-fa3f-4b22-b85a-cd50fbcd4020" containerName="sg-core" Nov 26 01:45:25 crc kubenswrapper[4766]: E1126 01:45:25.012916 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c577b8bb-a3c5-405d-8da8-0289e8483ba5" containerName="horizon-log" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.012921 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c577b8bb-a3c5-405d-8da8-0289e8483ba5" containerName="horizon-log" Nov 26 01:45:25 crc kubenswrapper[4766]: E1126 01:45:25.012938 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c577b8bb-a3c5-405d-8da8-0289e8483ba5" containerName="horizon" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.012944 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c577b8bb-a3c5-405d-8da8-0289e8483ba5" containerName="horizon" Nov 26 01:45:25 crc kubenswrapper[4766]: E1126 01:45:25.012956 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="656cb04b-f831-4671-a102-d01c6d01714f" containerName="horizon" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.012962 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="656cb04b-f831-4671-a102-d01c6d01714f" containerName="horizon" Nov 26 01:45:25 crc kubenswrapper[4766]: E1126 01:45:25.012971 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c561040f-fa3f-4b22-b85a-cd50fbcd4020" containerName="ceilometer-central-agent" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.012977 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c561040f-fa3f-4b22-b85a-cd50fbcd4020" containerName="ceilometer-central-agent" Nov 26 01:45:25 crc kubenswrapper[4766]: E1126 01:45:25.013004 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c561040f-fa3f-4b22-b85a-cd50fbcd4020" containerName="proxy-httpd" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.013010 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c561040f-fa3f-4b22-b85a-cd50fbcd4020" containerName="proxy-httpd" Nov 26 01:45:25 crc kubenswrapper[4766]: E1126 01:45:25.013027 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="656cb04b-f831-4671-a102-d01c6d01714f" containerName="horizon-log" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.013032 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="656cb04b-f831-4671-a102-d01c6d01714f" containerName="horizon-log" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.013246 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="c561040f-fa3f-4b22-b85a-cd50fbcd4020" containerName="proxy-httpd" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.013260 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="c577b8bb-a3c5-405d-8da8-0289e8483ba5" containerName="horizon-log" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.013290 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="c561040f-fa3f-4b22-b85a-cd50fbcd4020" containerName="ceilometer-central-agent" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.013301 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="c561040f-fa3f-4b22-b85a-cd50fbcd4020" containerName="sg-core" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.013319 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="656cb04b-f831-4671-a102-d01c6d01714f" containerName="horizon" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.013332 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="f586d76c-85fe-40dc-8158-52a1747703d7" containerName="dnsmasq-dns" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.013348 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="c561040f-fa3f-4b22-b85a-cd50fbcd4020" containerName="ceilometer-notification-agent" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.013360 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="c577b8bb-a3c5-405d-8da8-0289e8483ba5" containerName="horizon" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.013370 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="656cb04b-f831-4671-a102-d01c6d01714f" containerName="horizon-log" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.015223 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.019884 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.020012 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.020159 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.030202 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.084880 4766 scope.go:117] "RemoveContainer" containerID="304a9ad816704f348b8775d5c035d25428bd7a1f3031c7fdd4615b798e3fd175" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.115568 4766 scope.go:117] "RemoveContainer" containerID="804a79dc3edf3bad1d0d8f777c44cb0df6dabfbff35e3f379d3a18dcaafbdad8" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.145227 4766 scope.go:117] "RemoveContainer" containerID="ec47c2b1e5898d5099a1c792d9bb69f2c6e146c4b4afb81e6c0017b8d8c28663" Nov 26 01:45:25 crc kubenswrapper[4766]: E1126 01:45:25.145858 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec47c2b1e5898d5099a1c792d9bb69f2c6e146c4b4afb81e6c0017b8d8c28663\": container with ID starting with ec47c2b1e5898d5099a1c792d9bb69f2c6e146c4b4afb81e6c0017b8d8c28663 not found: ID does not exist" containerID="ec47c2b1e5898d5099a1c792d9bb69f2c6e146c4b4afb81e6c0017b8d8c28663" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.145919 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec47c2b1e5898d5099a1c792d9bb69f2c6e146c4b4afb81e6c0017b8d8c28663"} err="failed to get container status \"ec47c2b1e5898d5099a1c792d9bb69f2c6e146c4b4afb81e6c0017b8d8c28663\": rpc error: code = NotFound desc = could not find container \"ec47c2b1e5898d5099a1c792d9bb69f2c6e146c4b4afb81e6c0017b8d8c28663\": container with ID starting with ec47c2b1e5898d5099a1c792d9bb69f2c6e146c4b4afb81e6c0017b8d8c28663 not found: ID does not exist" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.145945 4766 scope.go:117] "RemoveContainer" containerID="ba517bafa142e7856741cb2767af497c86eb90b806f0bd6d8b696e4ba5fd9178" Nov 26 01:45:25 crc kubenswrapper[4766]: E1126 01:45:25.146454 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba517bafa142e7856741cb2767af497c86eb90b806f0bd6d8b696e4ba5fd9178\": container with ID starting with ba517bafa142e7856741cb2767af497c86eb90b806f0bd6d8b696e4ba5fd9178 not found: ID does not exist" containerID="ba517bafa142e7856741cb2767af497c86eb90b806f0bd6d8b696e4ba5fd9178" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.146482 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba517bafa142e7856741cb2767af497c86eb90b806f0bd6d8b696e4ba5fd9178"} err="failed to get container status \"ba517bafa142e7856741cb2767af497c86eb90b806f0bd6d8b696e4ba5fd9178\": rpc error: code = NotFound desc = could not find container \"ba517bafa142e7856741cb2767af497c86eb90b806f0bd6d8b696e4ba5fd9178\": container with ID starting with ba517bafa142e7856741cb2767af497c86eb90b806f0bd6d8b696e4ba5fd9178 not found: ID does not exist" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.146502 4766 scope.go:117] "RemoveContainer" containerID="304a9ad816704f348b8775d5c035d25428bd7a1f3031c7fdd4615b798e3fd175" Nov 26 01:45:25 crc kubenswrapper[4766]: E1126 01:45:25.147007 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"304a9ad816704f348b8775d5c035d25428bd7a1f3031c7fdd4615b798e3fd175\": container with ID starting with 304a9ad816704f348b8775d5c035d25428bd7a1f3031c7fdd4615b798e3fd175 not found: ID does not exist" containerID="304a9ad816704f348b8775d5c035d25428bd7a1f3031c7fdd4615b798e3fd175" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.147024 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"304a9ad816704f348b8775d5c035d25428bd7a1f3031c7fdd4615b798e3fd175"} err="failed to get container status \"304a9ad816704f348b8775d5c035d25428bd7a1f3031c7fdd4615b798e3fd175\": rpc error: code = NotFound desc = could not find container \"304a9ad816704f348b8775d5c035d25428bd7a1f3031c7fdd4615b798e3fd175\": container with ID starting with 304a9ad816704f348b8775d5c035d25428bd7a1f3031c7fdd4615b798e3fd175 not found: ID does not exist" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.147036 4766 scope.go:117] "RemoveContainer" containerID="804a79dc3edf3bad1d0d8f777c44cb0df6dabfbff35e3f379d3a18dcaafbdad8" Nov 26 01:45:25 crc kubenswrapper[4766]: E1126 01:45:25.147667 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"804a79dc3edf3bad1d0d8f777c44cb0df6dabfbff35e3f379d3a18dcaafbdad8\": container with ID starting with 804a79dc3edf3bad1d0d8f777c44cb0df6dabfbff35e3f379d3a18dcaafbdad8 not found: ID does not exist" containerID="804a79dc3edf3bad1d0d8f777c44cb0df6dabfbff35e3f379d3a18dcaafbdad8" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.147696 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"804a79dc3edf3bad1d0d8f777c44cb0df6dabfbff35e3f379d3a18dcaafbdad8"} err="failed to get container status \"804a79dc3edf3bad1d0d8f777c44cb0df6dabfbff35e3f379d3a18dcaafbdad8\": rpc error: code = NotFound desc = could not find container \"804a79dc3edf3bad1d0d8f777c44cb0df6dabfbff35e3f379d3a18dcaafbdad8\": container with ID starting with 804a79dc3edf3bad1d0d8f777c44cb0df6dabfbff35e3f379d3a18dcaafbdad8 not found: ID does not exist" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.163257 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c82c6d8-c395-4177-930f-f840d902df28-run-httpd\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.163307 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c82c6d8-c395-4177-930f-f840d902df28-scripts\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.163327 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c82c6d8-c395-4177-930f-f840d902df28-config-data\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.163577 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgfzl\" (UniqueName: \"kubernetes.io/projected/2c82c6d8-c395-4177-930f-f840d902df28-kube-api-access-fgfzl\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.163776 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c82c6d8-c395-4177-930f-f840d902df28-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.163832 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c82c6d8-c395-4177-930f-f840d902df28-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.163873 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c82c6d8-c395-4177-930f-f840d902df28-log-httpd\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.163923 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2c82c6d8-c395-4177-930f-f840d902df28-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.265676 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c82c6d8-c395-4177-930f-f840d902df28-run-httpd\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.265717 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c82c6d8-c395-4177-930f-f840d902df28-scripts\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.265730 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c82c6d8-c395-4177-930f-f840d902df28-config-data\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.265786 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgfzl\" (UniqueName: \"kubernetes.io/projected/2c82c6d8-c395-4177-930f-f840d902df28-kube-api-access-fgfzl\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.265824 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c82c6d8-c395-4177-930f-f840d902df28-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.265840 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c82c6d8-c395-4177-930f-f840d902df28-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.265861 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c82c6d8-c395-4177-930f-f840d902df28-log-httpd\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.265883 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2c82c6d8-c395-4177-930f-f840d902df28-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.268064 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c82c6d8-c395-4177-930f-f840d902df28-log-httpd\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.268158 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c82c6d8-c395-4177-930f-f840d902df28-run-httpd\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.272252 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c82c6d8-c395-4177-930f-f840d902df28-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.272905 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c82c6d8-c395-4177-930f-f840d902df28-scripts\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.273045 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c82c6d8-c395-4177-930f-f840d902df28-config-data\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.273591 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c82c6d8-c395-4177-930f-f840d902df28-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.283418 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgfzl\" (UniqueName: \"kubernetes.io/projected/2c82c6d8-c395-4177-930f-f840d902df28-kube-api-access-fgfzl\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.286219 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2c82c6d8-c395-4177-930f-f840d902df28-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2c82c6d8-c395-4177-930f-f840d902df28\") " pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.378930 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.856241 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c561040f-fa3f-4b22-b85a-cd50fbcd4020" path="/var/lib/kubelet/pods/c561040f-fa3f-4b22-b85a-cd50fbcd4020/volumes" Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.898748 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 01:45:25 crc kubenswrapper[4766]: W1126 01:45:25.949986 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c82c6d8_c395_4177_930f_f840d902df28.slice/crio-ab6b8b3fb1319bcbb6742f4f8c87ed01dd6cb3066e74935ec512ea82957ae6ec WatchSource:0}: Error finding container ab6b8b3fb1319bcbb6742f4f8c87ed01dd6cb3066e74935ec512ea82957ae6ec: Status 404 returned error can't find the container with id ab6b8b3fb1319bcbb6742f4f8c87ed01dd6cb3066e74935ec512ea82957ae6ec Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.954550 4766 generic.go:334] "Generic (PLEG): container finished" podID="1c921728-52e2-4b04-8202-ac19c632833c" containerID="d7357d8b357a23f6567bce95ef9894289ea485a98616ce1f81463ec94202a8b7" exitCode=0 Nov 26 01:45:25 crc kubenswrapper[4766]: I1126 01:45:25.954634 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"1c921728-52e2-4b04-8202-ac19c632833c","Type":"ContainerDied","Data":"d7357d8b357a23f6567bce95ef9894289ea485a98616ce1f81463ec94202a8b7"} Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.352274 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.501064 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-config-data-custom\") pod \"1c921728-52e2-4b04-8202-ac19c632833c\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.501165 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnq4r\" (UniqueName: \"kubernetes.io/projected/1c921728-52e2-4b04-8202-ac19c632833c-kube-api-access-mnq4r\") pod \"1c921728-52e2-4b04-8202-ac19c632833c\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.501285 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-config-data\") pod \"1c921728-52e2-4b04-8202-ac19c632833c\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.501422 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-scripts\") pod \"1c921728-52e2-4b04-8202-ac19c632833c\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.501441 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-combined-ca-bundle\") pod \"1c921728-52e2-4b04-8202-ac19c632833c\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.501853 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1c921728-52e2-4b04-8202-ac19c632833c-etc-machine-id\") pod \"1c921728-52e2-4b04-8202-ac19c632833c\" (UID: \"1c921728-52e2-4b04-8202-ac19c632833c\") " Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.501929 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c921728-52e2-4b04-8202-ac19c632833c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1c921728-52e2-4b04-8202-ac19c632833c" (UID: "1c921728-52e2-4b04-8202-ac19c632833c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.502431 4766 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1c921728-52e2-4b04-8202-ac19c632833c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.511305 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1c921728-52e2-4b04-8202-ac19c632833c" (UID: "1c921728-52e2-4b04-8202-ac19c632833c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.511332 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-scripts" (OuterVolumeSpecName: "scripts") pod "1c921728-52e2-4b04-8202-ac19c632833c" (UID: "1c921728-52e2-4b04-8202-ac19c632833c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.541094 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c921728-52e2-4b04-8202-ac19c632833c-kube-api-access-mnq4r" (OuterVolumeSpecName: "kube-api-access-mnq4r") pod "1c921728-52e2-4b04-8202-ac19c632833c" (UID: "1c921728-52e2-4b04-8202-ac19c632833c"). InnerVolumeSpecName "kube-api-access-mnq4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.605075 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.605436 4766 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.605451 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnq4r\" (UniqueName: \"kubernetes.io/projected/1c921728-52e2-4b04-8202-ac19c632833c-kube-api-access-mnq4r\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.622018 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1c921728-52e2-4b04-8202-ac19c632833c" (UID: "1c921728-52e2-4b04-8202-ac19c632833c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.645881 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-config-data" (OuterVolumeSpecName: "config-data") pod "1c921728-52e2-4b04-8202-ac19c632833c" (UID: "1c921728-52e2-4b04-8202-ac19c632833c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.707316 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.707346 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c921728-52e2-4b04-8202-ac19c632833c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.970992 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"1c921728-52e2-4b04-8202-ac19c632833c","Type":"ContainerDied","Data":"5fb34dd9c6155825e7d3e6c9c54a74e165886ad0d4c964fce3d80c53f7251bdb"} Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.971043 4766 scope.go:117] "RemoveContainer" containerID="6fe68e9a872e3c7c16975868bb45f96814c70638cbfce6ec0e77ad8f83406632" Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.971066 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.977404 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c82c6d8-c395-4177-930f-f840d902df28","Type":"ContainerStarted","Data":"dd6feb67c1e8e8c923c0464c8b6477e0baacbb5ad5f701f9c7e5f31e0c25887b"} Nov 26 01:45:26 crc kubenswrapper[4766]: I1126 01:45:26.977502 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c82c6d8-c395-4177-930f-f840d902df28","Type":"ContainerStarted","Data":"ab6b8b3fb1319bcbb6742f4f8c87ed01dd6cb3066e74935ec512ea82957ae6ec"} Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.082162 4766 scope.go:117] "RemoveContainer" containerID="d7357d8b357a23f6567bce95ef9894289ea485a98616ce1f81463ec94202a8b7" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.109816 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.131353 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.149720 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Nov 26 01:45:27 crc kubenswrapper[4766]: E1126 01:45:27.150195 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c921728-52e2-4b04-8202-ac19c632833c" containerName="manila-scheduler" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.150213 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c921728-52e2-4b04-8202-ac19c632833c" containerName="manila-scheduler" Nov 26 01:45:27 crc kubenswrapper[4766]: E1126 01:45:27.150247 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c921728-52e2-4b04-8202-ac19c632833c" containerName="probe" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.150253 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c921728-52e2-4b04-8202-ac19c632833c" containerName="probe" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.150470 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c921728-52e2-4b04-8202-ac19c632833c" containerName="manila-scheduler" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.150499 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c921728-52e2-4b04-8202-ac19c632833c" containerName="probe" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.151624 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.155004 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.175811 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.307731 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-65bd4d46f6-blndr" podUID="9e2311f3-2074-4336-ba7c-dc676f60d452" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.71:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.71:8443: connect: connection refused" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.321153 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a487b55f-cdaa-4351-95ed-7fb4a8ce382d-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"a487b55f-cdaa-4351-95ed-7fb4a8ce382d\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.321202 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a487b55f-cdaa-4351-95ed-7fb4a8ce382d-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"a487b55f-cdaa-4351-95ed-7fb4a8ce382d\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.321289 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcvwt\" (UniqueName: \"kubernetes.io/projected/a487b55f-cdaa-4351-95ed-7fb4a8ce382d-kube-api-access-zcvwt\") pod \"manila-scheduler-0\" (UID: \"a487b55f-cdaa-4351-95ed-7fb4a8ce382d\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.321322 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a487b55f-cdaa-4351-95ed-7fb4a8ce382d-scripts\") pod \"manila-scheduler-0\" (UID: \"a487b55f-cdaa-4351-95ed-7fb4a8ce382d\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.321594 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a487b55f-cdaa-4351-95ed-7fb4a8ce382d-config-data\") pod \"manila-scheduler-0\" (UID: \"a487b55f-cdaa-4351-95ed-7fb4a8ce382d\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.321759 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a487b55f-cdaa-4351-95ed-7fb4a8ce382d-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"a487b55f-cdaa-4351-95ed-7fb4a8ce382d\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.423680 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcvwt\" (UniqueName: \"kubernetes.io/projected/a487b55f-cdaa-4351-95ed-7fb4a8ce382d-kube-api-access-zcvwt\") pod \"manila-scheduler-0\" (UID: \"a487b55f-cdaa-4351-95ed-7fb4a8ce382d\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.423762 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a487b55f-cdaa-4351-95ed-7fb4a8ce382d-scripts\") pod \"manila-scheduler-0\" (UID: \"a487b55f-cdaa-4351-95ed-7fb4a8ce382d\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.423866 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a487b55f-cdaa-4351-95ed-7fb4a8ce382d-config-data\") pod \"manila-scheduler-0\" (UID: \"a487b55f-cdaa-4351-95ed-7fb4a8ce382d\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.423965 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a487b55f-cdaa-4351-95ed-7fb4a8ce382d-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"a487b55f-cdaa-4351-95ed-7fb4a8ce382d\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.424020 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a487b55f-cdaa-4351-95ed-7fb4a8ce382d-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"a487b55f-cdaa-4351-95ed-7fb4a8ce382d\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.424045 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a487b55f-cdaa-4351-95ed-7fb4a8ce382d-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"a487b55f-cdaa-4351-95ed-7fb4a8ce382d\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.426314 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a487b55f-cdaa-4351-95ed-7fb4a8ce382d-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"a487b55f-cdaa-4351-95ed-7fb4a8ce382d\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.429370 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a487b55f-cdaa-4351-95ed-7fb4a8ce382d-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"a487b55f-cdaa-4351-95ed-7fb4a8ce382d\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.431363 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a487b55f-cdaa-4351-95ed-7fb4a8ce382d-config-data\") pod \"manila-scheduler-0\" (UID: \"a487b55f-cdaa-4351-95ed-7fb4a8ce382d\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.440673 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a487b55f-cdaa-4351-95ed-7fb4a8ce382d-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"a487b55f-cdaa-4351-95ed-7fb4a8ce382d\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.443044 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a487b55f-cdaa-4351-95ed-7fb4a8ce382d-scripts\") pod \"manila-scheduler-0\" (UID: \"a487b55f-cdaa-4351-95ed-7fb4a8ce382d\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.443429 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcvwt\" (UniqueName: \"kubernetes.io/projected/a487b55f-cdaa-4351-95ed-7fb4a8ce382d-kube-api-access-zcvwt\") pod \"manila-scheduler-0\" (UID: \"a487b55f-cdaa-4351-95ed-7fb4a8ce382d\") " pod="openstack/manila-scheduler-0" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.524461 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.845209 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c921728-52e2-4b04-8202-ac19c632833c" path="/var/lib/kubelet/pods/1c921728-52e2-4b04-8202-ac19c632833c/volumes" Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.990182 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c82c6d8-c395-4177-930f-f840d902df28","Type":"ContainerStarted","Data":"e641631dfe7956ad3a0f34d533ceeb73b22161391c074fd6187d1139c9894012"} Nov 26 01:45:27 crc kubenswrapper[4766]: I1126 01:45:27.990219 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c82c6d8-c395-4177-930f-f840d902df28","Type":"ContainerStarted","Data":"3a0c29708833db52de36d7d61bdb2e2704c8e774924b034d3b18d8b5785d3fdb"} Nov 26 01:45:28 crc kubenswrapper[4766]: I1126 01:45:28.015333 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 26 01:45:29 crc kubenswrapper[4766]: I1126 01:45:29.017722 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"a487b55f-cdaa-4351-95ed-7fb4a8ce382d","Type":"ContainerStarted","Data":"01cc8151e99a0e48f931e57b85581ad6bb9f49562baa42b3aeb754290d9ea879"} Nov 26 01:45:29 crc kubenswrapper[4766]: I1126 01:45:29.017976 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"a487b55f-cdaa-4351-95ed-7fb4a8ce382d","Type":"ContainerStarted","Data":"04b30e2e888d5140ade8a89b5b2861b1170dfd0a0f371a71108fa7ef309183d6"} Nov 26 01:45:29 crc kubenswrapper[4766]: I1126 01:45:29.415975 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Nov 26 01:45:30 crc kubenswrapper[4766]: I1126 01:45:30.032211 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c82c6d8-c395-4177-930f-f840d902df28","Type":"ContainerStarted","Data":"da0e35f07c6c043e255a8a9c6a51819cba4c23fd2c1c417d089003890af6e825"} Nov 26 01:45:30 crc kubenswrapper[4766]: I1126 01:45:30.033281 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 26 01:45:30 crc kubenswrapper[4766]: I1126 01:45:30.035608 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"a487b55f-cdaa-4351-95ed-7fb4a8ce382d","Type":"ContainerStarted","Data":"bc60846023ed4fe3cd4ee14ac32bb600640cd3b7e5f057abd5c698c110017867"} Nov 26 01:45:30 crc kubenswrapper[4766]: I1126 01:45:30.063954 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.15908216 podStartE2EDuration="6.063935988s" podCreationTimestamp="2025-11-26 01:45:24 +0000 UTC" firstStartedPulling="2025-11-26 01:45:25.955145446 +0000 UTC m=+4906.803915876" lastFinishedPulling="2025-11-26 01:45:28.859999274 +0000 UTC m=+4909.708769704" observedRunningTime="2025-11-26 01:45:30.053716477 +0000 UTC m=+4910.902486907" watchObservedRunningTime="2025-11-26 01:45:30.063935988 +0000 UTC m=+4910.912706408" Nov 26 01:45:30 crc kubenswrapper[4766]: I1126 01:45:30.085697 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.085677971 podStartE2EDuration="3.085677971s" podCreationTimestamp="2025-11-26 01:45:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 01:45:30.075912892 +0000 UTC m=+4910.924683322" watchObservedRunningTime="2025-11-26 01:45:30.085677971 +0000 UTC m=+4910.934448401" Nov 26 01:45:33 crc kubenswrapper[4766]: I1126 01:45:33.390692 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Nov 26 01:45:33 crc kubenswrapper[4766]: I1126 01:45:33.483066 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Nov 26 01:45:34 crc kubenswrapper[4766]: I1126 01:45:34.084563 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="1118ff67-6cfc-487c-8887-071283f2cc18" containerName="manila-share" containerID="cri-o://7686ca0421b6dd799426d8a1e3aad0450a845c4e0d9ee7d3056a14e88cee7c54" gracePeriod=30 Nov 26 01:45:34 crc kubenswrapper[4766]: I1126 01:45:34.084675 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="1118ff67-6cfc-487c-8887-071283f2cc18" containerName="probe" containerID="cri-o://e7810811e77fd0d22b61f946b9119b8302e45e6df86a3ca99f2301baa3361aea" gracePeriod=30 Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.098971 4766 generic.go:334] "Generic (PLEG): container finished" podID="1118ff67-6cfc-487c-8887-071283f2cc18" containerID="e7810811e77fd0d22b61f946b9119b8302e45e6df86a3ca99f2301baa3361aea" exitCode=0 Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.099546 4766 generic.go:334] "Generic (PLEG): container finished" podID="1118ff67-6cfc-487c-8887-071283f2cc18" containerID="7686ca0421b6dd799426d8a1e3aad0450a845c4e0d9ee7d3056a14e88cee7c54" exitCode=1 Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.099575 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1118ff67-6cfc-487c-8887-071283f2cc18","Type":"ContainerDied","Data":"e7810811e77fd0d22b61f946b9119b8302e45e6df86a3ca99f2301baa3361aea"} Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.099607 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1118ff67-6cfc-487c-8887-071283f2cc18","Type":"ContainerDied","Data":"7686ca0421b6dd799426d8a1e3aad0450a845c4e0d9ee7d3056a14e88cee7c54"} Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.099623 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1118ff67-6cfc-487c-8887-071283f2cc18","Type":"ContainerDied","Data":"fcb98840cac5e4bdee2fa0111c4aa99dd4e411325a8c8829f0042f0cc4cc32ca"} Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.099636 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fcb98840cac5e4bdee2fa0111c4aa99dd4e411325a8c8829f0042f0cc4cc32ca" Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.174770 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.331926 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-config-data\") pod \"1118ff67-6cfc-487c-8887-071283f2cc18\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.332003 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7z8h\" (UniqueName: \"kubernetes.io/projected/1118ff67-6cfc-487c-8887-071283f2cc18-kube-api-access-v7z8h\") pod \"1118ff67-6cfc-487c-8887-071283f2cc18\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.332101 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1118ff67-6cfc-487c-8887-071283f2cc18-etc-machine-id\") pod \"1118ff67-6cfc-487c-8887-071283f2cc18\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.332141 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-scripts\") pod \"1118ff67-6cfc-487c-8887-071283f2cc18\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.332267 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/1118ff67-6cfc-487c-8887-071283f2cc18-var-lib-manila\") pod \"1118ff67-6cfc-487c-8887-071283f2cc18\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.332339 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1118ff67-6cfc-487c-8887-071283f2cc18-ceph\") pod \"1118ff67-6cfc-487c-8887-071283f2cc18\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.332388 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-combined-ca-bundle\") pod \"1118ff67-6cfc-487c-8887-071283f2cc18\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.332440 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-config-data-custom\") pod \"1118ff67-6cfc-487c-8887-071283f2cc18\" (UID: \"1118ff67-6cfc-487c-8887-071283f2cc18\") " Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.334718 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1118ff67-6cfc-487c-8887-071283f2cc18-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "1118ff67-6cfc-487c-8887-071283f2cc18" (UID: "1118ff67-6cfc-487c-8887-071283f2cc18"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.335443 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1118ff67-6cfc-487c-8887-071283f2cc18-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1118ff67-6cfc-487c-8887-071283f2cc18" (UID: "1118ff67-6cfc-487c-8887-071283f2cc18"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.338813 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1118ff67-6cfc-487c-8887-071283f2cc18-ceph" (OuterVolumeSpecName: "ceph") pod "1118ff67-6cfc-487c-8887-071283f2cc18" (UID: "1118ff67-6cfc-487c-8887-071283f2cc18"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.339455 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1118ff67-6cfc-487c-8887-071283f2cc18-kube-api-access-v7z8h" (OuterVolumeSpecName: "kube-api-access-v7z8h") pod "1118ff67-6cfc-487c-8887-071283f2cc18" (UID: "1118ff67-6cfc-487c-8887-071283f2cc18"). InnerVolumeSpecName "kube-api-access-v7z8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.349469 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-scripts" (OuterVolumeSpecName: "scripts") pod "1118ff67-6cfc-487c-8887-071283f2cc18" (UID: "1118ff67-6cfc-487c-8887-071283f2cc18"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.350239 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1118ff67-6cfc-487c-8887-071283f2cc18" (UID: "1118ff67-6cfc-487c-8887-071283f2cc18"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.436052 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.436420 4766 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/1118ff67-6cfc-487c-8887-071283f2cc18-var-lib-manila\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.436508 4766 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1118ff67-6cfc-487c-8887-071283f2cc18-ceph\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.436583 4766 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.436673 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7z8h\" (UniqueName: \"kubernetes.io/projected/1118ff67-6cfc-487c-8887-071283f2cc18-kube-api-access-v7z8h\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.436751 4766 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1118ff67-6cfc-487c-8887-071283f2cc18-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.442832 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1118ff67-6cfc-487c-8887-071283f2cc18" (UID: "1118ff67-6cfc-487c-8887-071283f2cc18"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.454496 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-config-data" (OuterVolumeSpecName: "config-data") pod "1118ff67-6cfc-487c-8887-071283f2cc18" (UID: "1118ff67-6cfc-487c-8887-071283f2cc18"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.539587 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:35 crc kubenswrapper[4766]: I1126 01:45:35.539642 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1118ff67-6cfc-487c-8887-071283f2cc18-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.113861 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.157981 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.172558 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.202277 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Nov 26 01:45:36 crc kubenswrapper[4766]: E1126 01:45:36.203245 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1118ff67-6cfc-487c-8887-071283f2cc18" containerName="probe" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.203283 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="1118ff67-6cfc-487c-8887-071283f2cc18" containerName="probe" Nov 26 01:45:36 crc kubenswrapper[4766]: E1126 01:45:36.203347 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1118ff67-6cfc-487c-8887-071283f2cc18" containerName="manila-share" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.203361 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="1118ff67-6cfc-487c-8887-071283f2cc18" containerName="manila-share" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.203840 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="1118ff67-6cfc-487c-8887-071283f2cc18" containerName="probe" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.203905 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="1118ff67-6cfc-487c-8887-071283f2cc18" containerName="manila-share" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.206369 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.208441 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.222140 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.359123 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2schg\" (UniqueName: \"kubernetes.io/projected/91ee515c-0d11-436f-b0c9-184f065d52eb-kube-api-access-2schg\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.359182 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/91ee515c-0d11-436f-b0c9-184f065d52eb-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.359213 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91ee515c-0d11-436f-b0c9-184f065d52eb-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.359310 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91ee515c-0d11-436f-b0c9-184f065d52eb-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.359489 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91ee515c-0d11-436f-b0c9-184f065d52eb-scripts\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.359531 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/91ee515c-0d11-436f-b0c9-184f065d52eb-ceph\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.359570 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/91ee515c-0d11-436f-b0c9-184f065d52eb-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.359646 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91ee515c-0d11-436f-b0c9-184f065d52eb-config-data\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.461553 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91ee515c-0d11-436f-b0c9-184f065d52eb-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.461751 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91ee515c-0d11-436f-b0c9-184f065d52eb-scripts\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.461795 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/91ee515c-0d11-436f-b0c9-184f065d52eb-ceph\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.461833 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/91ee515c-0d11-436f-b0c9-184f065d52eb-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.461904 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91ee515c-0d11-436f-b0c9-184f065d52eb-config-data\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.461952 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2schg\" (UniqueName: \"kubernetes.io/projected/91ee515c-0d11-436f-b0c9-184f065d52eb-kube-api-access-2schg\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.462015 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/91ee515c-0d11-436f-b0c9-184f065d52eb-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.462037 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91ee515c-0d11-436f-b0c9-184f065d52eb-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.462530 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/91ee515c-0d11-436f-b0c9-184f065d52eb-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.462973 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/91ee515c-0d11-436f-b0c9-184f065d52eb-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.466704 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/91ee515c-0d11-436f-b0c9-184f065d52eb-ceph\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.467537 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91ee515c-0d11-436f-b0c9-184f065d52eb-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.469345 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91ee515c-0d11-436f-b0c9-184f065d52eb-config-data\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.469371 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91ee515c-0d11-436f-b0c9-184f065d52eb-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.477857 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91ee515c-0d11-436f-b0c9-184f065d52eb-scripts\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.490573 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2schg\" (UniqueName: \"kubernetes.io/projected/91ee515c-0d11-436f-b0c9-184f065d52eb-kube-api-access-2schg\") pod \"manila-share-share1-0\" (UID: \"91ee515c-0d11-436f-b0c9-184f065d52eb\") " pod="openstack/manila-share-share1-0" Nov 26 01:45:36 crc kubenswrapper[4766]: I1126 01:45:36.530484 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 26 01:45:37 crc kubenswrapper[4766]: I1126 01:45:37.196329 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 26 01:45:37 crc kubenswrapper[4766]: I1126 01:45:37.309117 4766 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-65bd4d46f6-blndr" podUID="9e2311f3-2074-4336-ba7c-dc676f60d452" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.71:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.71:8443: connect: connection refused" Nov 26 01:45:37 crc kubenswrapper[4766]: I1126 01:45:37.311337 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:45:37 crc kubenswrapper[4766]: I1126 01:45:37.525614 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Nov 26 01:45:37 crc kubenswrapper[4766]: I1126 01:45:37.841043 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1118ff67-6cfc-487c-8887-071283f2cc18" path="/var/lib/kubelet/pods/1118ff67-6cfc-487c-8887-071283f2cc18/volumes" Nov 26 01:45:38 crc kubenswrapper[4766]: I1126 01:45:38.143981 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"91ee515c-0d11-436f-b0c9-184f065d52eb","Type":"ContainerStarted","Data":"81a9a6234636f16f573e2b2178ef33b4e00233a559d0b3fff09a0833d1379d34"} Nov 26 01:45:38 crc kubenswrapper[4766]: I1126 01:45:38.144024 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"91ee515c-0d11-436f-b0c9-184f065d52eb","Type":"ContainerStarted","Data":"ff288ab1c24bf8ff351496b77d8ad5b6b3fbdec84055b4085d594742f027cd06"} Nov 26 01:45:39 crc kubenswrapper[4766]: I1126 01:45:39.186867 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"91ee515c-0d11-436f-b0c9-184f065d52eb","Type":"ContainerStarted","Data":"3bd137007b79b579ff64f8b315e1bdfc66ba7ab508b06673d5f65d389013d66e"} Nov 26 01:45:39 crc kubenswrapper[4766]: I1126 01:45:39.216786 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.216643879 podStartE2EDuration="3.216643879s" podCreationTimestamp="2025-11-26 01:45:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 01:45:39.208424297 +0000 UTC m=+4920.057194727" watchObservedRunningTime="2025-11-26 01:45:39.216643879 +0000 UTC m=+4920.065414309" Nov 26 01:45:43 crc kubenswrapper[4766]: I1126 01:45:43.248619 4766 generic.go:334] "Generic (PLEG): container finished" podID="9e2311f3-2074-4336-ba7c-dc676f60d452" containerID="331dfc0a6e5abb0e07ca1ea14a78e5f63c07bd07abbf90a2b2dc0c61ad816a52" exitCode=137 Nov 26 01:45:43 crc kubenswrapper[4766]: I1126 01:45:43.248764 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65bd4d46f6-blndr" event={"ID":"9e2311f3-2074-4336-ba7c-dc676f60d452","Type":"ContainerDied","Data":"331dfc0a6e5abb0e07ca1ea14a78e5f63c07bd07abbf90a2b2dc0c61ad816a52"} Nov 26 01:45:43 crc kubenswrapper[4766]: I1126 01:45:43.249298 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65bd4d46f6-blndr" event={"ID":"9e2311f3-2074-4336-ba7c-dc676f60d452","Type":"ContainerDied","Data":"bbd4e6d9f2a4fe590fe2863099e55e529e07a462d9b22eb61a56a22301cc3930"} Nov 26 01:45:43 crc kubenswrapper[4766]: I1126 01:45:43.249315 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbd4e6d9f2a4fe590fe2863099e55e529e07a462d9b22eb61a56a22301cc3930" Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.198608 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.262091 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65bd4d46f6-blndr" Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.362456 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e2311f3-2074-4336-ba7c-dc676f60d452-config-data\") pod \"9e2311f3-2074-4336-ba7c-dc676f60d452\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.362570 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9e2311f3-2074-4336-ba7c-dc676f60d452-horizon-secret-key\") pod \"9e2311f3-2074-4336-ba7c-dc676f60d452\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.362609 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e2311f3-2074-4336-ba7c-dc676f60d452-scripts\") pod \"9e2311f3-2074-4336-ba7c-dc676f60d452\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.362715 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e2311f3-2074-4336-ba7c-dc676f60d452-horizon-tls-certs\") pod \"9e2311f3-2074-4336-ba7c-dc676f60d452\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.362791 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pbrs\" (UniqueName: \"kubernetes.io/projected/9e2311f3-2074-4336-ba7c-dc676f60d452-kube-api-access-9pbrs\") pod \"9e2311f3-2074-4336-ba7c-dc676f60d452\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.362830 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e2311f3-2074-4336-ba7c-dc676f60d452-combined-ca-bundle\") pod \"9e2311f3-2074-4336-ba7c-dc676f60d452\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.362876 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e2311f3-2074-4336-ba7c-dc676f60d452-logs\") pod \"9e2311f3-2074-4336-ba7c-dc676f60d452\" (UID: \"9e2311f3-2074-4336-ba7c-dc676f60d452\") " Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.363927 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e2311f3-2074-4336-ba7c-dc676f60d452-logs" (OuterVolumeSpecName: "logs") pod "9e2311f3-2074-4336-ba7c-dc676f60d452" (UID: "9e2311f3-2074-4336-ba7c-dc676f60d452"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.368632 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e2311f3-2074-4336-ba7c-dc676f60d452-kube-api-access-9pbrs" (OuterVolumeSpecName: "kube-api-access-9pbrs") pod "9e2311f3-2074-4336-ba7c-dc676f60d452" (UID: "9e2311f3-2074-4336-ba7c-dc676f60d452"). InnerVolumeSpecName "kube-api-access-9pbrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.372776 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e2311f3-2074-4336-ba7c-dc676f60d452-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "9e2311f3-2074-4336-ba7c-dc676f60d452" (UID: "9e2311f3-2074-4336-ba7c-dc676f60d452"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.393065 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e2311f3-2074-4336-ba7c-dc676f60d452-config-data" (OuterVolumeSpecName: "config-data") pod "9e2311f3-2074-4336-ba7c-dc676f60d452" (UID: "9e2311f3-2074-4336-ba7c-dc676f60d452"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.399928 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e2311f3-2074-4336-ba7c-dc676f60d452-scripts" (OuterVolumeSpecName: "scripts") pod "9e2311f3-2074-4336-ba7c-dc676f60d452" (UID: "9e2311f3-2074-4336-ba7c-dc676f60d452"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.406512 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e2311f3-2074-4336-ba7c-dc676f60d452-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e2311f3-2074-4336-ba7c-dc676f60d452" (UID: "9e2311f3-2074-4336-ba7c-dc676f60d452"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.423872 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e2311f3-2074-4336-ba7c-dc676f60d452-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "9e2311f3-2074-4336-ba7c-dc676f60d452" (UID: "9e2311f3-2074-4336-ba7c-dc676f60d452"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.465610 4766 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e2311f3-2074-4336-ba7c-dc676f60d452-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.465676 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pbrs\" (UniqueName: \"kubernetes.io/projected/9e2311f3-2074-4336-ba7c-dc676f60d452-kube-api-access-9pbrs\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.465692 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e2311f3-2074-4336-ba7c-dc676f60d452-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.465704 4766 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e2311f3-2074-4336-ba7c-dc676f60d452-logs\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.465716 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e2311f3-2074-4336-ba7c-dc676f60d452-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.465727 4766 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9e2311f3-2074-4336-ba7c-dc676f60d452-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.465738 4766 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e2311f3-2074-4336-ba7c-dc676f60d452-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.599345 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-65bd4d46f6-blndr"] Nov 26 01:45:44 crc kubenswrapper[4766]: I1126 01:45:44.617985 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-65bd4d46f6-blndr"] Nov 26 01:45:45 crc kubenswrapper[4766]: I1126 01:45:45.845783 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e2311f3-2074-4336-ba7c-dc676f60d452" path="/var/lib/kubelet/pods/9e2311f3-2074-4336-ba7c-dc676f60d452/volumes" Nov 26 01:45:46 crc kubenswrapper[4766]: I1126 01:45:46.531292 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Nov 26 01:45:49 crc kubenswrapper[4766]: I1126 01:45:49.102826 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Nov 26 01:45:55 crc kubenswrapper[4766]: I1126 01:45:55.393936 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 26 01:45:58 crc kubenswrapper[4766]: I1126 01:45:58.133206 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Nov 26 01:46:03 crc kubenswrapper[4766]: I1126 01:46:03.813925 4766 scope.go:117] "RemoveContainer" containerID="f6ac10ea28c25a969ab68c19eb363ff24c8ccc6bdd1e6a40648eca0ae7b16f5a" Nov 26 01:46:03 crc kubenswrapper[4766]: I1126 01:46:03.922722 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j4qzz"] Nov 26 01:46:03 crc kubenswrapper[4766]: E1126 01:46:03.923898 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e2311f3-2074-4336-ba7c-dc676f60d452" containerName="horizon" Nov 26 01:46:03 crc kubenswrapper[4766]: I1126 01:46:03.923920 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e2311f3-2074-4336-ba7c-dc676f60d452" containerName="horizon" Nov 26 01:46:03 crc kubenswrapper[4766]: E1126 01:46:03.923931 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e2311f3-2074-4336-ba7c-dc676f60d452" containerName="horizon-log" Nov 26 01:46:03 crc kubenswrapper[4766]: I1126 01:46:03.923939 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e2311f3-2074-4336-ba7c-dc676f60d452" containerName="horizon-log" Nov 26 01:46:03 crc kubenswrapper[4766]: I1126 01:46:03.924218 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e2311f3-2074-4336-ba7c-dc676f60d452" containerName="horizon-log" Nov 26 01:46:03 crc kubenswrapper[4766]: I1126 01:46:03.924253 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e2311f3-2074-4336-ba7c-dc676f60d452" containerName="horizon" Nov 26 01:46:03 crc kubenswrapper[4766]: I1126 01:46:03.926318 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j4qzz" Nov 26 01:46:03 crc kubenswrapper[4766]: I1126 01:46:03.959148 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j4qzz"] Nov 26 01:46:04 crc kubenswrapper[4766]: I1126 01:46:04.112096 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2808a50-1bd6-4360-9560-d7e47db12aa8-utilities\") pod \"redhat-operators-j4qzz\" (UID: \"c2808a50-1bd6-4360-9560-d7e47db12aa8\") " pod="openshift-marketplace/redhat-operators-j4qzz" Nov 26 01:46:04 crc kubenswrapper[4766]: I1126 01:46:04.112230 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2808a50-1bd6-4360-9560-d7e47db12aa8-catalog-content\") pod \"redhat-operators-j4qzz\" (UID: \"c2808a50-1bd6-4360-9560-d7e47db12aa8\") " pod="openshift-marketplace/redhat-operators-j4qzz" Nov 26 01:46:04 crc kubenswrapper[4766]: I1126 01:46:04.112352 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmww5\" (UniqueName: \"kubernetes.io/projected/c2808a50-1bd6-4360-9560-d7e47db12aa8-kube-api-access-tmww5\") pod \"redhat-operators-j4qzz\" (UID: \"c2808a50-1bd6-4360-9560-d7e47db12aa8\") " pod="openshift-marketplace/redhat-operators-j4qzz" Nov 26 01:46:04 crc kubenswrapper[4766]: I1126 01:46:04.215189 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmww5\" (UniqueName: \"kubernetes.io/projected/c2808a50-1bd6-4360-9560-d7e47db12aa8-kube-api-access-tmww5\") pod \"redhat-operators-j4qzz\" (UID: \"c2808a50-1bd6-4360-9560-d7e47db12aa8\") " pod="openshift-marketplace/redhat-operators-j4qzz" Nov 26 01:46:04 crc kubenswrapper[4766]: I1126 01:46:04.215400 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2808a50-1bd6-4360-9560-d7e47db12aa8-utilities\") pod \"redhat-operators-j4qzz\" (UID: \"c2808a50-1bd6-4360-9560-d7e47db12aa8\") " pod="openshift-marketplace/redhat-operators-j4qzz" Nov 26 01:46:04 crc kubenswrapper[4766]: I1126 01:46:04.215514 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2808a50-1bd6-4360-9560-d7e47db12aa8-catalog-content\") pod \"redhat-operators-j4qzz\" (UID: \"c2808a50-1bd6-4360-9560-d7e47db12aa8\") " pod="openshift-marketplace/redhat-operators-j4qzz" Nov 26 01:46:04 crc kubenswrapper[4766]: I1126 01:46:04.216320 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2808a50-1bd6-4360-9560-d7e47db12aa8-catalog-content\") pod \"redhat-operators-j4qzz\" (UID: \"c2808a50-1bd6-4360-9560-d7e47db12aa8\") " pod="openshift-marketplace/redhat-operators-j4qzz" Nov 26 01:46:04 crc kubenswrapper[4766]: I1126 01:46:04.216397 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2808a50-1bd6-4360-9560-d7e47db12aa8-utilities\") pod \"redhat-operators-j4qzz\" (UID: \"c2808a50-1bd6-4360-9560-d7e47db12aa8\") " pod="openshift-marketplace/redhat-operators-j4qzz" Nov 26 01:46:04 crc kubenswrapper[4766]: I1126 01:46:04.255912 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmww5\" (UniqueName: \"kubernetes.io/projected/c2808a50-1bd6-4360-9560-d7e47db12aa8-kube-api-access-tmww5\") pod \"redhat-operators-j4qzz\" (UID: \"c2808a50-1bd6-4360-9560-d7e47db12aa8\") " pod="openshift-marketplace/redhat-operators-j4qzz" Nov 26 01:46:04 crc kubenswrapper[4766]: I1126 01:46:04.546017 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j4qzz" Nov 26 01:46:05 crc kubenswrapper[4766]: I1126 01:46:05.070565 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j4qzz"] Nov 26 01:46:05 crc kubenswrapper[4766]: I1126 01:46:05.526312 4766 generic.go:334] "Generic (PLEG): container finished" podID="c2808a50-1bd6-4360-9560-d7e47db12aa8" containerID="5aa460e8de205abeff79730d04583ca06f06d130cfc10b5d7c30dc1be6ee7b25" exitCode=0 Nov 26 01:46:05 crc kubenswrapper[4766]: I1126 01:46:05.526433 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j4qzz" event={"ID":"c2808a50-1bd6-4360-9560-d7e47db12aa8","Type":"ContainerDied","Data":"5aa460e8de205abeff79730d04583ca06f06d130cfc10b5d7c30dc1be6ee7b25"} Nov 26 01:46:05 crc kubenswrapper[4766]: I1126 01:46:05.526635 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j4qzz" event={"ID":"c2808a50-1bd6-4360-9560-d7e47db12aa8","Type":"ContainerStarted","Data":"f1d0cbff0bedefcb6fd3e2ae62eb8f3770d41646add79cbd6de0ab6c13f67cba"} Nov 26 01:46:05 crc kubenswrapper[4766]: I1126 01:46:05.529350 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 01:46:06 crc kubenswrapper[4766]: I1126 01:46:06.550573 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j4qzz" event={"ID":"c2808a50-1bd6-4360-9560-d7e47db12aa8","Type":"ContainerStarted","Data":"ec5b0f3334e82b39187ffbbedda68563ca551ed7606912562e91a5e5ead58cba"} Nov 26 01:46:10 crc kubenswrapper[4766]: I1126 01:46:10.628770 4766 generic.go:334] "Generic (PLEG): container finished" podID="c2808a50-1bd6-4360-9560-d7e47db12aa8" containerID="ec5b0f3334e82b39187ffbbedda68563ca551ed7606912562e91a5e5ead58cba" exitCode=0 Nov 26 01:46:10 crc kubenswrapper[4766]: I1126 01:46:10.628938 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j4qzz" event={"ID":"c2808a50-1bd6-4360-9560-d7e47db12aa8","Type":"ContainerDied","Data":"ec5b0f3334e82b39187ffbbedda68563ca551ed7606912562e91a5e5ead58cba"} Nov 26 01:46:11 crc kubenswrapper[4766]: I1126 01:46:11.479463 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:46:11 crc kubenswrapper[4766]: I1126 01:46:11.479907 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:46:11 crc kubenswrapper[4766]: I1126 01:46:11.645856 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j4qzz" event={"ID":"c2808a50-1bd6-4360-9560-d7e47db12aa8","Type":"ContainerStarted","Data":"4f293f895e606643cedda195c1e0ac2eb83a386ca54cfbf3b05250fc243c51bb"} Nov 26 01:46:11 crc kubenswrapper[4766]: I1126 01:46:11.681130 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j4qzz" podStartSLOduration=3.18583739 podStartE2EDuration="8.681101782s" podCreationTimestamp="2025-11-26 01:46:03 +0000 UTC" firstStartedPulling="2025-11-26 01:46:05.529067979 +0000 UTC m=+4946.377838419" lastFinishedPulling="2025-11-26 01:46:11.024332351 +0000 UTC m=+4951.873102811" observedRunningTime="2025-11-26 01:46:11.663461599 +0000 UTC m=+4952.512232049" watchObservedRunningTime="2025-11-26 01:46:11.681101782 +0000 UTC m=+4952.529872212" Nov 26 01:46:14 crc kubenswrapper[4766]: I1126 01:46:14.546813 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j4qzz" Nov 26 01:46:14 crc kubenswrapper[4766]: I1126 01:46:14.548048 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j4qzz" Nov 26 01:46:15 crc kubenswrapper[4766]: I1126 01:46:15.619574 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j4qzz" podUID="c2808a50-1bd6-4360-9560-d7e47db12aa8" containerName="registry-server" probeResult="failure" output=< Nov 26 01:46:15 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 01:46:15 crc kubenswrapper[4766]: > Nov 26 01:46:25 crc kubenswrapper[4766]: I1126 01:46:25.604038 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j4qzz" podUID="c2808a50-1bd6-4360-9560-d7e47db12aa8" containerName="registry-server" probeResult="failure" output=< Nov 26 01:46:25 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 01:46:25 crc kubenswrapper[4766]: > Nov 26 01:46:30 crc kubenswrapper[4766]: I1126 01:46:30.804144 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tphd6"] Nov 26 01:46:30 crc kubenswrapper[4766]: I1126 01:46:30.810378 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tphd6" Nov 26 01:46:30 crc kubenswrapper[4766]: I1126 01:46:30.841130 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tphd6"] Nov 26 01:46:30 crc kubenswrapper[4766]: I1126 01:46:30.920413 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c10f8078-4645-4fe6-bd6e-3af086b70475-catalog-content\") pod \"redhat-marketplace-tphd6\" (UID: \"c10f8078-4645-4fe6-bd6e-3af086b70475\") " pod="openshift-marketplace/redhat-marketplace-tphd6" Nov 26 01:46:30 crc kubenswrapper[4766]: I1126 01:46:30.920497 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zbc2\" (UniqueName: \"kubernetes.io/projected/c10f8078-4645-4fe6-bd6e-3af086b70475-kube-api-access-6zbc2\") pod \"redhat-marketplace-tphd6\" (UID: \"c10f8078-4645-4fe6-bd6e-3af086b70475\") " pod="openshift-marketplace/redhat-marketplace-tphd6" Nov 26 01:46:30 crc kubenswrapper[4766]: I1126 01:46:30.921010 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c10f8078-4645-4fe6-bd6e-3af086b70475-utilities\") pod \"redhat-marketplace-tphd6\" (UID: \"c10f8078-4645-4fe6-bd6e-3af086b70475\") " pod="openshift-marketplace/redhat-marketplace-tphd6" Nov 26 01:46:31 crc kubenswrapper[4766]: I1126 01:46:31.023723 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c10f8078-4645-4fe6-bd6e-3af086b70475-catalog-content\") pod \"redhat-marketplace-tphd6\" (UID: \"c10f8078-4645-4fe6-bd6e-3af086b70475\") " pod="openshift-marketplace/redhat-marketplace-tphd6" Nov 26 01:46:31 crc kubenswrapper[4766]: I1126 01:46:31.023818 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zbc2\" (UniqueName: \"kubernetes.io/projected/c10f8078-4645-4fe6-bd6e-3af086b70475-kube-api-access-6zbc2\") pod \"redhat-marketplace-tphd6\" (UID: \"c10f8078-4645-4fe6-bd6e-3af086b70475\") " pod="openshift-marketplace/redhat-marketplace-tphd6" Nov 26 01:46:31 crc kubenswrapper[4766]: I1126 01:46:31.023977 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c10f8078-4645-4fe6-bd6e-3af086b70475-utilities\") pod \"redhat-marketplace-tphd6\" (UID: \"c10f8078-4645-4fe6-bd6e-3af086b70475\") " pod="openshift-marketplace/redhat-marketplace-tphd6" Nov 26 01:46:31 crc kubenswrapper[4766]: I1126 01:46:31.024442 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c10f8078-4645-4fe6-bd6e-3af086b70475-catalog-content\") pod \"redhat-marketplace-tphd6\" (UID: \"c10f8078-4645-4fe6-bd6e-3af086b70475\") " pod="openshift-marketplace/redhat-marketplace-tphd6" Nov 26 01:46:31 crc kubenswrapper[4766]: I1126 01:46:31.024489 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c10f8078-4645-4fe6-bd6e-3af086b70475-utilities\") pod \"redhat-marketplace-tphd6\" (UID: \"c10f8078-4645-4fe6-bd6e-3af086b70475\") " pod="openshift-marketplace/redhat-marketplace-tphd6" Nov 26 01:46:31 crc kubenswrapper[4766]: I1126 01:46:31.114486 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zbc2\" (UniqueName: \"kubernetes.io/projected/c10f8078-4645-4fe6-bd6e-3af086b70475-kube-api-access-6zbc2\") pod \"redhat-marketplace-tphd6\" (UID: \"c10f8078-4645-4fe6-bd6e-3af086b70475\") " pod="openshift-marketplace/redhat-marketplace-tphd6" Nov 26 01:46:31 crc kubenswrapper[4766]: I1126 01:46:31.151375 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tphd6" Nov 26 01:46:31 crc kubenswrapper[4766]: I1126 01:46:31.584341 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tphd6"] Nov 26 01:46:32 crc kubenswrapper[4766]: I1126 01:46:32.077999 4766 generic.go:334] "Generic (PLEG): container finished" podID="c10f8078-4645-4fe6-bd6e-3af086b70475" containerID="c8af5c47bddef3df6040b1eded3ed56965322725b011b9cc116b605842c61dde" exitCode=0 Nov 26 01:46:32 crc kubenswrapper[4766]: I1126 01:46:32.078121 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tphd6" event={"ID":"c10f8078-4645-4fe6-bd6e-3af086b70475","Type":"ContainerDied","Data":"c8af5c47bddef3df6040b1eded3ed56965322725b011b9cc116b605842c61dde"} Nov 26 01:46:32 crc kubenswrapper[4766]: I1126 01:46:32.078242 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tphd6" event={"ID":"c10f8078-4645-4fe6-bd6e-3af086b70475","Type":"ContainerStarted","Data":"7ccb5bbde69597f1b0718aecea86c2b42fdad968813258798d132fcd615e16d4"} Nov 26 01:46:34 crc kubenswrapper[4766]: I1126 01:46:34.106155 4766 generic.go:334] "Generic (PLEG): container finished" podID="c10f8078-4645-4fe6-bd6e-3af086b70475" containerID="f787d2c027d868dfd4c18eae64245a57bd114add6f7c61b4e3726f2e63161ec8" exitCode=0 Nov 26 01:46:34 crc kubenswrapper[4766]: I1126 01:46:34.106294 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tphd6" event={"ID":"c10f8078-4645-4fe6-bd6e-3af086b70475","Type":"ContainerDied","Data":"f787d2c027d868dfd4c18eae64245a57bd114add6f7c61b4e3726f2e63161ec8"} Nov 26 01:46:34 crc kubenswrapper[4766]: I1126 01:46:34.607296 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j4qzz" Nov 26 01:46:34 crc kubenswrapper[4766]: I1126 01:46:34.677254 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j4qzz" Nov 26 01:46:35 crc kubenswrapper[4766]: I1126 01:46:35.117267 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tphd6" event={"ID":"c10f8078-4645-4fe6-bd6e-3af086b70475","Type":"ContainerStarted","Data":"f7caf5fea554cc5e1b89788484b02156d5727c4afc7f238feb33f4adb2cbb413"} Nov 26 01:46:35 crc kubenswrapper[4766]: I1126 01:46:35.149146 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tphd6" podStartSLOduration=2.712436674 podStartE2EDuration="5.149123817s" podCreationTimestamp="2025-11-26 01:46:30 +0000 UTC" firstStartedPulling="2025-11-26 01:46:32.081897857 +0000 UTC m=+4972.930668277" lastFinishedPulling="2025-11-26 01:46:34.51858497 +0000 UTC m=+4975.367355420" observedRunningTime="2025-11-26 01:46:35.138825374 +0000 UTC m=+4975.987595814" watchObservedRunningTime="2025-11-26 01:46:35.149123817 +0000 UTC m=+4975.997894247" Nov 26 01:46:35 crc kubenswrapper[4766]: I1126 01:46:35.927081 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j4qzz"] Nov 26 01:46:36 crc kubenswrapper[4766]: I1126 01:46:36.129208 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j4qzz" podUID="c2808a50-1bd6-4360-9560-d7e47db12aa8" containerName="registry-server" containerID="cri-o://4f293f895e606643cedda195c1e0ac2eb83a386ca54cfbf3b05250fc243c51bb" gracePeriod=2 Nov 26 01:46:36 crc kubenswrapper[4766]: I1126 01:46:36.773316 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j4qzz" Nov 26 01:46:36 crc kubenswrapper[4766]: I1126 01:46:36.873460 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2808a50-1bd6-4360-9560-d7e47db12aa8-catalog-content\") pod \"c2808a50-1bd6-4360-9560-d7e47db12aa8\" (UID: \"c2808a50-1bd6-4360-9560-d7e47db12aa8\") " Nov 26 01:46:36 crc kubenswrapper[4766]: I1126 01:46:36.873621 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmww5\" (UniqueName: \"kubernetes.io/projected/c2808a50-1bd6-4360-9560-d7e47db12aa8-kube-api-access-tmww5\") pod \"c2808a50-1bd6-4360-9560-d7e47db12aa8\" (UID: \"c2808a50-1bd6-4360-9560-d7e47db12aa8\") " Nov 26 01:46:36 crc kubenswrapper[4766]: I1126 01:46:36.873766 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2808a50-1bd6-4360-9560-d7e47db12aa8-utilities\") pod \"c2808a50-1bd6-4360-9560-d7e47db12aa8\" (UID: \"c2808a50-1bd6-4360-9560-d7e47db12aa8\") " Nov 26 01:46:36 crc kubenswrapper[4766]: I1126 01:46:36.875098 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2808a50-1bd6-4360-9560-d7e47db12aa8-utilities" (OuterVolumeSpecName: "utilities") pod "c2808a50-1bd6-4360-9560-d7e47db12aa8" (UID: "c2808a50-1bd6-4360-9560-d7e47db12aa8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:46:36 crc kubenswrapper[4766]: I1126 01:46:36.876205 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2808a50-1bd6-4360-9560-d7e47db12aa8-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 01:46:36 crc kubenswrapper[4766]: I1126 01:46:36.881532 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2808a50-1bd6-4360-9560-d7e47db12aa8-kube-api-access-tmww5" (OuterVolumeSpecName: "kube-api-access-tmww5") pod "c2808a50-1bd6-4360-9560-d7e47db12aa8" (UID: "c2808a50-1bd6-4360-9560-d7e47db12aa8"). InnerVolumeSpecName "kube-api-access-tmww5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:46:36 crc kubenswrapper[4766]: I1126 01:46:36.977093 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2808a50-1bd6-4360-9560-d7e47db12aa8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c2808a50-1bd6-4360-9560-d7e47db12aa8" (UID: "c2808a50-1bd6-4360-9560-d7e47db12aa8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:46:36 crc kubenswrapper[4766]: I1126 01:46:36.978579 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2808a50-1bd6-4360-9560-d7e47db12aa8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 01:46:36 crc kubenswrapper[4766]: I1126 01:46:36.978600 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmww5\" (UniqueName: \"kubernetes.io/projected/c2808a50-1bd6-4360-9560-d7e47db12aa8-kube-api-access-tmww5\") on node \"crc\" DevicePath \"\"" Nov 26 01:46:37 crc kubenswrapper[4766]: I1126 01:46:37.146294 4766 generic.go:334] "Generic (PLEG): container finished" podID="c2808a50-1bd6-4360-9560-d7e47db12aa8" containerID="4f293f895e606643cedda195c1e0ac2eb83a386ca54cfbf3b05250fc243c51bb" exitCode=0 Nov 26 01:46:37 crc kubenswrapper[4766]: I1126 01:46:37.146318 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j4qzz" Nov 26 01:46:37 crc kubenswrapper[4766]: I1126 01:46:37.146353 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j4qzz" event={"ID":"c2808a50-1bd6-4360-9560-d7e47db12aa8","Type":"ContainerDied","Data":"4f293f895e606643cedda195c1e0ac2eb83a386ca54cfbf3b05250fc243c51bb"} Nov 26 01:46:37 crc kubenswrapper[4766]: I1126 01:46:37.148354 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j4qzz" event={"ID":"c2808a50-1bd6-4360-9560-d7e47db12aa8","Type":"ContainerDied","Data":"f1d0cbff0bedefcb6fd3e2ae62eb8f3770d41646add79cbd6de0ab6c13f67cba"} Nov 26 01:46:37 crc kubenswrapper[4766]: I1126 01:46:37.148403 4766 scope.go:117] "RemoveContainer" containerID="4f293f895e606643cedda195c1e0ac2eb83a386ca54cfbf3b05250fc243c51bb" Nov 26 01:46:37 crc kubenswrapper[4766]: I1126 01:46:37.201907 4766 scope.go:117] "RemoveContainer" containerID="ec5b0f3334e82b39187ffbbedda68563ca551ed7606912562e91a5e5ead58cba" Nov 26 01:46:37 crc kubenswrapper[4766]: I1126 01:46:37.215985 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j4qzz"] Nov 26 01:46:37 crc kubenswrapper[4766]: I1126 01:46:37.234237 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j4qzz"] Nov 26 01:46:37 crc kubenswrapper[4766]: I1126 01:46:37.272358 4766 scope.go:117] "RemoveContainer" containerID="5aa460e8de205abeff79730d04583ca06f06d130cfc10b5d7c30dc1be6ee7b25" Nov 26 01:46:37 crc kubenswrapper[4766]: I1126 01:46:37.312733 4766 scope.go:117] "RemoveContainer" containerID="4f293f895e606643cedda195c1e0ac2eb83a386ca54cfbf3b05250fc243c51bb" Nov 26 01:46:37 crc kubenswrapper[4766]: E1126 01:46:37.313298 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f293f895e606643cedda195c1e0ac2eb83a386ca54cfbf3b05250fc243c51bb\": container with ID starting with 4f293f895e606643cedda195c1e0ac2eb83a386ca54cfbf3b05250fc243c51bb not found: ID does not exist" containerID="4f293f895e606643cedda195c1e0ac2eb83a386ca54cfbf3b05250fc243c51bb" Nov 26 01:46:37 crc kubenswrapper[4766]: I1126 01:46:37.313358 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f293f895e606643cedda195c1e0ac2eb83a386ca54cfbf3b05250fc243c51bb"} err="failed to get container status \"4f293f895e606643cedda195c1e0ac2eb83a386ca54cfbf3b05250fc243c51bb\": rpc error: code = NotFound desc = could not find container \"4f293f895e606643cedda195c1e0ac2eb83a386ca54cfbf3b05250fc243c51bb\": container with ID starting with 4f293f895e606643cedda195c1e0ac2eb83a386ca54cfbf3b05250fc243c51bb not found: ID does not exist" Nov 26 01:46:37 crc kubenswrapper[4766]: I1126 01:46:37.313404 4766 scope.go:117] "RemoveContainer" containerID="ec5b0f3334e82b39187ffbbedda68563ca551ed7606912562e91a5e5ead58cba" Nov 26 01:46:37 crc kubenswrapper[4766]: E1126 01:46:37.313902 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec5b0f3334e82b39187ffbbedda68563ca551ed7606912562e91a5e5ead58cba\": container with ID starting with ec5b0f3334e82b39187ffbbedda68563ca551ed7606912562e91a5e5ead58cba not found: ID does not exist" containerID="ec5b0f3334e82b39187ffbbedda68563ca551ed7606912562e91a5e5ead58cba" Nov 26 01:46:37 crc kubenswrapper[4766]: I1126 01:46:37.313932 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec5b0f3334e82b39187ffbbedda68563ca551ed7606912562e91a5e5ead58cba"} err="failed to get container status \"ec5b0f3334e82b39187ffbbedda68563ca551ed7606912562e91a5e5ead58cba\": rpc error: code = NotFound desc = could not find container \"ec5b0f3334e82b39187ffbbedda68563ca551ed7606912562e91a5e5ead58cba\": container with ID starting with ec5b0f3334e82b39187ffbbedda68563ca551ed7606912562e91a5e5ead58cba not found: ID does not exist" Nov 26 01:46:37 crc kubenswrapper[4766]: I1126 01:46:37.313953 4766 scope.go:117] "RemoveContainer" containerID="5aa460e8de205abeff79730d04583ca06f06d130cfc10b5d7c30dc1be6ee7b25" Nov 26 01:46:37 crc kubenswrapper[4766]: E1126 01:46:37.314233 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5aa460e8de205abeff79730d04583ca06f06d130cfc10b5d7c30dc1be6ee7b25\": container with ID starting with 5aa460e8de205abeff79730d04583ca06f06d130cfc10b5d7c30dc1be6ee7b25 not found: ID does not exist" containerID="5aa460e8de205abeff79730d04583ca06f06d130cfc10b5d7c30dc1be6ee7b25" Nov 26 01:46:37 crc kubenswrapper[4766]: I1126 01:46:37.314301 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5aa460e8de205abeff79730d04583ca06f06d130cfc10b5d7c30dc1be6ee7b25"} err="failed to get container status \"5aa460e8de205abeff79730d04583ca06f06d130cfc10b5d7c30dc1be6ee7b25\": rpc error: code = NotFound desc = could not find container \"5aa460e8de205abeff79730d04583ca06f06d130cfc10b5d7c30dc1be6ee7b25\": container with ID starting with 5aa460e8de205abeff79730d04583ca06f06d130cfc10b5d7c30dc1be6ee7b25 not found: ID does not exist" Nov 26 01:46:37 crc kubenswrapper[4766]: I1126 01:46:37.861126 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2808a50-1bd6-4360-9560-d7e47db12aa8" path="/var/lib/kubelet/pods/c2808a50-1bd6-4360-9560-d7e47db12aa8/volumes" Nov 26 01:46:41 crc kubenswrapper[4766]: I1126 01:46:41.151849 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tphd6" Nov 26 01:46:41 crc kubenswrapper[4766]: I1126 01:46:41.152597 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tphd6" Nov 26 01:46:41 crc kubenswrapper[4766]: I1126 01:46:41.230157 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tphd6" Nov 26 01:46:41 crc kubenswrapper[4766]: I1126 01:46:41.312503 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tphd6" Nov 26 01:46:41 crc kubenswrapper[4766]: I1126 01:46:41.479889 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:46:41 crc kubenswrapper[4766]: I1126 01:46:41.480433 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:46:41 crc kubenswrapper[4766]: I1126 01:46:41.486240 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tphd6"] Nov 26 01:46:43 crc kubenswrapper[4766]: I1126 01:46:43.239432 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tphd6" podUID="c10f8078-4645-4fe6-bd6e-3af086b70475" containerName="registry-server" containerID="cri-o://f7caf5fea554cc5e1b89788484b02156d5727c4afc7f238feb33f4adb2cbb413" gracePeriod=2 Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.175706 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tphd6" Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.287488 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c10f8078-4645-4fe6-bd6e-3af086b70475-catalog-content\") pod \"c10f8078-4645-4fe6-bd6e-3af086b70475\" (UID: \"c10f8078-4645-4fe6-bd6e-3af086b70475\") " Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.287720 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zbc2\" (UniqueName: \"kubernetes.io/projected/c10f8078-4645-4fe6-bd6e-3af086b70475-kube-api-access-6zbc2\") pod \"c10f8078-4645-4fe6-bd6e-3af086b70475\" (UID: \"c10f8078-4645-4fe6-bd6e-3af086b70475\") " Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.287791 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c10f8078-4645-4fe6-bd6e-3af086b70475-utilities\") pod \"c10f8078-4645-4fe6-bd6e-3af086b70475\" (UID: \"c10f8078-4645-4fe6-bd6e-3af086b70475\") " Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.288993 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c10f8078-4645-4fe6-bd6e-3af086b70475-utilities" (OuterVolumeSpecName: "utilities") pod "c10f8078-4645-4fe6-bd6e-3af086b70475" (UID: "c10f8078-4645-4fe6-bd6e-3af086b70475"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.292952 4766 generic.go:334] "Generic (PLEG): container finished" podID="c10f8078-4645-4fe6-bd6e-3af086b70475" containerID="f7caf5fea554cc5e1b89788484b02156d5727c4afc7f238feb33f4adb2cbb413" exitCode=0 Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.292987 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tphd6" event={"ID":"c10f8078-4645-4fe6-bd6e-3af086b70475","Type":"ContainerDied","Data":"f7caf5fea554cc5e1b89788484b02156d5727c4afc7f238feb33f4adb2cbb413"} Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.293016 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tphd6" Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.293032 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tphd6" event={"ID":"c10f8078-4645-4fe6-bd6e-3af086b70475","Type":"ContainerDied","Data":"7ccb5bbde69597f1b0718aecea86c2b42fdad968813258798d132fcd615e16d4"} Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.293052 4766 scope.go:117] "RemoveContainer" containerID="f7caf5fea554cc5e1b89788484b02156d5727c4afc7f238feb33f4adb2cbb413" Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.304930 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c10f8078-4645-4fe6-bd6e-3af086b70475-kube-api-access-6zbc2" (OuterVolumeSpecName: "kube-api-access-6zbc2") pod "c10f8078-4645-4fe6-bd6e-3af086b70475" (UID: "c10f8078-4645-4fe6-bd6e-3af086b70475"). InnerVolumeSpecName "kube-api-access-6zbc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.318030 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c10f8078-4645-4fe6-bd6e-3af086b70475-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c10f8078-4645-4fe6-bd6e-3af086b70475" (UID: "c10f8078-4645-4fe6-bd6e-3af086b70475"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.390978 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zbc2\" (UniqueName: \"kubernetes.io/projected/c10f8078-4645-4fe6-bd6e-3af086b70475-kube-api-access-6zbc2\") on node \"crc\" DevicePath \"\"" Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.391012 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c10f8078-4645-4fe6-bd6e-3af086b70475-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.391025 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c10f8078-4645-4fe6-bd6e-3af086b70475-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.391782 4766 scope.go:117] "RemoveContainer" containerID="f787d2c027d868dfd4c18eae64245a57bd114add6f7c61b4e3726f2e63161ec8" Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.428621 4766 scope.go:117] "RemoveContainer" containerID="c8af5c47bddef3df6040b1eded3ed56965322725b011b9cc116b605842c61dde" Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.468365 4766 scope.go:117] "RemoveContainer" containerID="f7caf5fea554cc5e1b89788484b02156d5727c4afc7f238feb33f4adb2cbb413" Nov 26 01:46:44 crc kubenswrapper[4766]: E1126 01:46:44.469074 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7caf5fea554cc5e1b89788484b02156d5727c4afc7f238feb33f4adb2cbb413\": container with ID starting with f7caf5fea554cc5e1b89788484b02156d5727c4afc7f238feb33f4adb2cbb413 not found: ID does not exist" containerID="f7caf5fea554cc5e1b89788484b02156d5727c4afc7f238feb33f4adb2cbb413" Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.469135 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7caf5fea554cc5e1b89788484b02156d5727c4afc7f238feb33f4adb2cbb413"} err="failed to get container status \"f7caf5fea554cc5e1b89788484b02156d5727c4afc7f238feb33f4adb2cbb413\": rpc error: code = NotFound desc = could not find container \"f7caf5fea554cc5e1b89788484b02156d5727c4afc7f238feb33f4adb2cbb413\": container with ID starting with f7caf5fea554cc5e1b89788484b02156d5727c4afc7f238feb33f4adb2cbb413 not found: ID does not exist" Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.469167 4766 scope.go:117] "RemoveContainer" containerID="f787d2c027d868dfd4c18eae64245a57bd114add6f7c61b4e3726f2e63161ec8" Nov 26 01:46:44 crc kubenswrapper[4766]: E1126 01:46:44.469994 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f787d2c027d868dfd4c18eae64245a57bd114add6f7c61b4e3726f2e63161ec8\": container with ID starting with f787d2c027d868dfd4c18eae64245a57bd114add6f7c61b4e3726f2e63161ec8 not found: ID does not exist" containerID="f787d2c027d868dfd4c18eae64245a57bd114add6f7c61b4e3726f2e63161ec8" Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.470051 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f787d2c027d868dfd4c18eae64245a57bd114add6f7c61b4e3726f2e63161ec8"} err="failed to get container status \"f787d2c027d868dfd4c18eae64245a57bd114add6f7c61b4e3726f2e63161ec8\": rpc error: code = NotFound desc = could not find container \"f787d2c027d868dfd4c18eae64245a57bd114add6f7c61b4e3726f2e63161ec8\": container with ID starting with f787d2c027d868dfd4c18eae64245a57bd114add6f7c61b4e3726f2e63161ec8 not found: ID does not exist" Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.470086 4766 scope.go:117] "RemoveContainer" containerID="c8af5c47bddef3df6040b1eded3ed56965322725b011b9cc116b605842c61dde" Nov 26 01:46:44 crc kubenswrapper[4766]: E1126 01:46:44.470613 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8af5c47bddef3df6040b1eded3ed56965322725b011b9cc116b605842c61dde\": container with ID starting with c8af5c47bddef3df6040b1eded3ed56965322725b011b9cc116b605842c61dde not found: ID does not exist" containerID="c8af5c47bddef3df6040b1eded3ed56965322725b011b9cc116b605842c61dde" Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.470670 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8af5c47bddef3df6040b1eded3ed56965322725b011b9cc116b605842c61dde"} err="failed to get container status \"c8af5c47bddef3df6040b1eded3ed56965322725b011b9cc116b605842c61dde\": rpc error: code = NotFound desc = could not find container \"c8af5c47bddef3df6040b1eded3ed56965322725b011b9cc116b605842c61dde\": container with ID starting with c8af5c47bddef3df6040b1eded3ed56965322725b011b9cc116b605842c61dde not found: ID does not exist" Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.645047 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tphd6"] Nov 26 01:46:44 crc kubenswrapper[4766]: I1126 01:46:44.675724 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tphd6"] Nov 26 01:46:45 crc kubenswrapper[4766]: I1126 01:46:45.850412 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c10f8078-4645-4fe6-bd6e-3af086b70475" path="/var/lib/kubelet/pods/c10f8078-4645-4fe6-bd6e-3af086b70475/volumes" Nov 26 01:47:11 crc kubenswrapper[4766]: I1126 01:47:11.479960 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:47:11 crc kubenswrapper[4766]: I1126 01:47:11.480411 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:47:11 crc kubenswrapper[4766]: I1126 01:47:11.480451 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 01:47:11 crc kubenswrapper[4766]: I1126 01:47:11.480993 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 01:47:11 crc kubenswrapper[4766]: I1126 01:47:11.481036 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" gracePeriod=600 Nov 26 01:47:11 crc kubenswrapper[4766]: E1126 01:47:11.610018 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:47:11 crc kubenswrapper[4766]: I1126 01:47:11.711508 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" exitCode=0 Nov 26 01:47:11 crc kubenswrapper[4766]: I1126 01:47:11.711564 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479"} Nov 26 01:47:11 crc kubenswrapper[4766]: I1126 01:47:11.712009 4766 scope.go:117] "RemoveContainer" containerID="e9fdc469b39326ae18e21b32546bbc3ad05efbc565ef046911377215b98ca5e8" Nov 26 01:47:11 crc kubenswrapper[4766]: I1126 01:47:11.712941 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:47:11 crc kubenswrapper[4766]: E1126 01:47:11.713251 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:47:23 crc kubenswrapper[4766]: I1126 01:47:23.829018 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:47:23 crc kubenswrapper[4766]: E1126 01:47:23.830303 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:47:38 crc kubenswrapper[4766]: I1126 01:47:38.826389 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:47:38 crc kubenswrapper[4766]: E1126 01:47:38.827314 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:47:53 crc kubenswrapper[4766]: I1126 01:47:53.827761 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:47:53 crc kubenswrapper[4766]: E1126 01:47:53.828965 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:48:05 crc kubenswrapper[4766]: I1126 01:48:05.827574 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:48:05 crc kubenswrapper[4766]: E1126 01:48:05.828454 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:48:16 crc kubenswrapper[4766]: I1126 01:48:16.829699 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:48:16 crc kubenswrapper[4766]: E1126 01:48:16.830916 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:48:30 crc kubenswrapper[4766]: I1126 01:48:30.827903 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:48:30 crc kubenswrapper[4766]: E1126 01:48:30.829354 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:48:44 crc kubenswrapper[4766]: I1126 01:48:44.826639 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:48:44 crc kubenswrapper[4766]: E1126 01:48:44.827494 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:48:58 crc kubenswrapper[4766]: I1126 01:48:58.827517 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:48:58 crc kubenswrapper[4766]: E1126 01:48:58.829677 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:49:13 crc kubenswrapper[4766]: I1126 01:49:13.827551 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:49:13 crc kubenswrapper[4766]: E1126 01:49:13.828408 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:49:24 crc kubenswrapper[4766]: I1126 01:49:24.827082 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:49:24 crc kubenswrapper[4766]: E1126 01:49:24.827825 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:49:36 crc kubenswrapper[4766]: I1126 01:49:36.828224 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:49:36 crc kubenswrapper[4766]: E1126 01:49:36.831137 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:49:47 crc kubenswrapper[4766]: I1126 01:49:47.826430 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:49:47 crc kubenswrapper[4766]: E1126 01:49:47.827230 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:50:00 crc kubenswrapper[4766]: I1126 01:50:00.826696 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:50:00 crc kubenswrapper[4766]: E1126 01:50:00.827426 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:50:12 crc kubenswrapper[4766]: I1126 01:50:12.830329 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:50:12 crc kubenswrapper[4766]: E1126 01:50:12.831367 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:50:27 crc kubenswrapper[4766]: I1126 01:50:27.829368 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:50:27 crc kubenswrapper[4766]: E1126 01:50:27.830132 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:50:38 crc kubenswrapper[4766]: I1126 01:50:38.828067 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:50:38 crc kubenswrapper[4766]: E1126 01:50:38.829274 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:50:53 crc kubenswrapper[4766]: I1126 01:50:53.829050 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:50:53 crc kubenswrapper[4766]: E1126 01:50:53.829829 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:51:04 crc kubenswrapper[4766]: I1126 01:51:04.233618 4766 scope.go:117] "RemoveContainer" containerID="425ff6f4884d714d97e857b8513f474a1d431613a4dc5063d13ebf02dcb8adad" Nov 26 01:51:04 crc kubenswrapper[4766]: I1126 01:51:04.457314 4766 scope.go:117] "RemoveContainer" containerID="331dfc0a6e5abb0e07ca1ea14a78e5f63c07bd07abbf90a2b2dc0c61ad816a52" Nov 26 01:51:07 crc kubenswrapper[4766]: I1126 01:51:07.830314 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:51:07 crc kubenswrapper[4766]: E1126 01:51:07.831011 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:51:19 crc kubenswrapper[4766]: I1126 01:51:19.835479 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:51:19 crc kubenswrapper[4766]: E1126 01:51:19.837947 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:51:30 crc kubenswrapper[4766]: I1126 01:51:30.828293 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:51:30 crc kubenswrapper[4766]: E1126 01:51:30.829534 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:51:32 crc kubenswrapper[4766]: I1126 01:51:32.274470 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-z2q6r"] Nov 26 01:51:32 crc kubenswrapper[4766]: E1126 01:51:32.275379 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2808a50-1bd6-4360-9560-d7e47db12aa8" containerName="extract-content" Nov 26 01:51:32 crc kubenswrapper[4766]: I1126 01:51:32.275405 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2808a50-1bd6-4360-9560-d7e47db12aa8" containerName="extract-content" Nov 26 01:51:32 crc kubenswrapper[4766]: E1126 01:51:32.275464 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c10f8078-4645-4fe6-bd6e-3af086b70475" containerName="extract-content" Nov 26 01:51:32 crc kubenswrapper[4766]: I1126 01:51:32.275477 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c10f8078-4645-4fe6-bd6e-3af086b70475" containerName="extract-content" Nov 26 01:51:32 crc kubenswrapper[4766]: E1126 01:51:32.275498 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2808a50-1bd6-4360-9560-d7e47db12aa8" containerName="extract-utilities" Nov 26 01:51:32 crc kubenswrapper[4766]: I1126 01:51:32.275512 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2808a50-1bd6-4360-9560-d7e47db12aa8" containerName="extract-utilities" Nov 26 01:51:32 crc kubenswrapper[4766]: E1126 01:51:32.275536 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2808a50-1bd6-4360-9560-d7e47db12aa8" containerName="registry-server" Nov 26 01:51:32 crc kubenswrapper[4766]: I1126 01:51:32.275548 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2808a50-1bd6-4360-9560-d7e47db12aa8" containerName="registry-server" Nov 26 01:51:32 crc kubenswrapper[4766]: E1126 01:51:32.275570 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c10f8078-4645-4fe6-bd6e-3af086b70475" containerName="registry-server" Nov 26 01:51:32 crc kubenswrapper[4766]: I1126 01:51:32.275582 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c10f8078-4645-4fe6-bd6e-3af086b70475" containerName="registry-server" Nov 26 01:51:32 crc kubenswrapper[4766]: E1126 01:51:32.275607 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c10f8078-4645-4fe6-bd6e-3af086b70475" containerName="extract-utilities" Nov 26 01:51:32 crc kubenswrapper[4766]: I1126 01:51:32.275621 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c10f8078-4645-4fe6-bd6e-3af086b70475" containerName="extract-utilities" Nov 26 01:51:32 crc kubenswrapper[4766]: I1126 01:51:32.276045 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2808a50-1bd6-4360-9560-d7e47db12aa8" containerName="registry-server" Nov 26 01:51:32 crc kubenswrapper[4766]: I1126 01:51:32.276081 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="c10f8078-4645-4fe6-bd6e-3af086b70475" containerName="registry-server" Nov 26 01:51:32 crc kubenswrapper[4766]: I1126 01:51:32.278862 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2q6r" Nov 26 01:51:32 crc kubenswrapper[4766]: I1126 01:51:32.294390 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z2q6r"] Nov 26 01:51:32 crc kubenswrapper[4766]: I1126 01:51:32.416677 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6e0ae73-b8bd-4462-95cc-6203fbba9b30-catalog-content\") pod \"certified-operators-z2q6r\" (UID: \"b6e0ae73-b8bd-4462-95cc-6203fbba9b30\") " pod="openshift-marketplace/certified-operators-z2q6r" Nov 26 01:51:32 crc kubenswrapper[4766]: I1126 01:51:32.416954 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwhxb\" (UniqueName: \"kubernetes.io/projected/b6e0ae73-b8bd-4462-95cc-6203fbba9b30-kube-api-access-xwhxb\") pod \"certified-operators-z2q6r\" (UID: \"b6e0ae73-b8bd-4462-95cc-6203fbba9b30\") " pod="openshift-marketplace/certified-operators-z2q6r" Nov 26 01:51:32 crc kubenswrapper[4766]: I1126 01:51:32.417070 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6e0ae73-b8bd-4462-95cc-6203fbba9b30-utilities\") pod \"certified-operators-z2q6r\" (UID: \"b6e0ae73-b8bd-4462-95cc-6203fbba9b30\") " pod="openshift-marketplace/certified-operators-z2q6r" Nov 26 01:51:32 crc kubenswrapper[4766]: I1126 01:51:32.519761 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6e0ae73-b8bd-4462-95cc-6203fbba9b30-utilities\") pod \"certified-operators-z2q6r\" (UID: \"b6e0ae73-b8bd-4462-95cc-6203fbba9b30\") " pod="openshift-marketplace/certified-operators-z2q6r" Nov 26 01:51:32 crc kubenswrapper[4766]: I1126 01:51:32.519983 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6e0ae73-b8bd-4462-95cc-6203fbba9b30-catalog-content\") pod \"certified-operators-z2q6r\" (UID: \"b6e0ae73-b8bd-4462-95cc-6203fbba9b30\") " pod="openshift-marketplace/certified-operators-z2q6r" Nov 26 01:51:32 crc kubenswrapper[4766]: I1126 01:51:32.520171 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwhxb\" (UniqueName: \"kubernetes.io/projected/b6e0ae73-b8bd-4462-95cc-6203fbba9b30-kube-api-access-xwhxb\") pod \"certified-operators-z2q6r\" (UID: \"b6e0ae73-b8bd-4462-95cc-6203fbba9b30\") " pod="openshift-marketplace/certified-operators-z2q6r" Nov 26 01:51:32 crc kubenswrapper[4766]: I1126 01:51:32.520632 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6e0ae73-b8bd-4462-95cc-6203fbba9b30-utilities\") pod \"certified-operators-z2q6r\" (UID: \"b6e0ae73-b8bd-4462-95cc-6203fbba9b30\") " pod="openshift-marketplace/certified-operators-z2q6r" Nov 26 01:51:32 crc kubenswrapper[4766]: I1126 01:51:32.520632 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6e0ae73-b8bd-4462-95cc-6203fbba9b30-catalog-content\") pod \"certified-operators-z2q6r\" (UID: \"b6e0ae73-b8bd-4462-95cc-6203fbba9b30\") " pod="openshift-marketplace/certified-operators-z2q6r" Nov 26 01:51:32 crc kubenswrapper[4766]: I1126 01:51:32.545481 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwhxb\" (UniqueName: \"kubernetes.io/projected/b6e0ae73-b8bd-4462-95cc-6203fbba9b30-kube-api-access-xwhxb\") pod \"certified-operators-z2q6r\" (UID: \"b6e0ae73-b8bd-4462-95cc-6203fbba9b30\") " pod="openshift-marketplace/certified-operators-z2q6r" Nov 26 01:51:32 crc kubenswrapper[4766]: I1126 01:51:32.636831 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2q6r" Nov 26 01:51:33 crc kubenswrapper[4766]: I1126 01:51:33.176158 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z2q6r"] Nov 26 01:51:33 crc kubenswrapper[4766]: W1126 01:51:33.719951 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6e0ae73_b8bd_4462_95cc_6203fbba9b30.slice/crio-5af2d90576f724e288c831455a2550ede4073aba8ad0d32c427909ef39957fbc WatchSource:0}: Error finding container 5af2d90576f724e288c831455a2550ede4073aba8ad0d32c427909ef39957fbc: Status 404 returned error can't find the container with id 5af2d90576f724e288c831455a2550ede4073aba8ad0d32c427909ef39957fbc Nov 26 01:51:34 crc kubenswrapper[4766]: I1126 01:51:34.536390 4766 generic.go:334] "Generic (PLEG): container finished" podID="b6e0ae73-b8bd-4462-95cc-6203fbba9b30" containerID="da9a01636b88d259172bc51f5abe3672b8d352d2e31895862b02a95996574f11" exitCode=0 Nov 26 01:51:34 crc kubenswrapper[4766]: I1126 01:51:34.536486 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2q6r" event={"ID":"b6e0ae73-b8bd-4462-95cc-6203fbba9b30","Type":"ContainerDied","Data":"da9a01636b88d259172bc51f5abe3672b8d352d2e31895862b02a95996574f11"} Nov 26 01:51:34 crc kubenswrapper[4766]: I1126 01:51:34.536823 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2q6r" event={"ID":"b6e0ae73-b8bd-4462-95cc-6203fbba9b30","Type":"ContainerStarted","Data":"5af2d90576f724e288c831455a2550ede4073aba8ad0d32c427909ef39957fbc"} Nov 26 01:51:34 crc kubenswrapper[4766]: I1126 01:51:34.553138 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 01:51:36 crc kubenswrapper[4766]: I1126 01:51:36.571670 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2q6r" event={"ID":"b6e0ae73-b8bd-4462-95cc-6203fbba9b30","Type":"ContainerStarted","Data":"7f967c06955440151342c25d6ed756685105abf5f31b0cdd7c04fe49e67f5731"} Nov 26 01:51:38 crc kubenswrapper[4766]: I1126 01:51:38.602787 4766 generic.go:334] "Generic (PLEG): container finished" podID="b6e0ae73-b8bd-4462-95cc-6203fbba9b30" containerID="7f967c06955440151342c25d6ed756685105abf5f31b0cdd7c04fe49e67f5731" exitCode=0 Nov 26 01:51:38 crc kubenswrapper[4766]: I1126 01:51:38.602887 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2q6r" event={"ID":"b6e0ae73-b8bd-4462-95cc-6203fbba9b30","Type":"ContainerDied","Data":"7f967c06955440151342c25d6ed756685105abf5f31b0cdd7c04fe49e67f5731"} Nov 26 01:51:39 crc kubenswrapper[4766]: I1126 01:51:39.619839 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2q6r" event={"ID":"b6e0ae73-b8bd-4462-95cc-6203fbba9b30","Type":"ContainerStarted","Data":"82a0bc6309190fe35504ed582a6a2ea208dca89796471cb1feff1b1f22473db1"} Nov 26 01:51:39 crc kubenswrapper[4766]: I1126 01:51:39.649151 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-z2q6r" podStartSLOduration=3.164315909 podStartE2EDuration="7.649124886s" podCreationTimestamp="2025-11-26 01:51:32 +0000 UTC" firstStartedPulling="2025-11-26 01:51:34.552394124 +0000 UTC m=+5275.401164594" lastFinishedPulling="2025-11-26 01:51:39.037203131 +0000 UTC m=+5279.885973571" observedRunningTime="2025-11-26 01:51:39.645105308 +0000 UTC m=+5280.493875738" watchObservedRunningTime="2025-11-26 01:51:39.649124886 +0000 UTC m=+5280.497895316" Nov 26 01:51:42 crc kubenswrapper[4766]: I1126 01:51:42.637671 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-z2q6r" Nov 26 01:51:42 crc kubenswrapper[4766]: I1126 01:51:42.638296 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-z2q6r" Nov 26 01:51:42 crc kubenswrapper[4766]: I1126 01:51:42.723805 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-z2q6r" Nov 26 01:51:42 crc kubenswrapper[4766]: I1126 01:51:42.827177 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:51:42 crc kubenswrapper[4766]: E1126 01:51:42.827728 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:51:52 crc kubenswrapper[4766]: I1126 01:51:52.711003 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-z2q6r" Nov 26 01:51:52 crc kubenswrapper[4766]: I1126 01:51:52.883775 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kdnd8"] Nov 26 01:51:52 crc kubenswrapper[4766]: I1126 01:51:52.887011 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kdnd8" Nov 26 01:51:52 crc kubenswrapper[4766]: I1126 01:51:52.899089 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kdnd8"] Nov 26 01:51:52 crc kubenswrapper[4766]: I1126 01:51:52.948558 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e807a4b-80e7-4c70-b58a-89a2fcae7905-utilities\") pod \"community-operators-kdnd8\" (UID: \"7e807a4b-80e7-4c70-b58a-89a2fcae7905\") " pod="openshift-marketplace/community-operators-kdnd8" Nov 26 01:51:52 crc kubenswrapper[4766]: I1126 01:51:52.948627 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e807a4b-80e7-4c70-b58a-89a2fcae7905-catalog-content\") pod \"community-operators-kdnd8\" (UID: \"7e807a4b-80e7-4c70-b58a-89a2fcae7905\") " pod="openshift-marketplace/community-operators-kdnd8" Nov 26 01:51:52 crc kubenswrapper[4766]: I1126 01:51:52.948843 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckrrq\" (UniqueName: \"kubernetes.io/projected/7e807a4b-80e7-4c70-b58a-89a2fcae7905-kube-api-access-ckrrq\") pod \"community-operators-kdnd8\" (UID: \"7e807a4b-80e7-4c70-b58a-89a2fcae7905\") " pod="openshift-marketplace/community-operators-kdnd8" Nov 26 01:51:53 crc kubenswrapper[4766]: I1126 01:51:53.050786 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e807a4b-80e7-4c70-b58a-89a2fcae7905-utilities\") pod \"community-operators-kdnd8\" (UID: \"7e807a4b-80e7-4c70-b58a-89a2fcae7905\") " pod="openshift-marketplace/community-operators-kdnd8" Nov 26 01:51:53 crc kubenswrapper[4766]: I1126 01:51:53.050842 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e807a4b-80e7-4c70-b58a-89a2fcae7905-catalog-content\") pod \"community-operators-kdnd8\" (UID: \"7e807a4b-80e7-4c70-b58a-89a2fcae7905\") " pod="openshift-marketplace/community-operators-kdnd8" Nov 26 01:51:53 crc kubenswrapper[4766]: I1126 01:51:53.050914 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckrrq\" (UniqueName: \"kubernetes.io/projected/7e807a4b-80e7-4c70-b58a-89a2fcae7905-kube-api-access-ckrrq\") pod \"community-operators-kdnd8\" (UID: \"7e807a4b-80e7-4c70-b58a-89a2fcae7905\") " pod="openshift-marketplace/community-operators-kdnd8" Nov 26 01:51:53 crc kubenswrapper[4766]: I1126 01:51:53.051355 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e807a4b-80e7-4c70-b58a-89a2fcae7905-utilities\") pod \"community-operators-kdnd8\" (UID: \"7e807a4b-80e7-4c70-b58a-89a2fcae7905\") " pod="openshift-marketplace/community-operators-kdnd8" Nov 26 01:51:53 crc kubenswrapper[4766]: I1126 01:51:53.051735 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e807a4b-80e7-4c70-b58a-89a2fcae7905-catalog-content\") pod \"community-operators-kdnd8\" (UID: \"7e807a4b-80e7-4c70-b58a-89a2fcae7905\") " pod="openshift-marketplace/community-operators-kdnd8" Nov 26 01:51:53 crc kubenswrapper[4766]: I1126 01:51:53.074831 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckrrq\" (UniqueName: \"kubernetes.io/projected/7e807a4b-80e7-4c70-b58a-89a2fcae7905-kube-api-access-ckrrq\") pod \"community-operators-kdnd8\" (UID: \"7e807a4b-80e7-4c70-b58a-89a2fcae7905\") " pod="openshift-marketplace/community-operators-kdnd8" Nov 26 01:51:53 crc kubenswrapper[4766]: I1126 01:51:53.226995 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kdnd8" Nov 26 01:51:53 crc kubenswrapper[4766]: W1126 01:51:53.823628 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e807a4b_80e7_4c70_b58a_89a2fcae7905.slice/crio-9001c3bdf65d58df3c4b9afc334a2e84d940035229ccf849a5db21dc5cdfdf49 WatchSource:0}: Error finding container 9001c3bdf65d58df3c4b9afc334a2e84d940035229ccf849a5db21dc5cdfdf49: Status 404 returned error can't find the container with id 9001c3bdf65d58df3c4b9afc334a2e84d940035229ccf849a5db21dc5cdfdf49 Nov 26 01:51:53 crc kubenswrapper[4766]: I1126 01:51:53.860982 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kdnd8"] Nov 26 01:51:53 crc kubenswrapper[4766]: I1126 01:51:53.867022 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kdnd8" event={"ID":"7e807a4b-80e7-4c70-b58a-89a2fcae7905","Type":"ContainerStarted","Data":"9001c3bdf65d58df3c4b9afc334a2e84d940035229ccf849a5db21dc5cdfdf49"} Nov 26 01:51:54 crc kubenswrapper[4766]: I1126 01:51:54.880877 4766 generic.go:334] "Generic (PLEG): container finished" podID="7e807a4b-80e7-4c70-b58a-89a2fcae7905" containerID="8df3371a51cb7d0e9e59d7a38c432adf6fdbe76adb3ced528985524bd56ceb60" exitCode=0 Nov 26 01:51:54 crc kubenswrapper[4766]: I1126 01:51:54.880924 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kdnd8" event={"ID":"7e807a4b-80e7-4c70-b58a-89a2fcae7905","Type":"ContainerDied","Data":"8df3371a51cb7d0e9e59d7a38c432adf6fdbe76adb3ced528985524bd56ceb60"} Nov 26 01:51:55 crc kubenswrapper[4766]: I1126 01:51:55.457339 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z2q6r"] Nov 26 01:51:55 crc kubenswrapper[4766]: I1126 01:51:55.457872 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-z2q6r" podUID="b6e0ae73-b8bd-4462-95cc-6203fbba9b30" containerName="registry-server" containerID="cri-o://82a0bc6309190fe35504ed582a6a2ea208dca89796471cb1feff1b1f22473db1" gracePeriod=2 Nov 26 01:51:55 crc kubenswrapper[4766]: I1126 01:51:55.827445 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:51:55 crc kubenswrapper[4766]: E1126 01:51:55.828235 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:51:55 crc kubenswrapper[4766]: I1126 01:51:55.897988 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kdnd8" event={"ID":"7e807a4b-80e7-4c70-b58a-89a2fcae7905","Type":"ContainerStarted","Data":"1b040489d578f423d280201e8f40a5a9f97fb8e5063e3086b3ca18ffe1f4adff"} Nov 26 01:51:55 crc kubenswrapper[4766]: I1126 01:51:55.902509 4766 generic.go:334] "Generic (PLEG): container finished" podID="b6e0ae73-b8bd-4462-95cc-6203fbba9b30" containerID="82a0bc6309190fe35504ed582a6a2ea208dca89796471cb1feff1b1f22473db1" exitCode=0 Nov 26 01:51:55 crc kubenswrapper[4766]: I1126 01:51:55.902550 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2q6r" event={"ID":"b6e0ae73-b8bd-4462-95cc-6203fbba9b30","Type":"ContainerDied","Data":"82a0bc6309190fe35504ed582a6a2ea208dca89796471cb1feff1b1f22473db1"} Nov 26 01:51:56 crc kubenswrapper[4766]: I1126 01:51:56.082937 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2q6r" Nov 26 01:51:56 crc kubenswrapper[4766]: I1126 01:51:56.223812 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6e0ae73-b8bd-4462-95cc-6203fbba9b30-catalog-content\") pod \"b6e0ae73-b8bd-4462-95cc-6203fbba9b30\" (UID: \"b6e0ae73-b8bd-4462-95cc-6203fbba9b30\") " Nov 26 01:51:56 crc kubenswrapper[4766]: I1126 01:51:56.223940 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6e0ae73-b8bd-4462-95cc-6203fbba9b30-utilities\") pod \"b6e0ae73-b8bd-4462-95cc-6203fbba9b30\" (UID: \"b6e0ae73-b8bd-4462-95cc-6203fbba9b30\") " Nov 26 01:51:56 crc kubenswrapper[4766]: I1126 01:51:56.224043 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwhxb\" (UniqueName: \"kubernetes.io/projected/b6e0ae73-b8bd-4462-95cc-6203fbba9b30-kube-api-access-xwhxb\") pod \"b6e0ae73-b8bd-4462-95cc-6203fbba9b30\" (UID: \"b6e0ae73-b8bd-4462-95cc-6203fbba9b30\") " Nov 26 01:51:56 crc kubenswrapper[4766]: I1126 01:51:56.226077 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6e0ae73-b8bd-4462-95cc-6203fbba9b30-utilities" (OuterVolumeSpecName: "utilities") pod "b6e0ae73-b8bd-4462-95cc-6203fbba9b30" (UID: "b6e0ae73-b8bd-4462-95cc-6203fbba9b30"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:51:56 crc kubenswrapper[4766]: I1126 01:51:56.236829 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6e0ae73-b8bd-4462-95cc-6203fbba9b30-kube-api-access-xwhxb" (OuterVolumeSpecName: "kube-api-access-xwhxb") pod "b6e0ae73-b8bd-4462-95cc-6203fbba9b30" (UID: "b6e0ae73-b8bd-4462-95cc-6203fbba9b30"). InnerVolumeSpecName "kube-api-access-xwhxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:51:56 crc kubenswrapper[4766]: I1126 01:51:56.293849 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6e0ae73-b8bd-4462-95cc-6203fbba9b30-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b6e0ae73-b8bd-4462-95cc-6203fbba9b30" (UID: "b6e0ae73-b8bd-4462-95cc-6203fbba9b30"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:51:56 crc kubenswrapper[4766]: I1126 01:51:56.328085 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6e0ae73-b8bd-4462-95cc-6203fbba9b30-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 01:51:56 crc kubenswrapper[4766]: I1126 01:51:56.328667 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwhxb\" (UniqueName: \"kubernetes.io/projected/b6e0ae73-b8bd-4462-95cc-6203fbba9b30-kube-api-access-xwhxb\") on node \"crc\" DevicePath \"\"" Nov 26 01:51:56 crc kubenswrapper[4766]: I1126 01:51:56.328715 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6e0ae73-b8bd-4462-95cc-6203fbba9b30-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 01:51:56 crc kubenswrapper[4766]: I1126 01:51:56.922261 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2q6r" Nov 26 01:51:56 crc kubenswrapper[4766]: I1126 01:51:56.922262 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2q6r" event={"ID":"b6e0ae73-b8bd-4462-95cc-6203fbba9b30","Type":"ContainerDied","Data":"5af2d90576f724e288c831455a2550ede4073aba8ad0d32c427909ef39957fbc"} Nov 26 01:51:56 crc kubenswrapper[4766]: I1126 01:51:56.922711 4766 scope.go:117] "RemoveContainer" containerID="82a0bc6309190fe35504ed582a6a2ea208dca89796471cb1feff1b1f22473db1" Nov 26 01:51:56 crc kubenswrapper[4766]: I1126 01:51:56.970091 4766 scope.go:117] "RemoveContainer" containerID="7f967c06955440151342c25d6ed756685105abf5f31b0cdd7c04fe49e67f5731" Nov 26 01:51:56 crc kubenswrapper[4766]: I1126 01:51:56.974764 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z2q6r"] Nov 26 01:51:57 crc kubenswrapper[4766]: I1126 01:51:57.005785 4766 scope.go:117] "RemoveContainer" containerID="da9a01636b88d259172bc51f5abe3672b8d352d2e31895862b02a95996574f11" Nov 26 01:51:57 crc kubenswrapper[4766]: I1126 01:51:57.011061 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-z2q6r"] Nov 26 01:51:57 crc kubenswrapper[4766]: I1126 01:51:57.906834 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6e0ae73-b8bd-4462-95cc-6203fbba9b30" path="/var/lib/kubelet/pods/b6e0ae73-b8bd-4462-95cc-6203fbba9b30/volumes" Nov 26 01:51:58 crc kubenswrapper[4766]: I1126 01:51:58.945193 4766 generic.go:334] "Generic (PLEG): container finished" podID="7e807a4b-80e7-4c70-b58a-89a2fcae7905" containerID="1b040489d578f423d280201e8f40a5a9f97fb8e5063e3086b3ca18ffe1f4adff" exitCode=0 Nov 26 01:51:58 crc kubenswrapper[4766]: I1126 01:51:58.945540 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kdnd8" event={"ID":"7e807a4b-80e7-4c70-b58a-89a2fcae7905","Type":"ContainerDied","Data":"1b040489d578f423d280201e8f40a5a9f97fb8e5063e3086b3ca18ffe1f4adff"} Nov 26 01:52:00 crc kubenswrapper[4766]: I1126 01:52:00.983213 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kdnd8" event={"ID":"7e807a4b-80e7-4c70-b58a-89a2fcae7905","Type":"ContainerStarted","Data":"1130ab213e811941dd6b57cc23d9a5da098408ee7a75dabca8f03fb43b07e7f0"} Nov 26 01:52:01 crc kubenswrapper[4766]: I1126 01:52:01.016720 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kdnd8" podStartSLOduration=3.880332026 podStartE2EDuration="9.016697742s" podCreationTimestamp="2025-11-26 01:51:52 +0000 UTC" firstStartedPulling="2025-11-26 01:51:54.885534721 +0000 UTC m=+5295.734305161" lastFinishedPulling="2025-11-26 01:52:00.021900417 +0000 UTC m=+5300.870670877" observedRunningTime="2025-11-26 01:52:01.013584456 +0000 UTC m=+5301.862354886" watchObservedRunningTime="2025-11-26 01:52:01.016697742 +0000 UTC m=+5301.865468202" Nov 26 01:52:03 crc kubenswrapper[4766]: I1126 01:52:03.227513 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kdnd8" Nov 26 01:52:03 crc kubenswrapper[4766]: I1126 01:52:03.227986 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kdnd8" Nov 26 01:52:03 crc kubenswrapper[4766]: I1126 01:52:03.311439 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kdnd8" Nov 26 01:52:04 crc kubenswrapper[4766]: I1126 01:52:04.525082 4766 scope.go:117] "RemoveContainer" containerID="e7810811e77fd0d22b61f946b9119b8302e45e6df86a3ca99f2301baa3361aea" Nov 26 01:52:04 crc kubenswrapper[4766]: I1126 01:52:04.559679 4766 scope.go:117] "RemoveContainer" containerID="7686ca0421b6dd799426d8a1e3aad0450a845c4e0d9ee7d3056a14e88cee7c54" Nov 26 01:52:08 crc kubenswrapper[4766]: I1126 01:52:08.827069 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:52:08 crc kubenswrapper[4766]: E1126 01:52:08.827969 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:52:13 crc kubenswrapper[4766]: I1126 01:52:13.286932 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kdnd8" Nov 26 01:52:13 crc kubenswrapper[4766]: I1126 01:52:13.350894 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kdnd8"] Nov 26 01:52:14 crc kubenswrapper[4766]: I1126 01:52:14.165771 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kdnd8" podUID="7e807a4b-80e7-4c70-b58a-89a2fcae7905" containerName="registry-server" containerID="cri-o://1130ab213e811941dd6b57cc23d9a5da098408ee7a75dabca8f03fb43b07e7f0" gracePeriod=2 Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.004818 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kdnd8" Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.029172 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e807a4b-80e7-4c70-b58a-89a2fcae7905-utilities\") pod \"7e807a4b-80e7-4c70-b58a-89a2fcae7905\" (UID: \"7e807a4b-80e7-4c70-b58a-89a2fcae7905\") " Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.029563 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e807a4b-80e7-4c70-b58a-89a2fcae7905-catalog-content\") pod \"7e807a4b-80e7-4c70-b58a-89a2fcae7905\" (UID: \"7e807a4b-80e7-4c70-b58a-89a2fcae7905\") " Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.030677 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e807a4b-80e7-4c70-b58a-89a2fcae7905-utilities" (OuterVolumeSpecName: "utilities") pod "7e807a4b-80e7-4c70-b58a-89a2fcae7905" (UID: "7e807a4b-80e7-4c70-b58a-89a2fcae7905"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.031694 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckrrq\" (UniqueName: \"kubernetes.io/projected/7e807a4b-80e7-4c70-b58a-89a2fcae7905-kube-api-access-ckrrq\") pod \"7e807a4b-80e7-4c70-b58a-89a2fcae7905\" (UID: \"7e807a4b-80e7-4c70-b58a-89a2fcae7905\") " Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.036336 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e807a4b-80e7-4c70-b58a-89a2fcae7905-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.039874 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e807a4b-80e7-4c70-b58a-89a2fcae7905-kube-api-access-ckrrq" (OuterVolumeSpecName: "kube-api-access-ckrrq") pod "7e807a4b-80e7-4c70-b58a-89a2fcae7905" (UID: "7e807a4b-80e7-4c70-b58a-89a2fcae7905"). InnerVolumeSpecName "kube-api-access-ckrrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.091548 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e807a4b-80e7-4c70-b58a-89a2fcae7905-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7e807a4b-80e7-4c70-b58a-89a2fcae7905" (UID: "7e807a4b-80e7-4c70-b58a-89a2fcae7905"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.138621 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e807a4b-80e7-4c70-b58a-89a2fcae7905-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.138676 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckrrq\" (UniqueName: \"kubernetes.io/projected/7e807a4b-80e7-4c70-b58a-89a2fcae7905-kube-api-access-ckrrq\") on node \"crc\" DevicePath \"\"" Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.177668 4766 generic.go:334] "Generic (PLEG): container finished" podID="7e807a4b-80e7-4c70-b58a-89a2fcae7905" containerID="1130ab213e811941dd6b57cc23d9a5da098408ee7a75dabca8f03fb43b07e7f0" exitCode=0 Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.177719 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kdnd8" event={"ID":"7e807a4b-80e7-4c70-b58a-89a2fcae7905","Type":"ContainerDied","Data":"1130ab213e811941dd6b57cc23d9a5da098408ee7a75dabca8f03fb43b07e7f0"} Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.177729 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kdnd8" Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.177758 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kdnd8" event={"ID":"7e807a4b-80e7-4c70-b58a-89a2fcae7905","Type":"ContainerDied","Data":"9001c3bdf65d58df3c4b9afc334a2e84d940035229ccf849a5db21dc5cdfdf49"} Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.177782 4766 scope.go:117] "RemoveContainer" containerID="1130ab213e811941dd6b57cc23d9a5da098408ee7a75dabca8f03fb43b07e7f0" Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.201705 4766 scope.go:117] "RemoveContainer" containerID="1b040489d578f423d280201e8f40a5a9f97fb8e5063e3086b3ca18ffe1f4adff" Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.215479 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kdnd8"] Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.225675 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kdnd8"] Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.227737 4766 scope.go:117] "RemoveContainer" containerID="8df3371a51cb7d0e9e59d7a38c432adf6fdbe76adb3ced528985524bd56ceb60" Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.279075 4766 scope.go:117] "RemoveContainer" containerID="1130ab213e811941dd6b57cc23d9a5da098408ee7a75dabca8f03fb43b07e7f0" Nov 26 01:52:15 crc kubenswrapper[4766]: E1126 01:52:15.279721 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1130ab213e811941dd6b57cc23d9a5da098408ee7a75dabca8f03fb43b07e7f0\": container with ID starting with 1130ab213e811941dd6b57cc23d9a5da098408ee7a75dabca8f03fb43b07e7f0 not found: ID does not exist" containerID="1130ab213e811941dd6b57cc23d9a5da098408ee7a75dabca8f03fb43b07e7f0" Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.279762 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1130ab213e811941dd6b57cc23d9a5da098408ee7a75dabca8f03fb43b07e7f0"} err="failed to get container status \"1130ab213e811941dd6b57cc23d9a5da098408ee7a75dabca8f03fb43b07e7f0\": rpc error: code = NotFound desc = could not find container \"1130ab213e811941dd6b57cc23d9a5da098408ee7a75dabca8f03fb43b07e7f0\": container with ID starting with 1130ab213e811941dd6b57cc23d9a5da098408ee7a75dabca8f03fb43b07e7f0 not found: ID does not exist" Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.279787 4766 scope.go:117] "RemoveContainer" containerID="1b040489d578f423d280201e8f40a5a9f97fb8e5063e3086b3ca18ffe1f4adff" Nov 26 01:52:15 crc kubenswrapper[4766]: E1126 01:52:15.280182 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b040489d578f423d280201e8f40a5a9f97fb8e5063e3086b3ca18ffe1f4adff\": container with ID starting with 1b040489d578f423d280201e8f40a5a9f97fb8e5063e3086b3ca18ffe1f4adff not found: ID does not exist" containerID="1b040489d578f423d280201e8f40a5a9f97fb8e5063e3086b3ca18ffe1f4adff" Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.280214 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b040489d578f423d280201e8f40a5a9f97fb8e5063e3086b3ca18ffe1f4adff"} err="failed to get container status \"1b040489d578f423d280201e8f40a5a9f97fb8e5063e3086b3ca18ffe1f4adff\": rpc error: code = NotFound desc = could not find container \"1b040489d578f423d280201e8f40a5a9f97fb8e5063e3086b3ca18ffe1f4adff\": container with ID starting with 1b040489d578f423d280201e8f40a5a9f97fb8e5063e3086b3ca18ffe1f4adff not found: ID does not exist" Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.280252 4766 scope.go:117] "RemoveContainer" containerID="8df3371a51cb7d0e9e59d7a38c432adf6fdbe76adb3ced528985524bd56ceb60" Nov 26 01:52:15 crc kubenswrapper[4766]: E1126 01:52:15.280833 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8df3371a51cb7d0e9e59d7a38c432adf6fdbe76adb3ced528985524bd56ceb60\": container with ID starting with 8df3371a51cb7d0e9e59d7a38c432adf6fdbe76adb3ced528985524bd56ceb60 not found: ID does not exist" containerID="8df3371a51cb7d0e9e59d7a38c432adf6fdbe76adb3ced528985524bd56ceb60" Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.280934 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8df3371a51cb7d0e9e59d7a38c432adf6fdbe76adb3ced528985524bd56ceb60"} err="failed to get container status \"8df3371a51cb7d0e9e59d7a38c432adf6fdbe76adb3ced528985524bd56ceb60\": rpc error: code = NotFound desc = could not find container \"8df3371a51cb7d0e9e59d7a38c432adf6fdbe76adb3ced528985524bd56ceb60\": container with ID starting with 8df3371a51cb7d0e9e59d7a38c432adf6fdbe76adb3ced528985524bd56ceb60 not found: ID does not exist" Nov 26 01:52:15 crc kubenswrapper[4766]: I1126 01:52:15.847572 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e807a4b-80e7-4c70-b58a-89a2fcae7905" path="/var/lib/kubelet/pods/7e807a4b-80e7-4c70-b58a-89a2fcae7905/volumes" Nov 26 01:52:23 crc kubenswrapper[4766]: I1126 01:52:23.828165 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:52:24 crc kubenswrapper[4766]: I1126 01:52:24.300723 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"8cb92b56012b319b19fabfd8594272dd6949e442ab3db81280faac01a61052b7"} Nov 26 01:52:28 crc kubenswrapper[4766]: E1126 01:52:28.929511 4766 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.32:50560->38.102.83.32:36993: write tcp 38.102.83.32:50560->38.102.83.32:36993: write: broken pipe Nov 26 01:54:29 crc kubenswrapper[4766]: I1126 01:54:29.100722 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-7368-account-create-update-qcqxb"] Nov 26 01:54:29 crc kubenswrapper[4766]: I1126 01:54:29.119049 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-8lxnr"] Nov 26 01:54:29 crc kubenswrapper[4766]: I1126 01:54:29.138796 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-7368-account-create-update-qcqxb"] Nov 26 01:54:29 crc kubenswrapper[4766]: I1126 01:54:29.163846 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-8lxnr"] Nov 26 01:54:29 crc kubenswrapper[4766]: I1126 01:54:29.846695 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e" path="/var/lib/kubelet/pods/aa90b0ba-9aef-4898-ad1e-2d804aeb1b4e/volumes" Nov 26 01:54:29 crc kubenswrapper[4766]: I1126 01:54:29.848225 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcf71c2f-342b-4253-b16e-da87cebb754f" path="/var/lib/kubelet/pods/dcf71c2f-342b-4253-b16e-da87cebb754f/volumes" Nov 26 01:54:41 crc kubenswrapper[4766]: I1126 01:54:41.480096 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:54:41 crc kubenswrapper[4766]: I1126 01:54:41.480898 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:55:01 crc kubenswrapper[4766]: I1126 01:55:01.071955 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-dcdnc"] Nov 26 01:55:01 crc kubenswrapper[4766]: I1126 01:55:01.087959 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-dcdnc"] Nov 26 01:55:01 crc kubenswrapper[4766]: I1126 01:55:01.846615 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18eb3d3c-9197-4730-b085-31e77d7741f1" path="/var/lib/kubelet/pods/18eb3d3c-9197-4730-b085-31e77d7741f1/volumes" Nov 26 01:55:04 crc kubenswrapper[4766]: I1126 01:55:04.764482 4766 scope.go:117] "RemoveContainer" containerID="b944d21a05b42f79189879092c9064e4cbcb7f134a249ddc8945aec1c7805db2" Nov 26 01:55:04 crc kubenswrapper[4766]: I1126 01:55:04.791428 4766 scope.go:117] "RemoveContainer" containerID="0968be81c771f8ac5a4f846ffab730aa1e8eb3abf04354265c8586d7d2b46dc0" Nov 26 01:55:04 crc kubenswrapper[4766]: I1126 01:55:04.850533 4766 scope.go:117] "RemoveContainer" containerID="67dfd0eeeda51abb33036c92ddbe133254089a8baec53a765fef53db80e54c18" Nov 26 01:55:11 crc kubenswrapper[4766]: I1126 01:55:11.479306 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:55:11 crc kubenswrapper[4766]: I1126 01:55:11.479905 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:55:41 crc kubenswrapper[4766]: I1126 01:55:41.479419 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:55:41 crc kubenswrapper[4766]: I1126 01:55:41.480104 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:55:41 crc kubenswrapper[4766]: I1126 01:55:41.480192 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 01:55:41 crc kubenswrapper[4766]: I1126 01:55:41.481634 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8cb92b56012b319b19fabfd8594272dd6949e442ab3db81280faac01a61052b7"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 01:55:41 crc kubenswrapper[4766]: I1126 01:55:41.481847 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://8cb92b56012b319b19fabfd8594272dd6949e442ab3db81280faac01a61052b7" gracePeriod=600 Nov 26 01:55:42 crc kubenswrapper[4766]: I1126 01:55:42.051595 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="8cb92b56012b319b19fabfd8594272dd6949e442ab3db81280faac01a61052b7" exitCode=0 Nov 26 01:55:42 crc kubenswrapper[4766]: I1126 01:55:42.051759 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"8cb92b56012b319b19fabfd8594272dd6949e442ab3db81280faac01a61052b7"} Nov 26 01:55:42 crc kubenswrapper[4766]: I1126 01:55:42.052352 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135"} Nov 26 01:55:42 crc kubenswrapper[4766]: I1126 01:55:42.052389 4766 scope.go:117] "RemoveContainer" containerID="236fc89839d1e6fce15e5186f534208219da5c1f183634586e882ce7a7bdc479" Nov 26 01:57:08 crc kubenswrapper[4766]: I1126 01:57:08.806978 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="44f36c8b-51b8-4bce-afbb-1962eb317dec" containerName="galera" probeResult="failure" output="command timed out" Nov 26 01:57:41 crc kubenswrapper[4766]: I1126 01:57:41.479682 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:57:41 crc kubenswrapper[4766]: I1126 01:57:41.480343 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.060175 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8smb8"] Nov 26 01:57:43 crc kubenswrapper[4766]: E1126 01:57:43.061296 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e807a4b-80e7-4c70-b58a-89a2fcae7905" containerName="extract-utilities" Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.061332 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e807a4b-80e7-4c70-b58a-89a2fcae7905" containerName="extract-utilities" Nov 26 01:57:43 crc kubenswrapper[4766]: E1126 01:57:43.061371 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6e0ae73-b8bd-4462-95cc-6203fbba9b30" containerName="extract-utilities" Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.061386 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6e0ae73-b8bd-4462-95cc-6203fbba9b30" containerName="extract-utilities" Nov 26 01:57:43 crc kubenswrapper[4766]: E1126 01:57:43.061452 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6e0ae73-b8bd-4462-95cc-6203fbba9b30" containerName="extract-content" Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.061467 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6e0ae73-b8bd-4462-95cc-6203fbba9b30" containerName="extract-content" Nov 26 01:57:43 crc kubenswrapper[4766]: E1126 01:57:43.061494 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6e0ae73-b8bd-4462-95cc-6203fbba9b30" containerName="registry-server" Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.061505 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6e0ae73-b8bd-4462-95cc-6203fbba9b30" containerName="registry-server" Nov 26 01:57:43 crc kubenswrapper[4766]: E1126 01:57:43.061520 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e807a4b-80e7-4c70-b58a-89a2fcae7905" containerName="registry-server" Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.061532 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e807a4b-80e7-4c70-b58a-89a2fcae7905" containerName="registry-server" Nov 26 01:57:43 crc kubenswrapper[4766]: E1126 01:57:43.061574 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e807a4b-80e7-4c70-b58a-89a2fcae7905" containerName="extract-content" Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.061587 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e807a4b-80e7-4c70-b58a-89a2fcae7905" containerName="extract-content" Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.062140 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6e0ae73-b8bd-4462-95cc-6203fbba9b30" containerName="registry-server" Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.062185 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e807a4b-80e7-4c70-b58a-89a2fcae7905" containerName="registry-server" Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.063920 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8smb8" Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.087760 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8smb8"] Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.168953 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7-catalog-content\") pod \"redhat-marketplace-8smb8\" (UID: \"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7\") " pod="openshift-marketplace/redhat-marketplace-8smb8" Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.169034 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7-utilities\") pod \"redhat-marketplace-8smb8\" (UID: \"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7\") " pod="openshift-marketplace/redhat-marketplace-8smb8" Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.169117 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl77g\" (UniqueName: \"kubernetes.io/projected/ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7-kube-api-access-bl77g\") pod \"redhat-marketplace-8smb8\" (UID: \"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7\") " pod="openshift-marketplace/redhat-marketplace-8smb8" Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.270733 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl77g\" (UniqueName: \"kubernetes.io/projected/ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7-kube-api-access-bl77g\") pod \"redhat-marketplace-8smb8\" (UID: \"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7\") " pod="openshift-marketplace/redhat-marketplace-8smb8" Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.270885 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7-catalog-content\") pod \"redhat-marketplace-8smb8\" (UID: \"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7\") " pod="openshift-marketplace/redhat-marketplace-8smb8" Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.270933 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7-utilities\") pod \"redhat-marketplace-8smb8\" (UID: \"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7\") " pod="openshift-marketplace/redhat-marketplace-8smb8" Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.271506 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7-utilities\") pod \"redhat-marketplace-8smb8\" (UID: \"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7\") " pod="openshift-marketplace/redhat-marketplace-8smb8" Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.271507 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7-catalog-content\") pod \"redhat-marketplace-8smb8\" (UID: \"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7\") " pod="openshift-marketplace/redhat-marketplace-8smb8" Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.289961 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl77g\" (UniqueName: \"kubernetes.io/projected/ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7-kube-api-access-bl77g\") pod \"redhat-marketplace-8smb8\" (UID: \"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7\") " pod="openshift-marketplace/redhat-marketplace-8smb8" Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.389331 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8smb8" Nov 26 01:57:43 crc kubenswrapper[4766]: I1126 01:57:43.888115 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8smb8"] Nov 26 01:57:43 crc kubenswrapper[4766]: W1126 01:57:43.905423 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac8ffc97_8eeb_46f6_b97a_25532dcd7ae7.slice/crio-99902bd76aa9e94c3a2e32c44a817ab55e396dbae0c22045cd123beb0345a6ff WatchSource:0}: Error finding container 99902bd76aa9e94c3a2e32c44a817ab55e396dbae0c22045cd123beb0345a6ff: Status 404 returned error can't find the container with id 99902bd76aa9e94c3a2e32c44a817ab55e396dbae0c22045cd123beb0345a6ff Nov 26 01:57:44 crc kubenswrapper[4766]: I1126 01:57:44.786938 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8smb8" event={"ID":"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7","Type":"ContainerStarted","Data":"99902bd76aa9e94c3a2e32c44a817ab55e396dbae0c22045cd123beb0345a6ff"} Nov 26 01:57:45 crc kubenswrapper[4766]: I1126 01:57:45.805134 4766 generic.go:334] "Generic (PLEG): container finished" podID="ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7" containerID="9618d99e8b8260bed5459f295278497f0bbe6c6587a6c8cbb874f43fd71d0cb4" exitCode=0 Nov 26 01:57:45 crc kubenswrapper[4766]: I1126 01:57:45.805234 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8smb8" event={"ID":"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7","Type":"ContainerDied","Data":"9618d99e8b8260bed5459f295278497f0bbe6c6587a6c8cbb874f43fd71d0cb4"} Nov 26 01:57:45 crc kubenswrapper[4766]: I1126 01:57:45.808427 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 01:57:46 crc kubenswrapper[4766]: I1126 01:57:46.820772 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8smb8" event={"ID":"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7","Type":"ContainerStarted","Data":"9f624de3d1980e4edd2076ba8781eaf39080118e16f8f2964680e9d3f8b567ea"} Nov 26 01:57:47 crc kubenswrapper[4766]: I1126 01:57:47.840808 4766 generic.go:334] "Generic (PLEG): container finished" podID="ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7" containerID="9f624de3d1980e4edd2076ba8781eaf39080118e16f8f2964680e9d3f8b567ea" exitCode=0 Nov 26 01:57:47 crc kubenswrapper[4766]: I1126 01:57:47.849879 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8smb8" event={"ID":"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7","Type":"ContainerDied","Data":"9f624de3d1980e4edd2076ba8781eaf39080118e16f8f2964680e9d3f8b567ea"} Nov 26 01:57:48 crc kubenswrapper[4766]: I1126 01:57:48.910098 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8smb8" event={"ID":"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7","Type":"ContainerStarted","Data":"3470a1e471d892e8b14ae97f350a4886f4bfb97e13585eef36b654b4e86609c1"} Nov 26 01:57:48 crc kubenswrapper[4766]: I1126 01:57:48.945251 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8smb8" podStartSLOduration=3.44703295 podStartE2EDuration="5.945232451s" podCreationTimestamp="2025-11-26 01:57:43 +0000 UTC" firstStartedPulling="2025-11-26 01:57:45.808196161 +0000 UTC m=+5646.656966591" lastFinishedPulling="2025-11-26 01:57:48.306395632 +0000 UTC m=+5649.155166092" observedRunningTime="2025-11-26 01:57:48.937052692 +0000 UTC m=+5649.785823122" watchObservedRunningTime="2025-11-26 01:57:48.945232451 +0000 UTC m=+5649.794002881" Nov 26 01:57:53 crc kubenswrapper[4766]: I1126 01:57:53.389920 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8smb8" Nov 26 01:57:53 crc kubenswrapper[4766]: I1126 01:57:53.391723 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8smb8" Nov 26 01:57:53 crc kubenswrapper[4766]: I1126 01:57:53.481233 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8smb8" Nov 26 01:57:54 crc kubenswrapper[4766]: I1126 01:57:54.044564 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8smb8" Nov 26 01:57:54 crc kubenswrapper[4766]: I1126 01:57:54.111661 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8smb8"] Nov 26 01:57:56 crc kubenswrapper[4766]: I1126 01:57:56.010225 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8smb8" podUID="ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7" containerName="registry-server" containerID="cri-o://3470a1e471d892e8b14ae97f350a4886f4bfb97e13585eef36b654b4e86609c1" gracePeriod=2 Nov 26 01:57:56 crc kubenswrapper[4766]: I1126 01:57:56.585972 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8smb8" Nov 26 01:57:56 crc kubenswrapper[4766]: I1126 01:57:56.685592 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7-catalog-content\") pod \"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7\" (UID: \"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7\") " Nov 26 01:57:56 crc kubenswrapper[4766]: I1126 01:57:56.685986 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7-utilities\") pod \"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7\" (UID: \"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7\") " Nov 26 01:57:56 crc kubenswrapper[4766]: I1126 01:57:56.686039 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bl77g\" (UniqueName: \"kubernetes.io/projected/ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7-kube-api-access-bl77g\") pod \"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7\" (UID: \"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7\") " Nov 26 01:57:56 crc kubenswrapper[4766]: I1126 01:57:56.686845 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7-utilities" (OuterVolumeSpecName: "utilities") pod "ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7" (UID: "ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:57:56 crc kubenswrapper[4766]: I1126 01:57:56.695897 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7-kube-api-access-bl77g" (OuterVolumeSpecName: "kube-api-access-bl77g") pod "ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7" (UID: "ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7"). InnerVolumeSpecName "kube-api-access-bl77g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 01:57:56 crc kubenswrapper[4766]: I1126 01:57:56.717973 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7" (UID: "ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 01:57:56 crc kubenswrapper[4766]: I1126 01:57:56.788427 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 01:57:56 crc kubenswrapper[4766]: I1126 01:57:56.788763 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bl77g\" (UniqueName: \"kubernetes.io/projected/ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7-kube-api-access-bl77g\") on node \"crc\" DevicePath \"\"" Nov 26 01:57:56 crc kubenswrapper[4766]: I1126 01:57:56.788830 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 01:57:57 crc kubenswrapper[4766]: I1126 01:57:57.023164 4766 generic.go:334] "Generic (PLEG): container finished" podID="ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7" containerID="3470a1e471d892e8b14ae97f350a4886f4bfb97e13585eef36b654b4e86609c1" exitCode=0 Nov 26 01:57:57 crc kubenswrapper[4766]: I1126 01:57:57.023208 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8smb8" event={"ID":"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7","Type":"ContainerDied","Data":"3470a1e471d892e8b14ae97f350a4886f4bfb97e13585eef36b654b4e86609c1"} Nov 26 01:57:57 crc kubenswrapper[4766]: I1126 01:57:57.023241 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8smb8" event={"ID":"ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7","Type":"ContainerDied","Data":"99902bd76aa9e94c3a2e32c44a817ab55e396dbae0c22045cd123beb0345a6ff"} Nov 26 01:57:57 crc kubenswrapper[4766]: I1126 01:57:57.023260 4766 scope.go:117] "RemoveContainer" containerID="3470a1e471d892e8b14ae97f350a4886f4bfb97e13585eef36b654b4e86609c1" Nov 26 01:57:57 crc kubenswrapper[4766]: I1126 01:57:57.023257 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8smb8" Nov 26 01:57:57 crc kubenswrapper[4766]: I1126 01:57:57.052759 4766 scope.go:117] "RemoveContainer" containerID="9f624de3d1980e4edd2076ba8781eaf39080118e16f8f2964680e9d3f8b567ea" Nov 26 01:57:57 crc kubenswrapper[4766]: I1126 01:57:57.083013 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8smb8"] Nov 26 01:57:57 crc kubenswrapper[4766]: I1126 01:57:57.097968 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8smb8"] Nov 26 01:57:57 crc kubenswrapper[4766]: I1126 01:57:57.528866 4766 scope.go:117] "RemoveContainer" containerID="9618d99e8b8260bed5459f295278497f0bbe6c6587a6c8cbb874f43fd71d0cb4" Nov 26 01:57:57 crc kubenswrapper[4766]: I1126 01:57:57.608159 4766 scope.go:117] "RemoveContainer" containerID="3470a1e471d892e8b14ae97f350a4886f4bfb97e13585eef36b654b4e86609c1" Nov 26 01:57:57 crc kubenswrapper[4766]: E1126 01:57:57.608704 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3470a1e471d892e8b14ae97f350a4886f4bfb97e13585eef36b654b4e86609c1\": container with ID starting with 3470a1e471d892e8b14ae97f350a4886f4bfb97e13585eef36b654b4e86609c1 not found: ID does not exist" containerID="3470a1e471d892e8b14ae97f350a4886f4bfb97e13585eef36b654b4e86609c1" Nov 26 01:57:57 crc kubenswrapper[4766]: I1126 01:57:57.608797 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3470a1e471d892e8b14ae97f350a4886f4bfb97e13585eef36b654b4e86609c1"} err="failed to get container status \"3470a1e471d892e8b14ae97f350a4886f4bfb97e13585eef36b654b4e86609c1\": rpc error: code = NotFound desc = could not find container \"3470a1e471d892e8b14ae97f350a4886f4bfb97e13585eef36b654b4e86609c1\": container with ID starting with 3470a1e471d892e8b14ae97f350a4886f4bfb97e13585eef36b654b4e86609c1 not found: ID does not exist" Nov 26 01:57:57 crc kubenswrapper[4766]: I1126 01:57:57.608875 4766 scope.go:117] "RemoveContainer" containerID="9f624de3d1980e4edd2076ba8781eaf39080118e16f8f2964680e9d3f8b567ea" Nov 26 01:57:57 crc kubenswrapper[4766]: E1126 01:57:57.609445 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f624de3d1980e4edd2076ba8781eaf39080118e16f8f2964680e9d3f8b567ea\": container with ID starting with 9f624de3d1980e4edd2076ba8781eaf39080118e16f8f2964680e9d3f8b567ea not found: ID does not exist" containerID="9f624de3d1980e4edd2076ba8781eaf39080118e16f8f2964680e9d3f8b567ea" Nov 26 01:57:57 crc kubenswrapper[4766]: I1126 01:57:57.609475 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f624de3d1980e4edd2076ba8781eaf39080118e16f8f2964680e9d3f8b567ea"} err="failed to get container status \"9f624de3d1980e4edd2076ba8781eaf39080118e16f8f2964680e9d3f8b567ea\": rpc error: code = NotFound desc = could not find container \"9f624de3d1980e4edd2076ba8781eaf39080118e16f8f2964680e9d3f8b567ea\": container with ID starting with 9f624de3d1980e4edd2076ba8781eaf39080118e16f8f2964680e9d3f8b567ea not found: ID does not exist" Nov 26 01:57:57 crc kubenswrapper[4766]: I1126 01:57:57.609499 4766 scope.go:117] "RemoveContainer" containerID="9618d99e8b8260bed5459f295278497f0bbe6c6587a6c8cbb874f43fd71d0cb4" Nov 26 01:57:57 crc kubenswrapper[4766]: E1126 01:57:57.609972 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9618d99e8b8260bed5459f295278497f0bbe6c6587a6c8cbb874f43fd71d0cb4\": container with ID starting with 9618d99e8b8260bed5459f295278497f0bbe6c6587a6c8cbb874f43fd71d0cb4 not found: ID does not exist" containerID="9618d99e8b8260bed5459f295278497f0bbe6c6587a6c8cbb874f43fd71d0cb4" Nov 26 01:57:57 crc kubenswrapper[4766]: I1126 01:57:57.610046 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9618d99e8b8260bed5459f295278497f0bbe6c6587a6c8cbb874f43fd71d0cb4"} err="failed to get container status \"9618d99e8b8260bed5459f295278497f0bbe6c6587a6c8cbb874f43fd71d0cb4\": rpc error: code = NotFound desc = could not find container \"9618d99e8b8260bed5459f295278497f0bbe6c6587a6c8cbb874f43fd71d0cb4\": container with ID starting with 9618d99e8b8260bed5459f295278497f0bbe6c6587a6c8cbb874f43fd71d0cb4 not found: ID does not exist" Nov 26 01:57:57 crc kubenswrapper[4766]: I1126 01:57:57.849083 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7" path="/var/lib/kubelet/pods/ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7/volumes" Nov 26 01:58:11 crc kubenswrapper[4766]: I1126 01:58:11.497324 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:58:11 crc kubenswrapper[4766]: I1126 01:58:11.497812 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:58:41 crc kubenswrapper[4766]: I1126 01:58:41.479888 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 01:58:41 crc kubenswrapper[4766]: I1126 01:58:41.480192 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 01:58:41 crc kubenswrapper[4766]: I1126 01:58:41.480236 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 01:58:41 crc kubenswrapper[4766]: I1126 01:58:41.481041 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 01:58:41 crc kubenswrapper[4766]: I1126 01:58:41.481086 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" gracePeriod=600 Nov 26 01:58:41 crc kubenswrapper[4766]: E1126 01:58:41.604084 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:58:42 crc kubenswrapper[4766]: I1126 01:58:42.597190 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" exitCode=0 Nov 26 01:58:42 crc kubenswrapper[4766]: I1126 01:58:42.597304 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135"} Nov 26 01:58:42 crc kubenswrapper[4766]: I1126 01:58:42.597630 4766 scope.go:117] "RemoveContainer" containerID="8cb92b56012b319b19fabfd8594272dd6949e442ab3db81280faac01a61052b7" Nov 26 01:58:42 crc kubenswrapper[4766]: I1126 01:58:42.598622 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 01:58:42 crc kubenswrapper[4766]: E1126 01:58:42.599277 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:58:54 crc kubenswrapper[4766]: I1126 01:58:54.828407 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 01:58:54 crc kubenswrapper[4766]: E1126 01:58:54.829487 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:59:05 crc kubenswrapper[4766]: I1126 01:59:05.827987 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 01:59:05 crc kubenswrapper[4766]: E1126 01:59:05.829377 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:59:14 crc kubenswrapper[4766]: I1126 01:59:14.256607 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-g4bjk" podUID="abf46edd-34b9-42e7-94ca-a3066047cb53" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 01:59:20 crc kubenswrapper[4766]: I1126 01:59:20.827336 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 01:59:20 crc kubenswrapper[4766]: E1126 01:59:20.828625 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:59:31 crc kubenswrapper[4766]: I1126 01:59:31.826957 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 01:59:31 crc kubenswrapper[4766]: E1126 01:59:31.827866 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:59:45 crc kubenswrapper[4766]: I1126 01:59:45.838160 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 01:59:45 crc kubenswrapper[4766]: E1126 01:59:45.839195 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 01:59:59 crc kubenswrapper[4766]: I1126 01:59:59.837404 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 01:59:59 crc kubenswrapper[4766]: E1126 01:59:59.838440 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:00:00 crc kubenswrapper[4766]: I1126 02:00:00.201694 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt"] Nov 26 02:00:00 crc kubenswrapper[4766]: E1126 02:00:00.202340 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7" containerName="extract-utilities" Nov 26 02:00:00 crc kubenswrapper[4766]: I1126 02:00:00.202373 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7" containerName="extract-utilities" Nov 26 02:00:00 crc kubenswrapper[4766]: E1126 02:00:00.202436 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7" containerName="registry-server" Nov 26 02:00:00 crc kubenswrapper[4766]: I1126 02:00:00.202451 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7" containerName="registry-server" Nov 26 02:00:00 crc kubenswrapper[4766]: E1126 02:00:00.202493 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7" containerName="extract-content" Nov 26 02:00:00 crc kubenswrapper[4766]: I1126 02:00:00.202506 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7" containerName="extract-content" Nov 26 02:00:00 crc kubenswrapper[4766]: I1126 02:00:00.203126 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac8ffc97-8eeb-46f6-b97a-25532dcd7ae7" containerName="registry-server" Nov 26 02:00:00 crc kubenswrapper[4766]: I1126 02:00:00.204372 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt" Nov 26 02:00:00 crc kubenswrapper[4766]: I1126 02:00:00.207187 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 26 02:00:00 crc kubenswrapper[4766]: I1126 02:00:00.209929 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 26 02:00:00 crc kubenswrapper[4766]: I1126 02:00:00.218526 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cll75\" (UniqueName: \"kubernetes.io/projected/79ce3c1d-945d-44d0-a263-de55f580b561-kube-api-access-cll75\") pod \"collect-profiles-29402040-hg5nt\" (UID: \"79ce3c1d-945d-44d0-a263-de55f580b561\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt" Nov 26 02:00:00 crc kubenswrapper[4766]: I1126 02:00:00.218613 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/79ce3c1d-945d-44d0-a263-de55f580b561-config-volume\") pod \"collect-profiles-29402040-hg5nt\" (UID: \"79ce3c1d-945d-44d0-a263-de55f580b561\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt" Nov 26 02:00:00 crc kubenswrapper[4766]: I1126 02:00:00.218687 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/79ce3c1d-945d-44d0-a263-de55f580b561-secret-volume\") pod \"collect-profiles-29402040-hg5nt\" (UID: \"79ce3c1d-945d-44d0-a263-de55f580b561\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt" Nov 26 02:00:00 crc kubenswrapper[4766]: I1126 02:00:00.220572 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt"] Nov 26 02:00:00 crc kubenswrapper[4766]: I1126 02:00:00.322430 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cll75\" (UniqueName: \"kubernetes.io/projected/79ce3c1d-945d-44d0-a263-de55f580b561-kube-api-access-cll75\") pod \"collect-profiles-29402040-hg5nt\" (UID: \"79ce3c1d-945d-44d0-a263-de55f580b561\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt" Nov 26 02:00:00 crc kubenswrapper[4766]: I1126 02:00:00.322491 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/79ce3c1d-945d-44d0-a263-de55f580b561-config-volume\") pod \"collect-profiles-29402040-hg5nt\" (UID: \"79ce3c1d-945d-44d0-a263-de55f580b561\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt" Nov 26 02:00:00 crc kubenswrapper[4766]: I1126 02:00:00.322523 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/79ce3c1d-945d-44d0-a263-de55f580b561-secret-volume\") pod \"collect-profiles-29402040-hg5nt\" (UID: \"79ce3c1d-945d-44d0-a263-de55f580b561\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt" Nov 26 02:00:00 crc kubenswrapper[4766]: I1126 02:00:00.325305 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/79ce3c1d-945d-44d0-a263-de55f580b561-config-volume\") pod \"collect-profiles-29402040-hg5nt\" (UID: \"79ce3c1d-945d-44d0-a263-de55f580b561\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt" Nov 26 02:00:00 crc kubenswrapper[4766]: I1126 02:00:00.338694 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/79ce3c1d-945d-44d0-a263-de55f580b561-secret-volume\") pod \"collect-profiles-29402040-hg5nt\" (UID: \"79ce3c1d-945d-44d0-a263-de55f580b561\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt" Nov 26 02:00:00 crc kubenswrapper[4766]: I1126 02:00:00.350702 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cll75\" (UniqueName: \"kubernetes.io/projected/79ce3c1d-945d-44d0-a263-de55f580b561-kube-api-access-cll75\") pod \"collect-profiles-29402040-hg5nt\" (UID: \"79ce3c1d-945d-44d0-a263-de55f580b561\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt" Nov 26 02:00:00 crc kubenswrapper[4766]: I1126 02:00:00.520186 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt" Nov 26 02:00:01 crc kubenswrapper[4766]: I1126 02:00:01.014610 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt"] Nov 26 02:00:01 crc kubenswrapper[4766]: W1126 02:00:01.920027 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79ce3c1d_945d_44d0_a263_de55f580b561.slice/crio-2dc9b26c5ddaed89a379047ae6c742b7a27339800ce0b8444b202cc59827d84f WatchSource:0}: Error finding container 2dc9b26c5ddaed89a379047ae6c742b7a27339800ce0b8444b202cc59827d84f: Status 404 returned error can't find the container with id 2dc9b26c5ddaed89a379047ae6c742b7a27339800ce0b8444b202cc59827d84f Nov 26 02:00:02 crc kubenswrapper[4766]: I1126 02:00:02.509771 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt" event={"ID":"79ce3c1d-945d-44d0-a263-de55f580b561","Type":"ContainerStarted","Data":"6ae52fd9463077d7b32b88151ac7223a081d7a6fae971789b68c1f4b947a9b13"} Nov 26 02:00:02 crc kubenswrapper[4766]: I1126 02:00:02.510679 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt" event={"ID":"79ce3c1d-945d-44d0-a263-de55f580b561","Type":"ContainerStarted","Data":"2dc9b26c5ddaed89a379047ae6c742b7a27339800ce0b8444b202cc59827d84f"} Nov 26 02:00:02 crc kubenswrapper[4766]: I1126 02:00:02.539492 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt" podStartSLOduration=2.539468319 podStartE2EDuration="2.539468319s" podCreationTimestamp="2025-11-26 02:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 02:00:02.530024269 +0000 UTC m=+5783.378794699" watchObservedRunningTime="2025-11-26 02:00:02.539468319 +0000 UTC m=+5783.388238769" Nov 26 02:00:03 crc kubenswrapper[4766]: I1126 02:00:03.531422 4766 generic.go:334] "Generic (PLEG): container finished" podID="79ce3c1d-945d-44d0-a263-de55f580b561" containerID="6ae52fd9463077d7b32b88151ac7223a081d7a6fae971789b68c1f4b947a9b13" exitCode=0 Nov 26 02:00:03 crc kubenswrapper[4766]: I1126 02:00:03.531497 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt" event={"ID":"79ce3c1d-945d-44d0-a263-de55f580b561","Type":"ContainerDied","Data":"6ae52fd9463077d7b32b88151ac7223a081d7a6fae971789b68c1f4b947a9b13"} Nov 26 02:00:05 crc kubenswrapper[4766]: I1126 02:00:05.232884 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt" Nov 26 02:00:05 crc kubenswrapper[4766]: I1126 02:00:05.251186 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/79ce3c1d-945d-44d0-a263-de55f580b561-config-volume\") pod \"79ce3c1d-945d-44d0-a263-de55f580b561\" (UID: \"79ce3c1d-945d-44d0-a263-de55f580b561\") " Nov 26 02:00:05 crc kubenswrapper[4766]: I1126 02:00:05.251603 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/79ce3c1d-945d-44d0-a263-de55f580b561-secret-volume\") pod \"79ce3c1d-945d-44d0-a263-de55f580b561\" (UID: \"79ce3c1d-945d-44d0-a263-de55f580b561\") " Nov 26 02:00:05 crc kubenswrapper[4766]: I1126 02:00:05.251705 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cll75\" (UniqueName: \"kubernetes.io/projected/79ce3c1d-945d-44d0-a263-de55f580b561-kube-api-access-cll75\") pod \"79ce3c1d-945d-44d0-a263-de55f580b561\" (UID: \"79ce3c1d-945d-44d0-a263-de55f580b561\") " Nov 26 02:00:05 crc kubenswrapper[4766]: I1126 02:00:05.252031 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79ce3c1d-945d-44d0-a263-de55f580b561-config-volume" (OuterVolumeSpecName: "config-volume") pod "79ce3c1d-945d-44d0-a263-de55f580b561" (UID: "79ce3c1d-945d-44d0-a263-de55f580b561"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 02:00:05 crc kubenswrapper[4766]: I1126 02:00:05.252557 4766 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/79ce3c1d-945d-44d0-a263-de55f580b561-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 02:00:05 crc kubenswrapper[4766]: I1126 02:00:05.262816 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79ce3c1d-945d-44d0-a263-de55f580b561-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "79ce3c1d-945d-44d0-a263-de55f580b561" (UID: "79ce3c1d-945d-44d0-a263-de55f580b561"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 02:00:05 crc kubenswrapper[4766]: I1126 02:00:05.263135 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79ce3c1d-945d-44d0-a263-de55f580b561-kube-api-access-cll75" (OuterVolumeSpecName: "kube-api-access-cll75") pod "79ce3c1d-945d-44d0-a263-de55f580b561" (UID: "79ce3c1d-945d-44d0-a263-de55f580b561"). InnerVolumeSpecName "kube-api-access-cll75". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:00:05 crc kubenswrapper[4766]: I1126 02:00:05.354580 4766 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/79ce3c1d-945d-44d0-a263-de55f580b561-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 26 02:00:05 crc kubenswrapper[4766]: I1126 02:00:05.354623 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cll75\" (UniqueName: \"kubernetes.io/projected/79ce3c1d-945d-44d0-a263-de55f580b561-kube-api-access-cll75\") on node \"crc\" DevicePath \"\"" Nov 26 02:00:05 crc kubenswrapper[4766]: I1126 02:00:05.566032 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt" event={"ID":"79ce3c1d-945d-44d0-a263-de55f580b561","Type":"ContainerDied","Data":"2dc9b26c5ddaed89a379047ae6c742b7a27339800ce0b8444b202cc59827d84f"} Nov 26 02:00:05 crc kubenswrapper[4766]: I1126 02:00:05.566069 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2dc9b26c5ddaed89a379047ae6c742b7a27339800ce0b8444b202cc59827d84f" Nov 26 02:00:05 crc kubenswrapper[4766]: I1126 02:00:05.566133 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt" Nov 26 02:00:05 crc kubenswrapper[4766]: I1126 02:00:05.628623 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401995-pd7kg"] Nov 26 02:00:05 crc kubenswrapper[4766]: I1126 02:00:05.643505 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401995-pd7kg"] Nov 26 02:00:05 crc kubenswrapper[4766]: I1126 02:00:05.851244 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec347d52-c090-468b-8055-260ca66491bb" path="/var/lib/kubelet/pods/ec347d52-c090-468b-8055-260ca66491bb/volumes" Nov 26 02:00:14 crc kubenswrapper[4766]: I1126 02:00:14.827529 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 02:00:14 crc kubenswrapper[4766]: E1126 02:00:14.828879 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:00:17 crc kubenswrapper[4766]: I1126 02:00:17.692291 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kcjz9"] Nov 26 02:00:17 crc kubenswrapper[4766]: E1126 02:00:17.697336 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79ce3c1d-945d-44d0-a263-de55f580b561" containerName="collect-profiles" Nov 26 02:00:17 crc kubenswrapper[4766]: I1126 02:00:17.697363 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="79ce3c1d-945d-44d0-a263-de55f580b561" containerName="collect-profiles" Nov 26 02:00:17 crc kubenswrapper[4766]: I1126 02:00:17.697724 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="79ce3c1d-945d-44d0-a263-de55f580b561" containerName="collect-profiles" Nov 26 02:00:17 crc kubenswrapper[4766]: I1126 02:00:17.699874 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kcjz9" Nov 26 02:00:17 crc kubenswrapper[4766]: I1126 02:00:17.733035 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kcjz9"] Nov 26 02:00:17 crc kubenswrapper[4766]: I1126 02:00:17.876275 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f21a4925-e375-4826-9bf4-01e2c7966f9f-catalog-content\") pod \"redhat-operators-kcjz9\" (UID: \"f21a4925-e375-4826-9bf4-01e2c7966f9f\") " pod="openshift-marketplace/redhat-operators-kcjz9" Nov 26 02:00:17 crc kubenswrapper[4766]: I1126 02:00:17.876350 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f21a4925-e375-4826-9bf4-01e2c7966f9f-utilities\") pod \"redhat-operators-kcjz9\" (UID: \"f21a4925-e375-4826-9bf4-01e2c7966f9f\") " pod="openshift-marketplace/redhat-operators-kcjz9" Nov 26 02:00:17 crc kubenswrapper[4766]: I1126 02:00:17.876550 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn6gp\" (UniqueName: \"kubernetes.io/projected/f21a4925-e375-4826-9bf4-01e2c7966f9f-kube-api-access-zn6gp\") pod \"redhat-operators-kcjz9\" (UID: \"f21a4925-e375-4826-9bf4-01e2c7966f9f\") " pod="openshift-marketplace/redhat-operators-kcjz9" Nov 26 02:00:17 crc kubenswrapper[4766]: I1126 02:00:17.978851 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f21a4925-e375-4826-9bf4-01e2c7966f9f-catalog-content\") pod \"redhat-operators-kcjz9\" (UID: \"f21a4925-e375-4826-9bf4-01e2c7966f9f\") " pod="openshift-marketplace/redhat-operators-kcjz9" Nov 26 02:00:17 crc kubenswrapper[4766]: I1126 02:00:17.978935 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f21a4925-e375-4826-9bf4-01e2c7966f9f-utilities\") pod \"redhat-operators-kcjz9\" (UID: \"f21a4925-e375-4826-9bf4-01e2c7966f9f\") " pod="openshift-marketplace/redhat-operators-kcjz9" Nov 26 02:00:17 crc kubenswrapper[4766]: I1126 02:00:17.978997 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn6gp\" (UniqueName: \"kubernetes.io/projected/f21a4925-e375-4826-9bf4-01e2c7966f9f-kube-api-access-zn6gp\") pod \"redhat-operators-kcjz9\" (UID: \"f21a4925-e375-4826-9bf4-01e2c7966f9f\") " pod="openshift-marketplace/redhat-operators-kcjz9" Nov 26 02:00:17 crc kubenswrapper[4766]: I1126 02:00:17.979859 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f21a4925-e375-4826-9bf4-01e2c7966f9f-catalog-content\") pod \"redhat-operators-kcjz9\" (UID: \"f21a4925-e375-4826-9bf4-01e2c7966f9f\") " pod="openshift-marketplace/redhat-operators-kcjz9" Nov 26 02:00:17 crc kubenswrapper[4766]: I1126 02:00:17.980530 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f21a4925-e375-4826-9bf4-01e2c7966f9f-utilities\") pod \"redhat-operators-kcjz9\" (UID: \"f21a4925-e375-4826-9bf4-01e2c7966f9f\") " pod="openshift-marketplace/redhat-operators-kcjz9" Nov 26 02:00:18 crc kubenswrapper[4766]: I1126 02:00:18.002315 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn6gp\" (UniqueName: \"kubernetes.io/projected/f21a4925-e375-4826-9bf4-01e2c7966f9f-kube-api-access-zn6gp\") pod \"redhat-operators-kcjz9\" (UID: \"f21a4925-e375-4826-9bf4-01e2c7966f9f\") " pod="openshift-marketplace/redhat-operators-kcjz9" Nov 26 02:00:18 crc kubenswrapper[4766]: I1126 02:00:18.027717 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kcjz9" Nov 26 02:00:18 crc kubenswrapper[4766]: I1126 02:00:18.537809 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kcjz9"] Nov 26 02:00:19 crc kubenswrapper[4766]: W1126 02:00:19.222974 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf21a4925_e375_4826_9bf4_01e2c7966f9f.slice/crio-21a005d47461222afa5bbfaa219c56250222e8155cde3dd96613c3d22fc26cb5 WatchSource:0}: Error finding container 21a005d47461222afa5bbfaa219c56250222e8155cde3dd96613c3d22fc26cb5: Status 404 returned error can't find the container with id 21a005d47461222afa5bbfaa219c56250222e8155cde3dd96613c3d22fc26cb5 Nov 26 02:00:19 crc kubenswrapper[4766]: I1126 02:00:19.754948 4766 generic.go:334] "Generic (PLEG): container finished" podID="f21a4925-e375-4826-9bf4-01e2c7966f9f" containerID="10ce281879babc1c1083781dbb6a9753ab5d7c254a62ee23b31363f50c7c36a6" exitCode=0 Nov 26 02:00:19 crc kubenswrapper[4766]: I1126 02:00:19.754995 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kcjz9" event={"ID":"f21a4925-e375-4826-9bf4-01e2c7966f9f","Type":"ContainerDied","Data":"10ce281879babc1c1083781dbb6a9753ab5d7c254a62ee23b31363f50c7c36a6"} Nov 26 02:00:19 crc kubenswrapper[4766]: I1126 02:00:19.755777 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kcjz9" event={"ID":"f21a4925-e375-4826-9bf4-01e2c7966f9f","Type":"ContainerStarted","Data":"21a005d47461222afa5bbfaa219c56250222e8155cde3dd96613c3d22fc26cb5"} Nov 26 02:00:20 crc kubenswrapper[4766]: I1126 02:00:20.767958 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kcjz9" event={"ID":"f21a4925-e375-4826-9bf4-01e2c7966f9f","Type":"ContainerStarted","Data":"a691f1ca4e4f32f3f3ba0bbb26e9254fe701d36c0f14f4320fe79afb038c11eb"} Nov 26 02:00:24 crc kubenswrapper[4766]: I1126 02:00:24.826239 4766 generic.go:334] "Generic (PLEG): container finished" podID="f21a4925-e375-4826-9bf4-01e2c7966f9f" containerID="a691f1ca4e4f32f3f3ba0bbb26e9254fe701d36c0f14f4320fe79afb038c11eb" exitCode=0 Nov 26 02:00:24 crc kubenswrapper[4766]: I1126 02:00:24.826427 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kcjz9" event={"ID":"f21a4925-e375-4826-9bf4-01e2c7966f9f","Type":"ContainerDied","Data":"a691f1ca4e4f32f3f3ba0bbb26e9254fe701d36c0f14f4320fe79afb038c11eb"} Nov 26 02:00:25 crc kubenswrapper[4766]: I1126 02:00:25.852422 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kcjz9" event={"ID":"f21a4925-e375-4826-9bf4-01e2c7966f9f","Type":"ContainerStarted","Data":"d504f09afcfdc6c7a3d203cc51ae4e0ac9e60449bc6bf622426d13660ddf69ab"} Nov 26 02:00:25 crc kubenswrapper[4766]: I1126 02:00:25.890172 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kcjz9" podStartSLOduration=3.3555586760000002 podStartE2EDuration="8.890151899s" podCreationTimestamp="2025-11-26 02:00:17 +0000 UTC" firstStartedPulling="2025-11-26 02:00:19.75716809 +0000 UTC m=+5800.605938520" lastFinishedPulling="2025-11-26 02:00:25.291761273 +0000 UTC m=+5806.140531743" observedRunningTime="2025-11-26 02:00:25.879475019 +0000 UTC m=+5806.728245469" watchObservedRunningTime="2025-11-26 02:00:25.890151899 +0000 UTC m=+5806.738922339" Nov 26 02:00:26 crc kubenswrapper[4766]: I1126 02:00:26.828115 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 02:00:26 crc kubenswrapper[4766]: E1126 02:00:26.828414 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:00:28 crc kubenswrapper[4766]: I1126 02:00:28.029239 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kcjz9" Nov 26 02:00:28 crc kubenswrapper[4766]: I1126 02:00:28.031539 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kcjz9" Nov 26 02:00:29 crc kubenswrapper[4766]: I1126 02:00:29.108901 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kcjz9" podUID="f21a4925-e375-4826-9bf4-01e2c7966f9f" containerName="registry-server" probeResult="failure" output=< Nov 26 02:00:29 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 02:00:29 crc kubenswrapper[4766]: > Nov 26 02:00:37 crc kubenswrapper[4766]: I1126 02:00:37.827325 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 02:00:37 crc kubenswrapper[4766]: E1126 02:00:37.828699 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:00:38 crc kubenswrapper[4766]: I1126 02:00:38.098409 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kcjz9" Nov 26 02:00:38 crc kubenswrapper[4766]: I1126 02:00:38.163507 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kcjz9" Nov 26 02:00:38 crc kubenswrapper[4766]: I1126 02:00:38.366490 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kcjz9"] Nov 26 02:00:40 crc kubenswrapper[4766]: I1126 02:00:40.043102 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kcjz9" podUID="f21a4925-e375-4826-9bf4-01e2c7966f9f" containerName="registry-server" containerID="cri-o://d504f09afcfdc6c7a3d203cc51ae4e0ac9e60449bc6bf622426d13660ddf69ab" gracePeriod=2 Nov 26 02:00:40 crc kubenswrapper[4766]: I1126 02:00:40.695825 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kcjz9" Nov 26 02:00:40 crc kubenswrapper[4766]: I1126 02:00:40.789237 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zn6gp\" (UniqueName: \"kubernetes.io/projected/f21a4925-e375-4826-9bf4-01e2c7966f9f-kube-api-access-zn6gp\") pod \"f21a4925-e375-4826-9bf4-01e2c7966f9f\" (UID: \"f21a4925-e375-4826-9bf4-01e2c7966f9f\") " Nov 26 02:00:40 crc kubenswrapper[4766]: I1126 02:00:40.789681 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f21a4925-e375-4826-9bf4-01e2c7966f9f-utilities\") pod \"f21a4925-e375-4826-9bf4-01e2c7966f9f\" (UID: \"f21a4925-e375-4826-9bf4-01e2c7966f9f\") " Nov 26 02:00:40 crc kubenswrapper[4766]: I1126 02:00:40.789758 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f21a4925-e375-4826-9bf4-01e2c7966f9f-catalog-content\") pod \"f21a4925-e375-4826-9bf4-01e2c7966f9f\" (UID: \"f21a4925-e375-4826-9bf4-01e2c7966f9f\") " Nov 26 02:00:40 crc kubenswrapper[4766]: I1126 02:00:40.799129 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f21a4925-e375-4826-9bf4-01e2c7966f9f-utilities" (OuterVolumeSpecName: "utilities") pod "f21a4925-e375-4826-9bf4-01e2c7966f9f" (UID: "f21a4925-e375-4826-9bf4-01e2c7966f9f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:00:40 crc kubenswrapper[4766]: I1126 02:00:40.832916 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f21a4925-e375-4826-9bf4-01e2c7966f9f-kube-api-access-zn6gp" (OuterVolumeSpecName: "kube-api-access-zn6gp") pod "f21a4925-e375-4826-9bf4-01e2c7966f9f" (UID: "f21a4925-e375-4826-9bf4-01e2c7966f9f"). InnerVolumeSpecName "kube-api-access-zn6gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:00:40 crc kubenswrapper[4766]: I1126 02:00:40.892491 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f21a4925-e375-4826-9bf4-01e2c7966f9f-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 02:00:40 crc kubenswrapper[4766]: I1126 02:00:40.892522 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zn6gp\" (UniqueName: \"kubernetes.io/projected/f21a4925-e375-4826-9bf4-01e2c7966f9f-kube-api-access-zn6gp\") on node \"crc\" DevicePath \"\"" Nov 26 02:00:40 crc kubenswrapper[4766]: I1126 02:00:40.924761 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f21a4925-e375-4826-9bf4-01e2c7966f9f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f21a4925-e375-4826-9bf4-01e2c7966f9f" (UID: "f21a4925-e375-4826-9bf4-01e2c7966f9f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:00:40 crc kubenswrapper[4766]: I1126 02:00:40.995173 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f21a4925-e375-4826-9bf4-01e2c7966f9f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 02:00:41 crc kubenswrapper[4766]: I1126 02:00:41.054901 4766 generic.go:334] "Generic (PLEG): container finished" podID="f21a4925-e375-4826-9bf4-01e2c7966f9f" containerID="d504f09afcfdc6c7a3d203cc51ae4e0ac9e60449bc6bf622426d13660ddf69ab" exitCode=0 Nov 26 02:00:41 crc kubenswrapper[4766]: I1126 02:00:41.054937 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kcjz9" event={"ID":"f21a4925-e375-4826-9bf4-01e2c7966f9f","Type":"ContainerDied","Data":"d504f09afcfdc6c7a3d203cc51ae4e0ac9e60449bc6bf622426d13660ddf69ab"} Nov 26 02:00:41 crc kubenswrapper[4766]: I1126 02:00:41.054974 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kcjz9" Nov 26 02:00:41 crc kubenswrapper[4766]: I1126 02:00:41.054991 4766 scope.go:117] "RemoveContainer" containerID="d504f09afcfdc6c7a3d203cc51ae4e0ac9e60449bc6bf622426d13660ddf69ab" Nov 26 02:00:41 crc kubenswrapper[4766]: I1126 02:00:41.054977 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kcjz9" event={"ID":"f21a4925-e375-4826-9bf4-01e2c7966f9f","Type":"ContainerDied","Data":"21a005d47461222afa5bbfaa219c56250222e8155cde3dd96613c3d22fc26cb5"} Nov 26 02:00:41 crc kubenswrapper[4766]: I1126 02:00:41.102156 4766 scope.go:117] "RemoveContainer" containerID="a691f1ca4e4f32f3f3ba0bbb26e9254fe701d36c0f14f4320fe79afb038c11eb" Nov 26 02:00:41 crc kubenswrapper[4766]: I1126 02:00:41.122838 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kcjz9"] Nov 26 02:00:41 crc kubenswrapper[4766]: I1126 02:00:41.145128 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kcjz9"] Nov 26 02:00:41 crc kubenswrapper[4766]: I1126 02:00:41.148932 4766 scope.go:117] "RemoveContainer" containerID="10ce281879babc1c1083781dbb6a9753ab5d7c254a62ee23b31363f50c7c36a6" Nov 26 02:00:41 crc kubenswrapper[4766]: I1126 02:00:41.188159 4766 scope.go:117] "RemoveContainer" containerID="d504f09afcfdc6c7a3d203cc51ae4e0ac9e60449bc6bf622426d13660ddf69ab" Nov 26 02:00:41 crc kubenswrapper[4766]: E1126 02:00:41.189426 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d504f09afcfdc6c7a3d203cc51ae4e0ac9e60449bc6bf622426d13660ddf69ab\": container with ID starting with d504f09afcfdc6c7a3d203cc51ae4e0ac9e60449bc6bf622426d13660ddf69ab not found: ID does not exist" containerID="d504f09afcfdc6c7a3d203cc51ae4e0ac9e60449bc6bf622426d13660ddf69ab" Nov 26 02:00:41 crc kubenswrapper[4766]: I1126 02:00:41.189475 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d504f09afcfdc6c7a3d203cc51ae4e0ac9e60449bc6bf622426d13660ddf69ab"} err="failed to get container status \"d504f09afcfdc6c7a3d203cc51ae4e0ac9e60449bc6bf622426d13660ddf69ab\": rpc error: code = NotFound desc = could not find container \"d504f09afcfdc6c7a3d203cc51ae4e0ac9e60449bc6bf622426d13660ddf69ab\": container with ID starting with d504f09afcfdc6c7a3d203cc51ae4e0ac9e60449bc6bf622426d13660ddf69ab not found: ID does not exist" Nov 26 02:00:41 crc kubenswrapper[4766]: I1126 02:00:41.189504 4766 scope.go:117] "RemoveContainer" containerID="a691f1ca4e4f32f3f3ba0bbb26e9254fe701d36c0f14f4320fe79afb038c11eb" Nov 26 02:00:41 crc kubenswrapper[4766]: E1126 02:00:41.189948 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a691f1ca4e4f32f3f3ba0bbb26e9254fe701d36c0f14f4320fe79afb038c11eb\": container with ID starting with a691f1ca4e4f32f3f3ba0bbb26e9254fe701d36c0f14f4320fe79afb038c11eb not found: ID does not exist" containerID="a691f1ca4e4f32f3f3ba0bbb26e9254fe701d36c0f14f4320fe79afb038c11eb" Nov 26 02:00:41 crc kubenswrapper[4766]: I1126 02:00:41.190003 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a691f1ca4e4f32f3f3ba0bbb26e9254fe701d36c0f14f4320fe79afb038c11eb"} err="failed to get container status \"a691f1ca4e4f32f3f3ba0bbb26e9254fe701d36c0f14f4320fe79afb038c11eb\": rpc error: code = NotFound desc = could not find container \"a691f1ca4e4f32f3f3ba0bbb26e9254fe701d36c0f14f4320fe79afb038c11eb\": container with ID starting with a691f1ca4e4f32f3f3ba0bbb26e9254fe701d36c0f14f4320fe79afb038c11eb not found: ID does not exist" Nov 26 02:00:41 crc kubenswrapper[4766]: I1126 02:00:41.190043 4766 scope.go:117] "RemoveContainer" containerID="10ce281879babc1c1083781dbb6a9753ab5d7c254a62ee23b31363f50c7c36a6" Nov 26 02:00:41 crc kubenswrapper[4766]: E1126 02:00:41.190369 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10ce281879babc1c1083781dbb6a9753ab5d7c254a62ee23b31363f50c7c36a6\": container with ID starting with 10ce281879babc1c1083781dbb6a9753ab5d7c254a62ee23b31363f50c7c36a6 not found: ID does not exist" containerID="10ce281879babc1c1083781dbb6a9753ab5d7c254a62ee23b31363f50c7c36a6" Nov 26 02:00:41 crc kubenswrapper[4766]: I1126 02:00:41.190412 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10ce281879babc1c1083781dbb6a9753ab5d7c254a62ee23b31363f50c7c36a6"} err="failed to get container status \"10ce281879babc1c1083781dbb6a9753ab5d7c254a62ee23b31363f50c7c36a6\": rpc error: code = NotFound desc = could not find container \"10ce281879babc1c1083781dbb6a9753ab5d7c254a62ee23b31363f50c7c36a6\": container with ID starting with 10ce281879babc1c1083781dbb6a9753ab5d7c254a62ee23b31363f50c7c36a6 not found: ID does not exist" Nov 26 02:00:41 crc kubenswrapper[4766]: I1126 02:00:41.846277 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f21a4925-e375-4826-9bf4-01e2c7966f9f" path="/var/lib/kubelet/pods/f21a4925-e375-4826-9bf4-01e2c7966f9f/volumes" Nov 26 02:00:48 crc kubenswrapper[4766]: I1126 02:00:48.827331 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 02:00:48 crc kubenswrapper[4766]: E1126 02:00:48.828384 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:01:00 crc kubenswrapper[4766]: I1126 02:01:00.165183 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29402041-bjg4c"] Nov 26 02:01:00 crc kubenswrapper[4766]: E1126 02:01:00.166148 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f21a4925-e375-4826-9bf4-01e2c7966f9f" containerName="extract-utilities" Nov 26 02:01:00 crc kubenswrapper[4766]: I1126 02:01:00.166167 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="f21a4925-e375-4826-9bf4-01e2c7966f9f" containerName="extract-utilities" Nov 26 02:01:00 crc kubenswrapper[4766]: E1126 02:01:00.166180 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f21a4925-e375-4826-9bf4-01e2c7966f9f" containerName="registry-server" Nov 26 02:01:00 crc kubenswrapper[4766]: I1126 02:01:00.166187 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="f21a4925-e375-4826-9bf4-01e2c7966f9f" containerName="registry-server" Nov 26 02:01:00 crc kubenswrapper[4766]: E1126 02:01:00.166230 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f21a4925-e375-4826-9bf4-01e2c7966f9f" containerName="extract-content" Nov 26 02:01:00 crc kubenswrapper[4766]: I1126 02:01:00.166238 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="f21a4925-e375-4826-9bf4-01e2c7966f9f" containerName="extract-content" Nov 26 02:01:00 crc kubenswrapper[4766]: I1126 02:01:00.166457 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="f21a4925-e375-4826-9bf4-01e2c7966f9f" containerName="registry-server" Nov 26 02:01:00 crc kubenswrapper[4766]: I1126 02:01:00.167366 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29402041-bjg4c" Nov 26 02:01:00 crc kubenswrapper[4766]: I1126 02:01:00.200146 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29402041-bjg4c"] Nov 26 02:01:00 crc kubenswrapper[4766]: I1126 02:01:00.200268 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd82c837-925c-4835-a90f-3d5a873dfbae-config-data\") pod \"keystone-cron-29402041-bjg4c\" (UID: \"bd82c837-925c-4835-a90f-3d5a873dfbae\") " pod="openstack/keystone-cron-29402041-bjg4c" Nov 26 02:01:00 crc kubenswrapper[4766]: I1126 02:01:00.200462 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bd82c837-925c-4835-a90f-3d5a873dfbae-fernet-keys\") pod \"keystone-cron-29402041-bjg4c\" (UID: \"bd82c837-925c-4835-a90f-3d5a873dfbae\") " pod="openstack/keystone-cron-29402041-bjg4c" Nov 26 02:01:00 crc kubenswrapper[4766]: I1126 02:01:00.200572 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv5r8\" (UniqueName: \"kubernetes.io/projected/bd82c837-925c-4835-a90f-3d5a873dfbae-kube-api-access-zv5r8\") pod \"keystone-cron-29402041-bjg4c\" (UID: \"bd82c837-925c-4835-a90f-3d5a873dfbae\") " pod="openstack/keystone-cron-29402041-bjg4c" Nov 26 02:01:00 crc kubenswrapper[4766]: I1126 02:01:00.200612 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd82c837-925c-4835-a90f-3d5a873dfbae-combined-ca-bundle\") pod \"keystone-cron-29402041-bjg4c\" (UID: \"bd82c837-925c-4835-a90f-3d5a873dfbae\") " pod="openstack/keystone-cron-29402041-bjg4c" Nov 26 02:01:00 crc kubenswrapper[4766]: I1126 02:01:00.302231 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv5r8\" (UniqueName: \"kubernetes.io/projected/bd82c837-925c-4835-a90f-3d5a873dfbae-kube-api-access-zv5r8\") pod \"keystone-cron-29402041-bjg4c\" (UID: \"bd82c837-925c-4835-a90f-3d5a873dfbae\") " pod="openstack/keystone-cron-29402041-bjg4c" Nov 26 02:01:00 crc kubenswrapper[4766]: I1126 02:01:00.302289 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd82c837-925c-4835-a90f-3d5a873dfbae-combined-ca-bundle\") pod \"keystone-cron-29402041-bjg4c\" (UID: \"bd82c837-925c-4835-a90f-3d5a873dfbae\") " pod="openstack/keystone-cron-29402041-bjg4c" Nov 26 02:01:00 crc kubenswrapper[4766]: I1126 02:01:00.302350 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd82c837-925c-4835-a90f-3d5a873dfbae-config-data\") pod \"keystone-cron-29402041-bjg4c\" (UID: \"bd82c837-925c-4835-a90f-3d5a873dfbae\") " pod="openstack/keystone-cron-29402041-bjg4c" Nov 26 02:01:00 crc kubenswrapper[4766]: I1126 02:01:00.302455 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bd82c837-925c-4835-a90f-3d5a873dfbae-fernet-keys\") pod \"keystone-cron-29402041-bjg4c\" (UID: \"bd82c837-925c-4835-a90f-3d5a873dfbae\") " pod="openstack/keystone-cron-29402041-bjg4c" Nov 26 02:01:00 crc kubenswrapper[4766]: I1126 02:01:00.807944 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bd82c837-925c-4835-a90f-3d5a873dfbae-fernet-keys\") pod \"keystone-cron-29402041-bjg4c\" (UID: \"bd82c837-925c-4835-a90f-3d5a873dfbae\") " pod="openstack/keystone-cron-29402041-bjg4c" Nov 26 02:01:00 crc kubenswrapper[4766]: I1126 02:01:00.807951 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd82c837-925c-4835-a90f-3d5a873dfbae-combined-ca-bundle\") pod \"keystone-cron-29402041-bjg4c\" (UID: \"bd82c837-925c-4835-a90f-3d5a873dfbae\") " pod="openstack/keystone-cron-29402041-bjg4c" Nov 26 02:01:00 crc kubenswrapper[4766]: I1126 02:01:00.808290 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd82c837-925c-4835-a90f-3d5a873dfbae-config-data\") pod \"keystone-cron-29402041-bjg4c\" (UID: \"bd82c837-925c-4835-a90f-3d5a873dfbae\") " pod="openstack/keystone-cron-29402041-bjg4c" Nov 26 02:01:00 crc kubenswrapper[4766]: I1126 02:01:00.809048 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv5r8\" (UniqueName: \"kubernetes.io/projected/bd82c837-925c-4835-a90f-3d5a873dfbae-kube-api-access-zv5r8\") pod \"keystone-cron-29402041-bjg4c\" (UID: \"bd82c837-925c-4835-a90f-3d5a873dfbae\") " pod="openstack/keystone-cron-29402041-bjg4c" Nov 26 02:01:01 crc kubenswrapper[4766]: I1126 02:01:01.090184 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29402041-bjg4c" Nov 26 02:01:01 crc kubenswrapper[4766]: I1126 02:01:01.600805 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29402041-bjg4c"] Nov 26 02:01:01 crc kubenswrapper[4766]: I1126 02:01:01.827864 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 02:01:01 crc kubenswrapper[4766]: E1126 02:01:01.828682 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:01:02 crc kubenswrapper[4766]: I1126 02:01:02.350262 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29402041-bjg4c" event={"ID":"bd82c837-925c-4835-a90f-3d5a873dfbae","Type":"ContainerStarted","Data":"9d4c9c2e192a48b17f824dca6531675a18fff727f78bf35de0a8842ebe504216"} Nov 26 02:01:02 crc kubenswrapper[4766]: I1126 02:01:02.350576 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29402041-bjg4c" event={"ID":"bd82c837-925c-4835-a90f-3d5a873dfbae","Type":"ContainerStarted","Data":"2860b3b62e57c141f8b8309d11df5fa3f7de3444b00ba7bb8e25953c7b674976"} Nov 26 02:01:02 crc kubenswrapper[4766]: I1126 02:01:02.368435 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29402041-bjg4c" podStartSLOduration=2.368415405 podStartE2EDuration="2.368415405s" podCreationTimestamp="2025-11-26 02:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 02:01:02.367211966 +0000 UTC m=+5843.215982406" watchObservedRunningTime="2025-11-26 02:01:02.368415405 +0000 UTC m=+5843.217185845" Nov 26 02:01:05 crc kubenswrapper[4766]: I1126 02:01:05.182807 4766 scope.go:117] "RemoveContainer" containerID="47aed24158cb58dd47f194727b92c55ec5f1bca86c10d54bba7f3bdbc3a04684" Nov 26 02:01:05 crc kubenswrapper[4766]: I1126 02:01:05.400426 4766 generic.go:334] "Generic (PLEG): container finished" podID="bd82c837-925c-4835-a90f-3d5a873dfbae" containerID="9d4c9c2e192a48b17f824dca6531675a18fff727f78bf35de0a8842ebe504216" exitCode=0 Nov 26 02:01:05 crc kubenswrapper[4766]: I1126 02:01:05.400475 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29402041-bjg4c" event={"ID":"bd82c837-925c-4835-a90f-3d5a873dfbae","Type":"ContainerDied","Data":"9d4c9c2e192a48b17f824dca6531675a18fff727f78bf35de0a8842ebe504216"} Nov 26 02:01:06 crc kubenswrapper[4766]: I1126 02:01:06.909036 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29402041-bjg4c" Nov 26 02:01:07 crc kubenswrapper[4766]: I1126 02:01:07.000838 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd82c837-925c-4835-a90f-3d5a873dfbae-config-data\") pod \"bd82c837-925c-4835-a90f-3d5a873dfbae\" (UID: \"bd82c837-925c-4835-a90f-3d5a873dfbae\") " Nov 26 02:01:07 crc kubenswrapper[4766]: I1126 02:01:07.000976 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd82c837-925c-4835-a90f-3d5a873dfbae-combined-ca-bundle\") pod \"bd82c837-925c-4835-a90f-3d5a873dfbae\" (UID: \"bd82c837-925c-4835-a90f-3d5a873dfbae\") " Nov 26 02:01:07 crc kubenswrapper[4766]: I1126 02:01:07.001013 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bd82c837-925c-4835-a90f-3d5a873dfbae-fernet-keys\") pod \"bd82c837-925c-4835-a90f-3d5a873dfbae\" (UID: \"bd82c837-925c-4835-a90f-3d5a873dfbae\") " Nov 26 02:01:07 crc kubenswrapper[4766]: I1126 02:01:07.001123 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zv5r8\" (UniqueName: \"kubernetes.io/projected/bd82c837-925c-4835-a90f-3d5a873dfbae-kube-api-access-zv5r8\") pod \"bd82c837-925c-4835-a90f-3d5a873dfbae\" (UID: \"bd82c837-925c-4835-a90f-3d5a873dfbae\") " Nov 26 02:01:07 crc kubenswrapper[4766]: I1126 02:01:07.007406 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd82c837-925c-4835-a90f-3d5a873dfbae-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "bd82c837-925c-4835-a90f-3d5a873dfbae" (UID: "bd82c837-925c-4835-a90f-3d5a873dfbae"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 02:01:07 crc kubenswrapper[4766]: I1126 02:01:07.007645 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd82c837-925c-4835-a90f-3d5a873dfbae-kube-api-access-zv5r8" (OuterVolumeSpecName: "kube-api-access-zv5r8") pod "bd82c837-925c-4835-a90f-3d5a873dfbae" (UID: "bd82c837-925c-4835-a90f-3d5a873dfbae"). InnerVolumeSpecName "kube-api-access-zv5r8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:01:07 crc kubenswrapper[4766]: I1126 02:01:07.044201 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd82c837-925c-4835-a90f-3d5a873dfbae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd82c837-925c-4835-a90f-3d5a873dfbae" (UID: "bd82c837-925c-4835-a90f-3d5a873dfbae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 02:01:07 crc kubenswrapper[4766]: I1126 02:01:07.089199 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd82c837-925c-4835-a90f-3d5a873dfbae-config-data" (OuterVolumeSpecName: "config-data") pod "bd82c837-925c-4835-a90f-3d5a873dfbae" (UID: "bd82c837-925c-4835-a90f-3d5a873dfbae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 02:01:07 crc kubenswrapper[4766]: I1126 02:01:07.103526 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zv5r8\" (UniqueName: \"kubernetes.io/projected/bd82c837-925c-4835-a90f-3d5a873dfbae-kube-api-access-zv5r8\") on node \"crc\" DevicePath \"\"" Nov 26 02:01:07 crc kubenswrapper[4766]: I1126 02:01:07.103593 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd82c837-925c-4835-a90f-3d5a873dfbae-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 02:01:07 crc kubenswrapper[4766]: I1126 02:01:07.103607 4766 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd82c837-925c-4835-a90f-3d5a873dfbae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 02:01:07 crc kubenswrapper[4766]: I1126 02:01:07.103620 4766 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bd82c837-925c-4835-a90f-3d5a873dfbae-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 26 02:01:07 crc kubenswrapper[4766]: I1126 02:01:07.429004 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29402041-bjg4c" event={"ID":"bd82c837-925c-4835-a90f-3d5a873dfbae","Type":"ContainerDied","Data":"2860b3b62e57c141f8b8309d11df5fa3f7de3444b00ba7bb8e25953c7b674976"} Nov 26 02:01:07 crc kubenswrapper[4766]: I1126 02:01:07.429369 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2860b3b62e57c141f8b8309d11df5fa3f7de3444b00ba7bb8e25953c7b674976" Nov 26 02:01:07 crc kubenswrapper[4766]: I1126 02:01:07.429306 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29402041-bjg4c" Nov 26 02:01:15 crc kubenswrapper[4766]: I1126 02:01:15.828126 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 02:01:15 crc kubenswrapper[4766]: E1126 02:01:15.829767 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:01:30 crc kubenswrapper[4766]: I1126 02:01:30.827452 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 02:01:30 crc kubenswrapper[4766]: E1126 02:01:30.828314 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:01:45 crc kubenswrapper[4766]: I1126 02:01:45.827471 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 02:01:45 crc kubenswrapper[4766]: E1126 02:01:45.828871 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:01:58 crc kubenswrapper[4766]: I1126 02:01:58.826497 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 02:01:58 crc kubenswrapper[4766]: E1126 02:01:58.827267 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:02:12 crc kubenswrapper[4766]: I1126 02:02:12.826759 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 02:02:12 crc kubenswrapper[4766]: E1126 02:02:12.827553 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:02:27 crc kubenswrapper[4766]: I1126 02:02:27.829847 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 02:02:27 crc kubenswrapper[4766]: E1126 02:02:27.830593 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:02:42 crc kubenswrapper[4766]: I1126 02:02:42.827635 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 02:02:42 crc kubenswrapper[4766]: E1126 02:02:42.828582 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:02:56 crc kubenswrapper[4766]: I1126 02:02:56.827365 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 02:02:56 crc kubenswrapper[4766]: E1126 02:02:56.828602 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:03:00 crc kubenswrapper[4766]: I1126 02:03:00.818880 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t9kmw"] Nov 26 02:03:00 crc kubenswrapper[4766]: E1126 02:03:00.820357 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd82c837-925c-4835-a90f-3d5a873dfbae" containerName="keystone-cron" Nov 26 02:03:00 crc kubenswrapper[4766]: I1126 02:03:00.820384 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd82c837-925c-4835-a90f-3d5a873dfbae" containerName="keystone-cron" Nov 26 02:03:00 crc kubenswrapper[4766]: I1126 02:03:00.820815 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd82c837-925c-4835-a90f-3d5a873dfbae" containerName="keystone-cron" Nov 26 02:03:00 crc kubenswrapper[4766]: I1126 02:03:00.824025 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t9kmw" Nov 26 02:03:00 crc kubenswrapper[4766]: I1126 02:03:00.834250 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t9kmw"] Nov 26 02:03:00 crc kubenswrapper[4766]: I1126 02:03:00.846483 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd5b8566-d77c-44d3-af0f-fdeb68471054-utilities\") pod \"certified-operators-t9kmw\" (UID: \"cd5b8566-d77c-44d3-af0f-fdeb68471054\") " pod="openshift-marketplace/certified-operators-t9kmw" Nov 26 02:03:00 crc kubenswrapper[4766]: I1126 02:03:00.846839 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd5b8566-d77c-44d3-af0f-fdeb68471054-catalog-content\") pod \"certified-operators-t9kmw\" (UID: \"cd5b8566-d77c-44d3-af0f-fdeb68471054\") " pod="openshift-marketplace/certified-operators-t9kmw" Nov 26 02:03:00 crc kubenswrapper[4766]: I1126 02:03:00.846945 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzsnp\" (UniqueName: \"kubernetes.io/projected/cd5b8566-d77c-44d3-af0f-fdeb68471054-kube-api-access-rzsnp\") pod \"certified-operators-t9kmw\" (UID: \"cd5b8566-d77c-44d3-af0f-fdeb68471054\") " pod="openshift-marketplace/certified-operators-t9kmw" Nov 26 02:03:00 crc kubenswrapper[4766]: I1126 02:03:00.948062 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd5b8566-d77c-44d3-af0f-fdeb68471054-catalog-content\") pod \"certified-operators-t9kmw\" (UID: \"cd5b8566-d77c-44d3-af0f-fdeb68471054\") " pod="openshift-marketplace/certified-operators-t9kmw" Nov 26 02:03:00 crc kubenswrapper[4766]: I1126 02:03:00.948351 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzsnp\" (UniqueName: \"kubernetes.io/projected/cd5b8566-d77c-44d3-af0f-fdeb68471054-kube-api-access-rzsnp\") pod \"certified-operators-t9kmw\" (UID: \"cd5b8566-d77c-44d3-af0f-fdeb68471054\") " pod="openshift-marketplace/certified-operators-t9kmw" Nov 26 02:03:00 crc kubenswrapper[4766]: I1126 02:03:00.948506 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd5b8566-d77c-44d3-af0f-fdeb68471054-utilities\") pod \"certified-operators-t9kmw\" (UID: \"cd5b8566-d77c-44d3-af0f-fdeb68471054\") " pod="openshift-marketplace/certified-operators-t9kmw" Nov 26 02:03:00 crc kubenswrapper[4766]: I1126 02:03:00.948796 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd5b8566-d77c-44d3-af0f-fdeb68471054-catalog-content\") pod \"certified-operators-t9kmw\" (UID: \"cd5b8566-d77c-44d3-af0f-fdeb68471054\") " pod="openshift-marketplace/certified-operators-t9kmw" Nov 26 02:03:00 crc kubenswrapper[4766]: I1126 02:03:00.949012 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd5b8566-d77c-44d3-af0f-fdeb68471054-utilities\") pod \"certified-operators-t9kmw\" (UID: \"cd5b8566-d77c-44d3-af0f-fdeb68471054\") " pod="openshift-marketplace/certified-operators-t9kmw" Nov 26 02:03:00 crc kubenswrapper[4766]: I1126 02:03:00.968441 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzsnp\" (UniqueName: \"kubernetes.io/projected/cd5b8566-d77c-44d3-af0f-fdeb68471054-kube-api-access-rzsnp\") pod \"certified-operators-t9kmw\" (UID: \"cd5b8566-d77c-44d3-af0f-fdeb68471054\") " pod="openshift-marketplace/certified-operators-t9kmw" Nov 26 02:03:01 crc kubenswrapper[4766]: I1126 02:03:01.174439 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t9kmw" Nov 26 02:03:01 crc kubenswrapper[4766]: I1126 02:03:01.741228 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t9kmw"] Nov 26 02:03:01 crc kubenswrapper[4766]: W1126 02:03:01.744381 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd5b8566_d77c_44d3_af0f_fdeb68471054.slice/crio-0a62e21f0c853a3808ccf20907e80a748692ce67a97c42b0e14226ec3ada9d24 WatchSource:0}: Error finding container 0a62e21f0c853a3808ccf20907e80a748692ce67a97c42b0e14226ec3ada9d24: Status 404 returned error can't find the container with id 0a62e21f0c853a3808ccf20907e80a748692ce67a97c42b0e14226ec3ada9d24 Nov 26 02:03:01 crc kubenswrapper[4766]: I1126 02:03:01.929069 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9kmw" event={"ID":"cd5b8566-d77c-44d3-af0f-fdeb68471054","Type":"ContainerStarted","Data":"0a62e21f0c853a3808ccf20907e80a748692ce67a97c42b0e14226ec3ada9d24"} Nov 26 02:03:02 crc kubenswrapper[4766]: I1126 02:03:02.943872 4766 generic.go:334] "Generic (PLEG): container finished" podID="cd5b8566-d77c-44d3-af0f-fdeb68471054" containerID="bb677ddde3f79124b6cfb0c3d7e9e1afc8ee0758a73a47530c95473a6cf5924c" exitCode=0 Nov 26 02:03:02 crc kubenswrapper[4766]: I1126 02:03:02.943912 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9kmw" event={"ID":"cd5b8566-d77c-44d3-af0f-fdeb68471054","Type":"ContainerDied","Data":"bb677ddde3f79124b6cfb0c3d7e9e1afc8ee0758a73a47530c95473a6cf5924c"} Nov 26 02:03:02 crc kubenswrapper[4766]: I1126 02:03:02.946750 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 02:03:03 crc kubenswrapper[4766]: I1126 02:03:03.956633 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9kmw" event={"ID":"cd5b8566-d77c-44d3-af0f-fdeb68471054","Type":"ContainerStarted","Data":"7fcc6cf44cd41c3e37b953e1bb5d71268d27586a16e7ce31de6553185d420b1c"} Nov 26 02:03:05 crc kubenswrapper[4766]: I1126 02:03:05.982711 4766 generic.go:334] "Generic (PLEG): container finished" podID="cd5b8566-d77c-44d3-af0f-fdeb68471054" containerID="7fcc6cf44cd41c3e37b953e1bb5d71268d27586a16e7ce31de6553185d420b1c" exitCode=0 Nov 26 02:03:05 crc kubenswrapper[4766]: I1126 02:03:05.982818 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9kmw" event={"ID":"cd5b8566-d77c-44d3-af0f-fdeb68471054","Type":"ContainerDied","Data":"7fcc6cf44cd41c3e37b953e1bb5d71268d27586a16e7ce31de6553185d420b1c"} Nov 26 02:03:06 crc kubenswrapper[4766]: I1126 02:03:06.996991 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9kmw" event={"ID":"cd5b8566-d77c-44d3-af0f-fdeb68471054","Type":"ContainerStarted","Data":"e29926bdece71f94ac2bbb01a11d3e5620051bccc8cafa1847e2173838de6613"} Nov 26 02:03:07 crc kubenswrapper[4766]: I1126 02:03:07.051606 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t9kmw" podStartSLOduration=3.534937119 podStartE2EDuration="7.051582423s" podCreationTimestamp="2025-11-26 02:03:00 +0000 UTC" firstStartedPulling="2025-11-26 02:03:02.946546365 +0000 UTC m=+5963.795316795" lastFinishedPulling="2025-11-26 02:03:06.463191629 +0000 UTC m=+5967.311962099" observedRunningTime="2025-11-26 02:03:07.015170437 +0000 UTC m=+5967.863940867" watchObservedRunningTime="2025-11-26 02:03:07.051582423 +0000 UTC m=+5967.900352863" Nov 26 02:03:10 crc kubenswrapper[4766]: I1126 02:03:10.826795 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 02:03:10 crc kubenswrapper[4766]: E1126 02:03:10.829211 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:03:11 crc kubenswrapper[4766]: I1126 02:03:11.176466 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t9kmw" Nov 26 02:03:11 crc kubenswrapper[4766]: I1126 02:03:11.176557 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t9kmw" Nov 26 02:03:11 crc kubenswrapper[4766]: I1126 02:03:11.270516 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t9kmw" Nov 26 02:03:12 crc kubenswrapper[4766]: I1126 02:03:12.135159 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t9kmw" Nov 26 02:03:12 crc kubenswrapper[4766]: I1126 02:03:12.197315 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t9kmw"] Nov 26 02:03:14 crc kubenswrapper[4766]: I1126 02:03:14.079308 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t9kmw" podUID="cd5b8566-d77c-44d3-af0f-fdeb68471054" containerName="registry-server" containerID="cri-o://e29926bdece71f94ac2bbb01a11d3e5620051bccc8cafa1847e2173838de6613" gracePeriod=2 Nov 26 02:03:14 crc kubenswrapper[4766]: I1126 02:03:14.676713 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t9kmw" Nov 26 02:03:14 crc kubenswrapper[4766]: I1126 02:03:14.829161 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd5b8566-d77c-44d3-af0f-fdeb68471054-utilities\") pod \"cd5b8566-d77c-44d3-af0f-fdeb68471054\" (UID: \"cd5b8566-d77c-44d3-af0f-fdeb68471054\") " Nov 26 02:03:14 crc kubenswrapper[4766]: I1126 02:03:14.829527 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzsnp\" (UniqueName: \"kubernetes.io/projected/cd5b8566-d77c-44d3-af0f-fdeb68471054-kube-api-access-rzsnp\") pod \"cd5b8566-d77c-44d3-af0f-fdeb68471054\" (UID: \"cd5b8566-d77c-44d3-af0f-fdeb68471054\") " Nov 26 02:03:14 crc kubenswrapper[4766]: I1126 02:03:14.829706 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd5b8566-d77c-44d3-af0f-fdeb68471054-catalog-content\") pod \"cd5b8566-d77c-44d3-af0f-fdeb68471054\" (UID: \"cd5b8566-d77c-44d3-af0f-fdeb68471054\") " Nov 26 02:03:14 crc kubenswrapper[4766]: I1126 02:03:14.829904 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd5b8566-d77c-44d3-af0f-fdeb68471054-utilities" (OuterVolumeSpecName: "utilities") pod "cd5b8566-d77c-44d3-af0f-fdeb68471054" (UID: "cd5b8566-d77c-44d3-af0f-fdeb68471054"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:03:14 crc kubenswrapper[4766]: I1126 02:03:14.830643 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd5b8566-d77c-44d3-af0f-fdeb68471054-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 02:03:14 crc kubenswrapper[4766]: I1126 02:03:14.840920 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd5b8566-d77c-44d3-af0f-fdeb68471054-kube-api-access-rzsnp" (OuterVolumeSpecName: "kube-api-access-rzsnp") pod "cd5b8566-d77c-44d3-af0f-fdeb68471054" (UID: "cd5b8566-d77c-44d3-af0f-fdeb68471054"). InnerVolumeSpecName "kube-api-access-rzsnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:03:14 crc kubenswrapper[4766]: I1126 02:03:14.877842 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd5b8566-d77c-44d3-af0f-fdeb68471054-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd5b8566-d77c-44d3-af0f-fdeb68471054" (UID: "cd5b8566-d77c-44d3-af0f-fdeb68471054"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:03:14 crc kubenswrapper[4766]: I1126 02:03:14.935501 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzsnp\" (UniqueName: \"kubernetes.io/projected/cd5b8566-d77c-44d3-af0f-fdeb68471054-kube-api-access-rzsnp\") on node \"crc\" DevicePath \"\"" Nov 26 02:03:14 crc kubenswrapper[4766]: I1126 02:03:14.935567 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd5b8566-d77c-44d3-af0f-fdeb68471054-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 02:03:15 crc kubenswrapper[4766]: I1126 02:03:15.094603 4766 generic.go:334] "Generic (PLEG): container finished" podID="cd5b8566-d77c-44d3-af0f-fdeb68471054" containerID="e29926bdece71f94ac2bbb01a11d3e5620051bccc8cafa1847e2173838de6613" exitCode=0 Nov 26 02:03:15 crc kubenswrapper[4766]: I1126 02:03:15.094648 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9kmw" event={"ID":"cd5b8566-d77c-44d3-af0f-fdeb68471054","Type":"ContainerDied","Data":"e29926bdece71f94ac2bbb01a11d3e5620051bccc8cafa1847e2173838de6613"} Nov 26 02:03:15 crc kubenswrapper[4766]: I1126 02:03:15.094701 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9kmw" event={"ID":"cd5b8566-d77c-44d3-af0f-fdeb68471054","Type":"ContainerDied","Data":"0a62e21f0c853a3808ccf20907e80a748692ce67a97c42b0e14226ec3ada9d24"} Nov 26 02:03:15 crc kubenswrapper[4766]: I1126 02:03:15.094721 4766 scope.go:117] "RemoveContainer" containerID="e29926bdece71f94ac2bbb01a11d3e5620051bccc8cafa1847e2173838de6613" Nov 26 02:03:15 crc kubenswrapper[4766]: I1126 02:03:15.094891 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t9kmw" Nov 26 02:03:15 crc kubenswrapper[4766]: I1126 02:03:15.169708 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t9kmw"] Nov 26 02:03:15 crc kubenswrapper[4766]: I1126 02:03:15.176252 4766 scope.go:117] "RemoveContainer" containerID="7fcc6cf44cd41c3e37b953e1bb5d71268d27586a16e7ce31de6553185d420b1c" Nov 26 02:03:15 crc kubenswrapper[4766]: I1126 02:03:15.189449 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t9kmw"] Nov 26 02:03:15 crc kubenswrapper[4766]: I1126 02:03:15.203563 4766 scope.go:117] "RemoveContainer" containerID="bb677ddde3f79124b6cfb0c3d7e9e1afc8ee0758a73a47530c95473a6cf5924c" Nov 26 02:03:15 crc kubenswrapper[4766]: I1126 02:03:15.282478 4766 scope.go:117] "RemoveContainer" containerID="e29926bdece71f94ac2bbb01a11d3e5620051bccc8cafa1847e2173838de6613" Nov 26 02:03:15 crc kubenswrapper[4766]: E1126 02:03:15.283046 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e29926bdece71f94ac2bbb01a11d3e5620051bccc8cafa1847e2173838de6613\": container with ID starting with e29926bdece71f94ac2bbb01a11d3e5620051bccc8cafa1847e2173838de6613 not found: ID does not exist" containerID="e29926bdece71f94ac2bbb01a11d3e5620051bccc8cafa1847e2173838de6613" Nov 26 02:03:15 crc kubenswrapper[4766]: I1126 02:03:15.283112 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e29926bdece71f94ac2bbb01a11d3e5620051bccc8cafa1847e2173838de6613"} err="failed to get container status \"e29926bdece71f94ac2bbb01a11d3e5620051bccc8cafa1847e2173838de6613\": rpc error: code = NotFound desc = could not find container \"e29926bdece71f94ac2bbb01a11d3e5620051bccc8cafa1847e2173838de6613\": container with ID starting with e29926bdece71f94ac2bbb01a11d3e5620051bccc8cafa1847e2173838de6613 not found: ID does not exist" Nov 26 02:03:15 crc kubenswrapper[4766]: I1126 02:03:15.283154 4766 scope.go:117] "RemoveContainer" containerID="7fcc6cf44cd41c3e37b953e1bb5d71268d27586a16e7ce31de6553185d420b1c" Nov 26 02:03:15 crc kubenswrapper[4766]: E1126 02:03:15.283633 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fcc6cf44cd41c3e37b953e1bb5d71268d27586a16e7ce31de6553185d420b1c\": container with ID starting with 7fcc6cf44cd41c3e37b953e1bb5d71268d27586a16e7ce31de6553185d420b1c not found: ID does not exist" containerID="7fcc6cf44cd41c3e37b953e1bb5d71268d27586a16e7ce31de6553185d420b1c" Nov 26 02:03:15 crc kubenswrapper[4766]: I1126 02:03:15.283707 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fcc6cf44cd41c3e37b953e1bb5d71268d27586a16e7ce31de6553185d420b1c"} err="failed to get container status \"7fcc6cf44cd41c3e37b953e1bb5d71268d27586a16e7ce31de6553185d420b1c\": rpc error: code = NotFound desc = could not find container \"7fcc6cf44cd41c3e37b953e1bb5d71268d27586a16e7ce31de6553185d420b1c\": container with ID starting with 7fcc6cf44cd41c3e37b953e1bb5d71268d27586a16e7ce31de6553185d420b1c not found: ID does not exist" Nov 26 02:03:15 crc kubenswrapper[4766]: I1126 02:03:15.283743 4766 scope.go:117] "RemoveContainer" containerID="bb677ddde3f79124b6cfb0c3d7e9e1afc8ee0758a73a47530c95473a6cf5924c" Nov 26 02:03:15 crc kubenswrapper[4766]: E1126 02:03:15.284116 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb677ddde3f79124b6cfb0c3d7e9e1afc8ee0758a73a47530c95473a6cf5924c\": container with ID starting with bb677ddde3f79124b6cfb0c3d7e9e1afc8ee0758a73a47530c95473a6cf5924c not found: ID does not exist" containerID="bb677ddde3f79124b6cfb0c3d7e9e1afc8ee0758a73a47530c95473a6cf5924c" Nov 26 02:03:15 crc kubenswrapper[4766]: I1126 02:03:15.284209 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb677ddde3f79124b6cfb0c3d7e9e1afc8ee0758a73a47530c95473a6cf5924c"} err="failed to get container status \"bb677ddde3f79124b6cfb0c3d7e9e1afc8ee0758a73a47530c95473a6cf5924c\": rpc error: code = NotFound desc = could not find container \"bb677ddde3f79124b6cfb0c3d7e9e1afc8ee0758a73a47530c95473a6cf5924c\": container with ID starting with bb677ddde3f79124b6cfb0c3d7e9e1afc8ee0758a73a47530c95473a6cf5924c not found: ID does not exist" Nov 26 02:03:15 crc kubenswrapper[4766]: I1126 02:03:15.850001 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd5b8566-d77c-44d3-af0f-fdeb68471054" path="/var/lib/kubelet/pods/cd5b8566-d77c-44d3-af0f-fdeb68471054/volumes" Nov 26 02:03:24 crc kubenswrapper[4766]: I1126 02:03:24.829785 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 02:03:24 crc kubenswrapper[4766]: E1126 02:03:24.830413 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:03:35 crc kubenswrapper[4766]: I1126 02:03:35.827909 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 02:03:35 crc kubenswrapper[4766]: E1126 02:03:35.829238 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:03:50 crc kubenswrapper[4766]: I1126 02:03:50.827771 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 02:03:51 crc kubenswrapper[4766]: I1126 02:03:51.601378 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"e70127ee22b463f69db81a0069e5277e44a8b961e37fbeac182f3a366ae58c79"} Nov 26 02:03:59 crc kubenswrapper[4766]: I1126 02:03:59.917159 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qsx26"] Nov 26 02:03:59 crc kubenswrapper[4766]: E1126 02:03:59.918557 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd5b8566-d77c-44d3-af0f-fdeb68471054" containerName="extract-content" Nov 26 02:03:59 crc kubenswrapper[4766]: I1126 02:03:59.918801 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd5b8566-d77c-44d3-af0f-fdeb68471054" containerName="extract-content" Nov 26 02:03:59 crc kubenswrapper[4766]: E1126 02:03:59.918841 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd5b8566-d77c-44d3-af0f-fdeb68471054" containerName="extract-utilities" Nov 26 02:03:59 crc kubenswrapper[4766]: I1126 02:03:59.918854 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd5b8566-d77c-44d3-af0f-fdeb68471054" containerName="extract-utilities" Nov 26 02:03:59 crc kubenswrapper[4766]: E1126 02:03:59.918896 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd5b8566-d77c-44d3-af0f-fdeb68471054" containerName="registry-server" Nov 26 02:03:59 crc kubenswrapper[4766]: I1126 02:03:59.918910 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd5b8566-d77c-44d3-af0f-fdeb68471054" containerName="registry-server" Nov 26 02:03:59 crc kubenswrapper[4766]: I1126 02:03:59.919385 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd5b8566-d77c-44d3-af0f-fdeb68471054" containerName="registry-server" Nov 26 02:03:59 crc kubenswrapper[4766]: I1126 02:03:59.922509 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qsx26" Nov 26 02:03:59 crc kubenswrapper[4766]: I1126 02:03:59.938330 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qsx26"] Nov 26 02:04:00 crc kubenswrapper[4766]: I1126 02:04:00.003798 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7mxb\" (UniqueName: \"kubernetes.io/projected/d4a65e75-b0d1-4ca6-a36c-58aab4b638b0-kube-api-access-q7mxb\") pod \"community-operators-qsx26\" (UID: \"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0\") " pod="openshift-marketplace/community-operators-qsx26" Nov 26 02:04:00 crc kubenswrapper[4766]: I1126 02:04:00.003968 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a65e75-b0d1-4ca6-a36c-58aab4b638b0-utilities\") pod \"community-operators-qsx26\" (UID: \"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0\") " pod="openshift-marketplace/community-operators-qsx26" Nov 26 02:04:00 crc kubenswrapper[4766]: I1126 02:04:00.004036 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a65e75-b0d1-4ca6-a36c-58aab4b638b0-catalog-content\") pod \"community-operators-qsx26\" (UID: \"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0\") " pod="openshift-marketplace/community-operators-qsx26" Nov 26 02:04:00 crc kubenswrapper[4766]: I1126 02:04:00.107386 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7mxb\" (UniqueName: \"kubernetes.io/projected/d4a65e75-b0d1-4ca6-a36c-58aab4b638b0-kube-api-access-q7mxb\") pod \"community-operators-qsx26\" (UID: \"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0\") " pod="openshift-marketplace/community-operators-qsx26" Nov 26 02:04:00 crc kubenswrapper[4766]: I1126 02:04:00.107538 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a65e75-b0d1-4ca6-a36c-58aab4b638b0-utilities\") pod \"community-operators-qsx26\" (UID: \"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0\") " pod="openshift-marketplace/community-operators-qsx26" Nov 26 02:04:00 crc kubenswrapper[4766]: I1126 02:04:00.107591 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a65e75-b0d1-4ca6-a36c-58aab4b638b0-catalog-content\") pod \"community-operators-qsx26\" (UID: \"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0\") " pod="openshift-marketplace/community-operators-qsx26" Nov 26 02:04:00 crc kubenswrapper[4766]: I1126 02:04:00.108495 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a65e75-b0d1-4ca6-a36c-58aab4b638b0-utilities\") pod \"community-operators-qsx26\" (UID: \"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0\") " pod="openshift-marketplace/community-operators-qsx26" Nov 26 02:04:00 crc kubenswrapper[4766]: I1126 02:04:00.108588 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a65e75-b0d1-4ca6-a36c-58aab4b638b0-catalog-content\") pod \"community-operators-qsx26\" (UID: \"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0\") " pod="openshift-marketplace/community-operators-qsx26" Nov 26 02:04:00 crc kubenswrapper[4766]: I1126 02:04:00.133170 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7mxb\" (UniqueName: \"kubernetes.io/projected/d4a65e75-b0d1-4ca6-a36c-58aab4b638b0-kube-api-access-q7mxb\") pod \"community-operators-qsx26\" (UID: \"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0\") " pod="openshift-marketplace/community-operators-qsx26" Nov 26 02:04:00 crc kubenswrapper[4766]: I1126 02:04:00.307019 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qsx26" Nov 26 02:04:00 crc kubenswrapper[4766]: I1126 02:04:00.930988 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qsx26"] Nov 26 02:04:01 crc kubenswrapper[4766]: I1126 02:04:01.725770 4766 generic.go:334] "Generic (PLEG): container finished" podID="d4a65e75-b0d1-4ca6-a36c-58aab4b638b0" containerID="39ec41e6535bc6d6ac6dc336b346a84ef1aef7bc5321058c2ee5f3b4dbe91cbb" exitCode=0 Nov 26 02:04:01 crc kubenswrapper[4766]: I1126 02:04:01.725869 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qsx26" event={"ID":"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0","Type":"ContainerDied","Data":"39ec41e6535bc6d6ac6dc336b346a84ef1aef7bc5321058c2ee5f3b4dbe91cbb"} Nov 26 02:04:01 crc kubenswrapper[4766]: I1126 02:04:01.726290 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qsx26" event={"ID":"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0","Type":"ContainerStarted","Data":"f29625e7c64ba73f0558f83b17a96a016b4cc2e42caa845f21bc89cf6683fbc1"} Nov 26 02:04:03 crc kubenswrapper[4766]: I1126 02:04:03.757846 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qsx26" event={"ID":"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0","Type":"ContainerStarted","Data":"ee0b14822ef844c30d525e9d4ff9f90b08e7a3330b8e4488f39673564f672a3b"} Nov 26 02:04:04 crc kubenswrapper[4766]: I1126 02:04:04.775161 4766 generic.go:334] "Generic (PLEG): container finished" podID="d4a65e75-b0d1-4ca6-a36c-58aab4b638b0" containerID="ee0b14822ef844c30d525e9d4ff9f90b08e7a3330b8e4488f39673564f672a3b" exitCode=0 Nov 26 02:04:04 crc kubenswrapper[4766]: I1126 02:04:04.775228 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qsx26" event={"ID":"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0","Type":"ContainerDied","Data":"ee0b14822ef844c30d525e9d4ff9f90b08e7a3330b8e4488f39673564f672a3b"} Nov 26 02:04:05 crc kubenswrapper[4766]: I1126 02:04:05.808519 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qsx26" event={"ID":"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0","Type":"ContainerStarted","Data":"aa0ebc67dc03343e1d06dcb2bdaa3ef41910e37c4f38817755af26925f5bd7cc"} Nov 26 02:04:05 crc kubenswrapper[4766]: I1126 02:04:05.837510 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qsx26" podStartSLOduration=3.370473838 podStartE2EDuration="6.837491325s" podCreationTimestamp="2025-11-26 02:03:59 +0000 UTC" firstStartedPulling="2025-11-26 02:04:01.728057241 +0000 UTC m=+6022.576827671" lastFinishedPulling="2025-11-26 02:04:05.195074728 +0000 UTC m=+6026.043845158" observedRunningTime="2025-11-26 02:04:05.831710485 +0000 UTC m=+6026.680480925" watchObservedRunningTime="2025-11-26 02:04:05.837491325 +0000 UTC m=+6026.686261765" Nov 26 02:04:10 crc kubenswrapper[4766]: I1126 02:04:10.307790 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qsx26" Nov 26 02:04:10 crc kubenswrapper[4766]: I1126 02:04:10.310206 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qsx26" Nov 26 02:04:10 crc kubenswrapper[4766]: I1126 02:04:10.388190 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qsx26" Nov 26 02:04:10 crc kubenswrapper[4766]: I1126 02:04:10.975440 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qsx26" Nov 26 02:04:11 crc kubenswrapper[4766]: I1126 02:04:11.038727 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qsx26"] Nov 26 02:04:12 crc kubenswrapper[4766]: I1126 02:04:12.932813 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qsx26" podUID="d4a65e75-b0d1-4ca6-a36c-58aab4b638b0" containerName="registry-server" containerID="cri-o://aa0ebc67dc03343e1d06dcb2bdaa3ef41910e37c4f38817755af26925f5bd7cc" gracePeriod=2 Nov 26 02:04:13 crc kubenswrapper[4766]: I1126 02:04:13.562907 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qsx26" Nov 26 02:04:13 crc kubenswrapper[4766]: I1126 02:04:13.672054 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7mxb\" (UniqueName: \"kubernetes.io/projected/d4a65e75-b0d1-4ca6-a36c-58aab4b638b0-kube-api-access-q7mxb\") pod \"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0\" (UID: \"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0\") " Nov 26 02:04:13 crc kubenswrapper[4766]: I1126 02:04:13.672157 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a65e75-b0d1-4ca6-a36c-58aab4b638b0-utilities\") pod \"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0\" (UID: \"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0\") " Nov 26 02:04:13 crc kubenswrapper[4766]: I1126 02:04:13.672211 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a65e75-b0d1-4ca6-a36c-58aab4b638b0-catalog-content\") pod \"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0\" (UID: \"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0\") " Nov 26 02:04:13 crc kubenswrapper[4766]: I1126 02:04:13.673357 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4a65e75-b0d1-4ca6-a36c-58aab4b638b0-utilities" (OuterVolumeSpecName: "utilities") pod "d4a65e75-b0d1-4ca6-a36c-58aab4b638b0" (UID: "d4a65e75-b0d1-4ca6-a36c-58aab4b638b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:04:13 crc kubenswrapper[4766]: I1126 02:04:13.678743 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4a65e75-b0d1-4ca6-a36c-58aab4b638b0-kube-api-access-q7mxb" (OuterVolumeSpecName: "kube-api-access-q7mxb") pod "d4a65e75-b0d1-4ca6-a36c-58aab4b638b0" (UID: "d4a65e75-b0d1-4ca6-a36c-58aab4b638b0"). InnerVolumeSpecName "kube-api-access-q7mxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:04:13 crc kubenswrapper[4766]: I1126 02:04:13.688134 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7mxb\" (UniqueName: \"kubernetes.io/projected/d4a65e75-b0d1-4ca6-a36c-58aab4b638b0-kube-api-access-q7mxb\") on node \"crc\" DevicePath \"\"" Nov 26 02:04:13 crc kubenswrapper[4766]: I1126 02:04:13.688185 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a65e75-b0d1-4ca6-a36c-58aab4b638b0-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 02:04:13 crc kubenswrapper[4766]: I1126 02:04:13.749802 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4a65e75-b0d1-4ca6-a36c-58aab4b638b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4a65e75-b0d1-4ca6-a36c-58aab4b638b0" (UID: "d4a65e75-b0d1-4ca6-a36c-58aab4b638b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:04:13 crc kubenswrapper[4766]: I1126 02:04:13.790293 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a65e75-b0d1-4ca6-a36c-58aab4b638b0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 02:04:13 crc kubenswrapper[4766]: I1126 02:04:13.945041 4766 generic.go:334] "Generic (PLEG): container finished" podID="d4a65e75-b0d1-4ca6-a36c-58aab4b638b0" containerID="aa0ebc67dc03343e1d06dcb2bdaa3ef41910e37c4f38817755af26925f5bd7cc" exitCode=0 Nov 26 02:04:13 crc kubenswrapper[4766]: I1126 02:04:13.945099 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qsx26" event={"ID":"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0","Type":"ContainerDied","Data":"aa0ebc67dc03343e1d06dcb2bdaa3ef41910e37c4f38817755af26925f5bd7cc"} Nov 26 02:04:13 crc kubenswrapper[4766]: I1126 02:04:13.945135 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qsx26" event={"ID":"d4a65e75-b0d1-4ca6-a36c-58aab4b638b0","Type":"ContainerDied","Data":"f29625e7c64ba73f0558f83b17a96a016b4cc2e42caa845f21bc89cf6683fbc1"} Nov 26 02:04:13 crc kubenswrapper[4766]: I1126 02:04:13.945162 4766 scope.go:117] "RemoveContainer" containerID="aa0ebc67dc03343e1d06dcb2bdaa3ef41910e37c4f38817755af26925f5bd7cc" Nov 26 02:04:13 crc kubenswrapper[4766]: I1126 02:04:13.946383 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qsx26" Nov 26 02:04:13 crc kubenswrapper[4766]: I1126 02:04:13.970543 4766 scope.go:117] "RemoveContainer" containerID="ee0b14822ef844c30d525e9d4ff9f90b08e7a3330b8e4488f39673564f672a3b" Nov 26 02:04:13 crc kubenswrapper[4766]: I1126 02:04:13.970702 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qsx26"] Nov 26 02:04:13 crc kubenswrapper[4766]: I1126 02:04:13.982727 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qsx26"] Nov 26 02:04:14 crc kubenswrapper[4766]: I1126 02:04:14.002351 4766 scope.go:117] "RemoveContainer" containerID="39ec41e6535bc6d6ac6dc336b346a84ef1aef7bc5321058c2ee5f3b4dbe91cbb" Nov 26 02:04:14 crc kubenswrapper[4766]: I1126 02:04:14.054323 4766 scope.go:117] "RemoveContainer" containerID="aa0ebc67dc03343e1d06dcb2bdaa3ef41910e37c4f38817755af26925f5bd7cc" Nov 26 02:04:14 crc kubenswrapper[4766]: E1126 02:04:14.054788 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa0ebc67dc03343e1d06dcb2bdaa3ef41910e37c4f38817755af26925f5bd7cc\": container with ID starting with aa0ebc67dc03343e1d06dcb2bdaa3ef41910e37c4f38817755af26925f5bd7cc not found: ID does not exist" containerID="aa0ebc67dc03343e1d06dcb2bdaa3ef41910e37c4f38817755af26925f5bd7cc" Nov 26 02:04:14 crc kubenswrapper[4766]: I1126 02:04:14.054819 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa0ebc67dc03343e1d06dcb2bdaa3ef41910e37c4f38817755af26925f5bd7cc"} err="failed to get container status \"aa0ebc67dc03343e1d06dcb2bdaa3ef41910e37c4f38817755af26925f5bd7cc\": rpc error: code = NotFound desc = could not find container \"aa0ebc67dc03343e1d06dcb2bdaa3ef41910e37c4f38817755af26925f5bd7cc\": container with ID starting with aa0ebc67dc03343e1d06dcb2bdaa3ef41910e37c4f38817755af26925f5bd7cc not found: ID does not exist" Nov 26 02:04:14 crc kubenswrapper[4766]: I1126 02:04:14.054837 4766 scope.go:117] "RemoveContainer" containerID="ee0b14822ef844c30d525e9d4ff9f90b08e7a3330b8e4488f39673564f672a3b" Nov 26 02:04:14 crc kubenswrapper[4766]: E1126 02:04:14.055046 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee0b14822ef844c30d525e9d4ff9f90b08e7a3330b8e4488f39673564f672a3b\": container with ID starting with ee0b14822ef844c30d525e9d4ff9f90b08e7a3330b8e4488f39673564f672a3b not found: ID does not exist" containerID="ee0b14822ef844c30d525e9d4ff9f90b08e7a3330b8e4488f39673564f672a3b" Nov 26 02:04:14 crc kubenswrapper[4766]: I1126 02:04:14.055064 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee0b14822ef844c30d525e9d4ff9f90b08e7a3330b8e4488f39673564f672a3b"} err="failed to get container status \"ee0b14822ef844c30d525e9d4ff9f90b08e7a3330b8e4488f39673564f672a3b\": rpc error: code = NotFound desc = could not find container \"ee0b14822ef844c30d525e9d4ff9f90b08e7a3330b8e4488f39673564f672a3b\": container with ID starting with ee0b14822ef844c30d525e9d4ff9f90b08e7a3330b8e4488f39673564f672a3b not found: ID does not exist" Nov 26 02:04:14 crc kubenswrapper[4766]: I1126 02:04:14.055075 4766 scope.go:117] "RemoveContainer" containerID="39ec41e6535bc6d6ac6dc336b346a84ef1aef7bc5321058c2ee5f3b4dbe91cbb" Nov 26 02:04:14 crc kubenswrapper[4766]: E1126 02:04:14.055203 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39ec41e6535bc6d6ac6dc336b346a84ef1aef7bc5321058c2ee5f3b4dbe91cbb\": container with ID starting with 39ec41e6535bc6d6ac6dc336b346a84ef1aef7bc5321058c2ee5f3b4dbe91cbb not found: ID does not exist" containerID="39ec41e6535bc6d6ac6dc336b346a84ef1aef7bc5321058c2ee5f3b4dbe91cbb" Nov 26 02:04:14 crc kubenswrapper[4766]: I1126 02:04:14.055216 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39ec41e6535bc6d6ac6dc336b346a84ef1aef7bc5321058c2ee5f3b4dbe91cbb"} err="failed to get container status \"39ec41e6535bc6d6ac6dc336b346a84ef1aef7bc5321058c2ee5f3b4dbe91cbb\": rpc error: code = NotFound desc = could not find container \"39ec41e6535bc6d6ac6dc336b346a84ef1aef7bc5321058c2ee5f3b4dbe91cbb\": container with ID starting with 39ec41e6535bc6d6ac6dc336b346a84ef1aef7bc5321058c2ee5f3b4dbe91cbb not found: ID does not exist" Nov 26 02:04:15 crc kubenswrapper[4766]: I1126 02:04:15.841489 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4a65e75-b0d1-4ca6-a36c-58aab4b638b0" path="/var/lib/kubelet/pods/d4a65e75-b0d1-4ca6-a36c-58aab4b638b0/volumes" Nov 26 02:06:11 crc kubenswrapper[4766]: I1126 02:06:11.479463 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:06:11 crc kubenswrapper[4766]: I1126 02:06:11.480425 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:06:41 crc kubenswrapper[4766]: I1126 02:06:41.479725 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:06:41 crc kubenswrapper[4766]: I1126 02:06:41.481043 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:07:11 crc kubenswrapper[4766]: I1126 02:07:11.480170 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:07:11 crc kubenswrapper[4766]: I1126 02:07:11.481084 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:07:11 crc kubenswrapper[4766]: I1126 02:07:11.481168 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 02:07:11 crc kubenswrapper[4766]: I1126 02:07:11.482844 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e70127ee22b463f69db81a0069e5277e44a8b961e37fbeac182f3a366ae58c79"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 02:07:11 crc kubenswrapper[4766]: I1126 02:07:11.482994 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://e70127ee22b463f69db81a0069e5277e44a8b961e37fbeac182f3a366ae58c79" gracePeriod=600 Nov 26 02:07:12 crc kubenswrapper[4766]: I1126 02:07:12.363901 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="e70127ee22b463f69db81a0069e5277e44a8b961e37fbeac182f3a366ae58c79" exitCode=0 Nov 26 02:07:12 crc kubenswrapper[4766]: I1126 02:07:12.363987 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"e70127ee22b463f69db81a0069e5277e44a8b961e37fbeac182f3a366ae58c79"} Nov 26 02:07:12 crc kubenswrapper[4766]: I1126 02:07:12.364261 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b"} Nov 26 02:07:12 crc kubenswrapper[4766]: I1126 02:07:12.364284 4766 scope.go:117] "RemoveContainer" containerID="a86e085eea6de161a4eb7311e29d3e68900108076a440b440953ce703b3ac135" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.785306 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 26 02:08:10 crc kubenswrapper[4766]: E1126 02:08:10.786491 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a65e75-b0d1-4ca6-a36c-58aab4b638b0" containerName="extract-content" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.786509 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a65e75-b0d1-4ca6-a36c-58aab4b638b0" containerName="extract-content" Nov 26 02:08:10 crc kubenswrapper[4766]: E1126 02:08:10.786536 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a65e75-b0d1-4ca6-a36c-58aab4b638b0" containerName="registry-server" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.786544 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a65e75-b0d1-4ca6-a36c-58aab4b638b0" containerName="registry-server" Nov 26 02:08:10 crc kubenswrapper[4766]: E1126 02:08:10.786582 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a65e75-b0d1-4ca6-a36c-58aab4b638b0" containerName="extract-utilities" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.786591 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a65e75-b0d1-4ca6-a36c-58aab4b638b0" containerName="extract-utilities" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.786871 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4a65e75-b0d1-4ca6-a36c-58aab4b638b0" containerName="registry-server" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.787831 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.789372 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.789696 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.789946 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.790750 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-kq4g7" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.804087 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.853317 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.853486 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-config-data\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.853524 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.956180 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.956336 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.956385 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.956438 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.956472 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-config-data\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.956505 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.956610 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vcrc\" (UniqueName: \"kubernetes.io/projected/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-kube-api-access-6vcrc\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.956694 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.956716 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.957990 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.958357 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-config-data\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:10 crc kubenswrapper[4766]: I1126 02:08:10.982037 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:11 crc kubenswrapper[4766]: I1126 02:08:11.058488 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vcrc\" (UniqueName: \"kubernetes.io/projected/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-kube-api-access-6vcrc\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:11 crc kubenswrapper[4766]: I1126 02:08:11.058577 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:11 crc kubenswrapper[4766]: I1126 02:08:11.058607 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:11 crc kubenswrapper[4766]: I1126 02:08:11.058824 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:11 crc kubenswrapper[4766]: I1126 02:08:11.058868 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:11 crc kubenswrapper[4766]: I1126 02:08:11.058911 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:11 crc kubenswrapper[4766]: I1126 02:08:11.059091 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:11 crc kubenswrapper[4766]: I1126 02:08:11.059444 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:11 crc kubenswrapper[4766]: I1126 02:08:11.061457 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/tempest-tests-tempest" Nov 26 02:08:11 crc kubenswrapper[4766]: I1126 02:08:11.063264 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:11 crc kubenswrapper[4766]: I1126 02:08:11.064265 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:11 crc kubenswrapper[4766]: I1126 02:08:11.077375 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vcrc\" (UniqueName: \"kubernetes.io/projected/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-kube-api-access-6vcrc\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:11 crc kubenswrapper[4766]: I1126 02:08:11.129573 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " pod="openstack/tempest-tests-tempest" Nov 26 02:08:11 crc kubenswrapper[4766]: I1126 02:08:11.425488 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 26 02:08:11 crc kubenswrapper[4766]: I1126 02:08:11.968026 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 26 02:08:11 crc kubenswrapper[4766]: I1126 02:08:11.970697 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 02:08:12 crc kubenswrapper[4766]: I1126 02:08:12.231192 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99","Type":"ContainerStarted","Data":"ef1d76ebb4a4de9034c8eb4c573b72645fa07647927c25ac5c62f0025758a60f"} Nov 26 02:08:17 crc kubenswrapper[4766]: I1126 02:08:17.274292 4766 trace.go:236] Trace[1572732678]: "Calculate volume metrics of catalog-content for pod openshift-marketplace/redhat-marketplace-vhw86" (26-Nov-2025 02:08:16.083) (total time: 1187ms): Nov 26 02:08:17 crc kubenswrapper[4766]: Trace[1572732678]: [1.187534496s] [1.187534496s] END Nov 26 02:08:24 crc kubenswrapper[4766]: I1126 02:08:24.254314 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v5hwl"] Nov 26 02:08:24 crc kubenswrapper[4766]: I1126 02:08:24.257173 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5hwl" Nov 26 02:08:24 crc kubenswrapper[4766]: I1126 02:08:24.279681 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5hwl"] Nov 26 02:08:24 crc kubenswrapper[4766]: I1126 02:08:24.388446 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/990efaec-0d56-4b23-b2f9-7714e0a0efe9-catalog-content\") pod \"redhat-marketplace-v5hwl\" (UID: \"990efaec-0d56-4b23-b2f9-7714e0a0efe9\") " pod="openshift-marketplace/redhat-marketplace-v5hwl" Nov 26 02:08:24 crc kubenswrapper[4766]: I1126 02:08:24.388503 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/990efaec-0d56-4b23-b2f9-7714e0a0efe9-utilities\") pod \"redhat-marketplace-v5hwl\" (UID: \"990efaec-0d56-4b23-b2f9-7714e0a0efe9\") " pod="openshift-marketplace/redhat-marketplace-v5hwl" Nov 26 02:08:24 crc kubenswrapper[4766]: I1126 02:08:24.388822 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lfkn\" (UniqueName: \"kubernetes.io/projected/990efaec-0d56-4b23-b2f9-7714e0a0efe9-kube-api-access-4lfkn\") pod \"redhat-marketplace-v5hwl\" (UID: \"990efaec-0d56-4b23-b2f9-7714e0a0efe9\") " pod="openshift-marketplace/redhat-marketplace-v5hwl" Nov 26 02:08:24 crc kubenswrapper[4766]: I1126 02:08:24.491172 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/990efaec-0d56-4b23-b2f9-7714e0a0efe9-catalog-content\") pod \"redhat-marketplace-v5hwl\" (UID: \"990efaec-0d56-4b23-b2f9-7714e0a0efe9\") " pod="openshift-marketplace/redhat-marketplace-v5hwl" Nov 26 02:08:24 crc kubenswrapper[4766]: I1126 02:08:24.491235 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/990efaec-0d56-4b23-b2f9-7714e0a0efe9-utilities\") pod \"redhat-marketplace-v5hwl\" (UID: \"990efaec-0d56-4b23-b2f9-7714e0a0efe9\") " pod="openshift-marketplace/redhat-marketplace-v5hwl" Nov 26 02:08:24 crc kubenswrapper[4766]: I1126 02:08:24.491344 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lfkn\" (UniqueName: \"kubernetes.io/projected/990efaec-0d56-4b23-b2f9-7714e0a0efe9-kube-api-access-4lfkn\") pod \"redhat-marketplace-v5hwl\" (UID: \"990efaec-0d56-4b23-b2f9-7714e0a0efe9\") " pod="openshift-marketplace/redhat-marketplace-v5hwl" Nov 26 02:08:24 crc kubenswrapper[4766]: I1126 02:08:24.491899 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/990efaec-0d56-4b23-b2f9-7714e0a0efe9-catalog-content\") pod \"redhat-marketplace-v5hwl\" (UID: \"990efaec-0d56-4b23-b2f9-7714e0a0efe9\") " pod="openshift-marketplace/redhat-marketplace-v5hwl" Nov 26 02:08:24 crc kubenswrapper[4766]: I1126 02:08:24.491943 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/990efaec-0d56-4b23-b2f9-7714e0a0efe9-utilities\") pod \"redhat-marketplace-v5hwl\" (UID: \"990efaec-0d56-4b23-b2f9-7714e0a0efe9\") " pod="openshift-marketplace/redhat-marketplace-v5hwl" Nov 26 02:08:24 crc kubenswrapper[4766]: I1126 02:08:24.520735 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lfkn\" (UniqueName: \"kubernetes.io/projected/990efaec-0d56-4b23-b2f9-7714e0a0efe9-kube-api-access-4lfkn\") pod \"redhat-marketplace-v5hwl\" (UID: \"990efaec-0d56-4b23-b2f9-7714e0a0efe9\") " pod="openshift-marketplace/redhat-marketplace-v5hwl" Nov 26 02:08:24 crc kubenswrapper[4766]: I1126 02:08:24.587871 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5hwl" Nov 26 02:08:41 crc kubenswrapper[4766]: E1126 02:08:41.343878 4766 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Nov 26 02:08:41 crc kubenswrapper[4766]: E1126 02:08:41.348043 4766 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6vcrc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(edfc5f7f-1ff2-4147-b68e-b9be01ff6a99): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 02:08:41 crc kubenswrapper[4766]: E1126 02:08:41.349226 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="edfc5f7f-1ff2-4147-b68e-b9be01ff6a99" Nov 26 02:08:41 crc kubenswrapper[4766]: E1126 02:08:41.608929 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="edfc5f7f-1ff2-4147-b68e-b9be01ff6a99" Nov 26 02:08:41 crc kubenswrapper[4766]: I1126 02:08:41.762459 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5hwl"] Nov 26 02:08:42 crc kubenswrapper[4766]: I1126 02:08:42.626469 4766 generic.go:334] "Generic (PLEG): container finished" podID="990efaec-0d56-4b23-b2f9-7714e0a0efe9" containerID="38310b158d674c810debbec513bfb87c344a8d7757297f42579798d399f31b1a" exitCode=0 Nov 26 02:08:42 crc kubenswrapper[4766]: I1126 02:08:42.626591 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5hwl" event={"ID":"990efaec-0d56-4b23-b2f9-7714e0a0efe9","Type":"ContainerDied","Data":"38310b158d674c810debbec513bfb87c344a8d7757297f42579798d399f31b1a"} Nov 26 02:08:42 crc kubenswrapper[4766]: I1126 02:08:42.626927 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5hwl" event={"ID":"990efaec-0d56-4b23-b2f9-7714e0a0efe9","Type":"ContainerStarted","Data":"72c91d3f39eb3b37950890a02a0c4804b18f77cf142ef7904fe8bad611ca0f99"} Nov 26 02:08:43 crc kubenswrapper[4766]: I1126 02:08:43.643890 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5hwl" event={"ID":"990efaec-0d56-4b23-b2f9-7714e0a0efe9","Type":"ContainerStarted","Data":"591920a559a742d473808d68ef0bf7e0563a7547cb71f6a03a931981ccbf712e"} Nov 26 02:08:44 crc kubenswrapper[4766]: I1126 02:08:44.665980 4766 generic.go:334] "Generic (PLEG): container finished" podID="990efaec-0d56-4b23-b2f9-7714e0a0efe9" containerID="591920a559a742d473808d68ef0bf7e0563a7547cb71f6a03a931981ccbf712e" exitCode=0 Nov 26 02:08:44 crc kubenswrapper[4766]: I1126 02:08:44.666095 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5hwl" event={"ID":"990efaec-0d56-4b23-b2f9-7714e0a0efe9","Type":"ContainerDied","Data":"591920a559a742d473808d68ef0bf7e0563a7547cb71f6a03a931981ccbf712e"} Nov 26 02:08:45 crc kubenswrapper[4766]: I1126 02:08:45.678037 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5hwl" event={"ID":"990efaec-0d56-4b23-b2f9-7714e0a0efe9","Type":"ContainerStarted","Data":"c1f00601facf6ee4a5fd56f400a54cb17fa3c776ff9fbbbf19409b2e5b6c929c"} Nov 26 02:08:45 crc kubenswrapper[4766]: I1126 02:08:45.710294 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v5hwl" podStartSLOduration=19.251415166 podStartE2EDuration="21.710264402s" podCreationTimestamp="2025-11-26 02:08:24 +0000 UTC" firstStartedPulling="2025-11-26 02:08:42.629982472 +0000 UTC m=+6303.478752942" lastFinishedPulling="2025-11-26 02:08:45.088831748 +0000 UTC m=+6305.937602178" observedRunningTime="2025-11-26 02:08:45.701828557 +0000 UTC m=+6306.550598987" watchObservedRunningTime="2025-11-26 02:08:45.710264402 +0000 UTC m=+6306.559034872" Nov 26 02:08:53 crc kubenswrapper[4766]: I1126 02:08:53.292321 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 26 02:08:54 crc kubenswrapper[4766]: I1126 02:08:54.588710 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v5hwl" Nov 26 02:08:54 crc kubenswrapper[4766]: I1126 02:08:54.589351 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v5hwl" Nov 26 02:08:54 crc kubenswrapper[4766]: I1126 02:08:54.637020 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v5hwl" Nov 26 02:08:54 crc kubenswrapper[4766]: I1126 02:08:54.802682 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99","Type":"ContainerStarted","Data":"d6cb0b3456d7352838582860f7ebfad091a9a9316ec28d54973e7e546b75c0e6"} Nov 26 02:08:54 crc kubenswrapper[4766]: I1126 02:08:54.851216 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.533219965 podStartE2EDuration="45.851192237s" podCreationTimestamp="2025-11-26 02:08:09 +0000 UTC" firstStartedPulling="2025-11-26 02:08:11.970311868 +0000 UTC m=+6272.819082338" lastFinishedPulling="2025-11-26 02:08:53.28828414 +0000 UTC m=+6314.137054610" observedRunningTime="2025-11-26 02:08:54.835003994 +0000 UTC m=+6315.683774464" watchObservedRunningTime="2025-11-26 02:08:54.851192237 +0000 UTC m=+6315.699962677" Nov 26 02:08:54 crc kubenswrapper[4766]: I1126 02:08:54.882080 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v5hwl" Nov 26 02:08:55 crc kubenswrapper[4766]: I1126 02:08:55.450213 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5hwl"] Nov 26 02:08:56 crc kubenswrapper[4766]: I1126 02:08:56.837117 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v5hwl" podUID="990efaec-0d56-4b23-b2f9-7714e0a0efe9" containerName="registry-server" containerID="cri-o://c1f00601facf6ee4a5fd56f400a54cb17fa3c776ff9fbbbf19409b2e5b6c929c" gracePeriod=2 Nov 26 02:08:57 crc kubenswrapper[4766]: I1126 02:08:57.380389 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5hwl" Nov 26 02:08:57 crc kubenswrapper[4766]: I1126 02:08:57.489753 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/990efaec-0d56-4b23-b2f9-7714e0a0efe9-catalog-content\") pod \"990efaec-0d56-4b23-b2f9-7714e0a0efe9\" (UID: \"990efaec-0d56-4b23-b2f9-7714e0a0efe9\") " Nov 26 02:08:57 crc kubenswrapper[4766]: I1126 02:08:57.489873 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/990efaec-0d56-4b23-b2f9-7714e0a0efe9-utilities\") pod \"990efaec-0d56-4b23-b2f9-7714e0a0efe9\" (UID: \"990efaec-0d56-4b23-b2f9-7714e0a0efe9\") " Nov 26 02:08:57 crc kubenswrapper[4766]: I1126 02:08:57.489976 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lfkn\" (UniqueName: \"kubernetes.io/projected/990efaec-0d56-4b23-b2f9-7714e0a0efe9-kube-api-access-4lfkn\") pod \"990efaec-0d56-4b23-b2f9-7714e0a0efe9\" (UID: \"990efaec-0d56-4b23-b2f9-7714e0a0efe9\") " Nov 26 02:08:57 crc kubenswrapper[4766]: I1126 02:08:57.491294 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/990efaec-0d56-4b23-b2f9-7714e0a0efe9-utilities" (OuterVolumeSpecName: "utilities") pod "990efaec-0d56-4b23-b2f9-7714e0a0efe9" (UID: "990efaec-0d56-4b23-b2f9-7714e0a0efe9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:08:57 crc kubenswrapper[4766]: I1126 02:08:57.497437 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/990efaec-0d56-4b23-b2f9-7714e0a0efe9-kube-api-access-4lfkn" (OuterVolumeSpecName: "kube-api-access-4lfkn") pod "990efaec-0d56-4b23-b2f9-7714e0a0efe9" (UID: "990efaec-0d56-4b23-b2f9-7714e0a0efe9"). InnerVolumeSpecName "kube-api-access-4lfkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:08:57 crc kubenswrapper[4766]: I1126 02:08:57.515265 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/990efaec-0d56-4b23-b2f9-7714e0a0efe9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "990efaec-0d56-4b23-b2f9-7714e0a0efe9" (UID: "990efaec-0d56-4b23-b2f9-7714e0a0efe9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:08:57 crc kubenswrapper[4766]: I1126 02:08:57.593150 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lfkn\" (UniqueName: \"kubernetes.io/projected/990efaec-0d56-4b23-b2f9-7714e0a0efe9-kube-api-access-4lfkn\") on node \"crc\" DevicePath \"\"" Nov 26 02:08:57 crc kubenswrapper[4766]: I1126 02:08:57.593210 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/990efaec-0d56-4b23-b2f9-7714e0a0efe9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 02:08:57 crc kubenswrapper[4766]: I1126 02:08:57.593230 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/990efaec-0d56-4b23-b2f9-7714e0a0efe9-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 02:08:57 crc kubenswrapper[4766]: I1126 02:08:57.854438 4766 generic.go:334] "Generic (PLEG): container finished" podID="990efaec-0d56-4b23-b2f9-7714e0a0efe9" containerID="c1f00601facf6ee4a5fd56f400a54cb17fa3c776ff9fbbbf19409b2e5b6c929c" exitCode=0 Nov 26 02:08:57 crc kubenswrapper[4766]: I1126 02:08:57.854507 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5hwl" Nov 26 02:08:57 crc kubenswrapper[4766]: I1126 02:08:57.854530 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5hwl" event={"ID":"990efaec-0d56-4b23-b2f9-7714e0a0efe9","Type":"ContainerDied","Data":"c1f00601facf6ee4a5fd56f400a54cb17fa3c776ff9fbbbf19409b2e5b6c929c"} Nov 26 02:08:57 crc kubenswrapper[4766]: I1126 02:08:57.856893 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5hwl" event={"ID":"990efaec-0d56-4b23-b2f9-7714e0a0efe9","Type":"ContainerDied","Data":"72c91d3f39eb3b37950890a02a0c4804b18f77cf142ef7904fe8bad611ca0f99"} Nov 26 02:08:57 crc kubenswrapper[4766]: I1126 02:08:57.856928 4766 scope.go:117] "RemoveContainer" containerID="c1f00601facf6ee4a5fd56f400a54cb17fa3c776ff9fbbbf19409b2e5b6c929c" Nov 26 02:08:57 crc kubenswrapper[4766]: I1126 02:08:57.893515 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5hwl"] Nov 26 02:08:57 crc kubenswrapper[4766]: I1126 02:08:57.903408 4766 scope.go:117] "RemoveContainer" containerID="591920a559a742d473808d68ef0bf7e0563a7547cb71f6a03a931981ccbf712e" Nov 26 02:08:57 crc kubenswrapper[4766]: I1126 02:08:57.915519 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5hwl"] Nov 26 02:08:57 crc kubenswrapper[4766]: I1126 02:08:57.944506 4766 scope.go:117] "RemoveContainer" containerID="38310b158d674c810debbec513bfb87c344a8d7757297f42579798d399f31b1a" Nov 26 02:08:58 crc kubenswrapper[4766]: I1126 02:08:58.001485 4766 scope.go:117] "RemoveContainer" containerID="c1f00601facf6ee4a5fd56f400a54cb17fa3c776ff9fbbbf19409b2e5b6c929c" Nov 26 02:08:58 crc kubenswrapper[4766]: E1126 02:08:58.012225 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1f00601facf6ee4a5fd56f400a54cb17fa3c776ff9fbbbf19409b2e5b6c929c\": container with ID starting with c1f00601facf6ee4a5fd56f400a54cb17fa3c776ff9fbbbf19409b2e5b6c929c not found: ID does not exist" containerID="c1f00601facf6ee4a5fd56f400a54cb17fa3c776ff9fbbbf19409b2e5b6c929c" Nov 26 02:08:58 crc kubenswrapper[4766]: I1126 02:08:58.012276 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1f00601facf6ee4a5fd56f400a54cb17fa3c776ff9fbbbf19409b2e5b6c929c"} err="failed to get container status \"c1f00601facf6ee4a5fd56f400a54cb17fa3c776ff9fbbbf19409b2e5b6c929c\": rpc error: code = NotFound desc = could not find container \"c1f00601facf6ee4a5fd56f400a54cb17fa3c776ff9fbbbf19409b2e5b6c929c\": container with ID starting with c1f00601facf6ee4a5fd56f400a54cb17fa3c776ff9fbbbf19409b2e5b6c929c not found: ID does not exist" Nov 26 02:08:58 crc kubenswrapper[4766]: I1126 02:08:58.012302 4766 scope.go:117] "RemoveContainer" containerID="591920a559a742d473808d68ef0bf7e0563a7547cb71f6a03a931981ccbf712e" Nov 26 02:08:58 crc kubenswrapper[4766]: E1126 02:08:58.012797 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"591920a559a742d473808d68ef0bf7e0563a7547cb71f6a03a931981ccbf712e\": container with ID starting with 591920a559a742d473808d68ef0bf7e0563a7547cb71f6a03a931981ccbf712e not found: ID does not exist" containerID="591920a559a742d473808d68ef0bf7e0563a7547cb71f6a03a931981ccbf712e" Nov 26 02:08:58 crc kubenswrapper[4766]: I1126 02:08:58.012838 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"591920a559a742d473808d68ef0bf7e0563a7547cb71f6a03a931981ccbf712e"} err="failed to get container status \"591920a559a742d473808d68ef0bf7e0563a7547cb71f6a03a931981ccbf712e\": rpc error: code = NotFound desc = could not find container \"591920a559a742d473808d68ef0bf7e0563a7547cb71f6a03a931981ccbf712e\": container with ID starting with 591920a559a742d473808d68ef0bf7e0563a7547cb71f6a03a931981ccbf712e not found: ID does not exist" Nov 26 02:08:58 crc kubenswrapper[4766]: I1126 02:08:58.012863 4766 scope.go:117] "RemoveContainer" containerID="38310b158d674c810debbec513bfb87c344a8d7757297f42579798d399f31b1a" Nov 26 02:08:58 crc kubenswrapper[4766]: E1126 02:08:58.013341 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38310b158d674c810debbec513bfb87c344a8d7757297f42579798d399f31b1a\": container with ID starting with 38310b158d674c810debbec513bfb87c344a8d7757297f42579798d399f31b1a not found: ID does not exist" containerID="38310b158d674c810debbec513bfb87c344a8d7757297f42579798d399f31b1a" Nov 26 02:08:58 crc kubenswrapper[4766]: I1126 02:08:58.013372 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38310b158d674c810debbec513bfb87c344a8d7757297f42579798d399f31b1a"} err="failed to get container status \"38310b158d674c810debbec513bfb87c344a8d7757297f42579798d399f31b1a\": rpc error: code = NotFound desc = could not find container \"38310b158d674c810debbec513bfb87c344a8d7757297f42579798d399f31b1a\": container with ID starting with 38310b158d674c810debbec513bfb87c344a8d7757297f42579798d399f31b1a not found: ID does not exist" Nov 26 02:08:59 crc kubenswrapper[4766]: I1126 02:08:59.843213 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="990efaec-0d56-4b23-b2f9-7714e0a0efe9" path="/var/lib/kubelet/pods/990efaec-0d56-4b23-b2f9-7714e0a0efe9/volumes" Nov 26 02:09:11 crc kubenswrapper[4766]: I1126 02:09:11.479987 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:09:11 crc kubenswrapper[4766]: I1126 02:09:11.480768 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:09:41 crc kubenswrapper[4766]: I1126 02:09:41.479825 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:09:41 crc kubenswrapper[4766]: I1126 02:09:41.480471 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:10:11 crc kubenswrapper[4766]: I1126 02:10:11.479472 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:10:11 crc kubenswrapper[4766]: I1126 02:10:11.480151 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:10:11 crc kubenswrapper[4766]: I1126 02:10:11.480221 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 02:10:11 crc kubenswrapper[4766]: I1126 02:10:11.481119 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 02:10:11 crc kubenswrapper[4766]: I1126 02:10:11.481197 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" gracePeriod=600 Nov 26 02:10:11 crc kubenswrapper[4766]: E1126 02:10:11.606869 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:10:11 crc kubenswrapper[4766]: I1126 02:10:11.745776 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" exitCode=0 Nov 26 02:10:11 crc kubenswrapper[4766]: I1126 02:10:11.745841 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b"} Nov 26 02:10:11 crc kubenswrapper[4766]: I1126 02:10:11.746175 4766 scope.go:117] "RemoveContainer" containerID="e70127ee22b463f69db81a0069e5277e44a8b961e37fbeac182f3a366ae58c79" Nov 26 02:10:11 crc kubenswrapper[4766]: I1126 02:10:11.747238 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:10:11 crc kubenswrapper[4766]: E1126 02:10:11.747623 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:10:23 crc kubenswrapper[4766]: I1126 02:10:23.828488 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:10:23 crc kubenswrapper[4766]: E1126 02:10:23.829311 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:10:38 crc kubenswrapper[4766]: I1126 02:10:38.827078 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:10:38 crc kubenswrapper[4766]: E1126 02:10:38.827959 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:10:51 crc kubenswrapper[4766]: I1126 02:10:51.827278 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:10:51 crc kubenswrapper[4766]: E1126 02:10:51.828160 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:11:00 crc kubenswrapper[4766]: I1126 02:11:00.680848 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d4pbk"] Nov 26 02:11:00 crc kubenswrapper[4766]: E1126 02:11:00.686950 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="990efaec-0d56-4b23-b2f9-7714e0a0efe9" containerName="extract-content" Nov 26 02:11:00 crc kubenswrapper[4766]: I1126 02:11:00.687116 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="990efaec-0d56-4b23-b2f9-7714e0a0efe9" containerName="extract-content" Nov 26 02:11:00 crc kubenswrapper[4766]: E1126 02:11:00.687191 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="990efaec-0d56-4b23-b2f9-7714e0a0efe9" containerName="extract-utilities" Nov 26 02:11:00 crc kubenswrapper[4766]: I1126 02:11:00.687246 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="990efaec-0d56-4b23-b2f9-7714e0a0efe9" containerName="extract-utilities" Nov 26 02:11:00 crc kubenswrapper[4766]: E1126 02:11:00.687326 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="990efaec-0d56-4b23-b2f9-7714e0a0efe9" containerName="registry-server" Nov 26 02:11:00 crc kubenswrapper[4766]: I1126 02:11:00.687383 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="990efaec-0d56-4b23-b2f9-7714e0a0efe9" containerName="registry-server" Nov 26 02:11:00 crc kubenswrapper[4766]: I1126 02:11:00.687672 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="990efaec-0d56-4b23-b2f9-7714e0a0efe9" containerName="registry-server" Nov 26 02:11:00 crc kubenswrapper[4766]: I1126 02:11:00.689370 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d4pbk" Nov 26 02:11:00 crc kubenswrapper[4766]: I1126 02:11:00.704082 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d4pbk"] Nov 26 02:11:00 crc kubenswrapper[4766]: I1126 02:11:00.768152 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9d5ffd2-0ec1-4162-b59a-bac4c8265789-catalog-content\") pod \"redhat-operators-d4pbk\" (UID: \"d9d5ffd2-0ec1-4162-b59a-bac4c8265789\") " pod="openshift-marketplace/redhat-operators-d4pbk" Nov 26 02:11:00 crc kubenswrapper[4766]: I1126 02:11:00.768238 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j9nk\" (UniqueName: \"kubernetes.io/projected/d9d5ffd2-0ec1-4162-b59a-bac4c8265789-kube-api-access-4j9nk\") pod \"redhat-operators-d4pbk\" (UID: \"d9d5ffd2-0ec1-4162-b59a-bac4c8265789\") " pod="openshift-marketplace/redhat-operators-d4pbk" Nov 26 02:11:00 crc kubenswrapper[4766]: I1126 02:11:00.768281 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9d5ffd2-0ec1-4162-b59a-bac4c8265789-utilities\") pod \"redhat-operators-d4pbk\" (UID: \"d9d5ffd2-0ec1-4162-b59a-bac4c8265789\") " pod="openshift-marketplace/redhat-operators-d4pbk" Nov 26 02:11:00 crc kubenswrapper[4766]: I1126 02:11:00.870139 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9d5ffd2-0ec1-4162-b59a-bac4c8265789-utilities\") pod \"redhat-operators-d4pbk\" (UID: \"d9d5ffd2-0ec1-4162-b59a-bac4c8265789\") " pod="openshift-marketplace/redhat-operators-d4pbk" Nov 26 02:11:00 crc kubenswrapper[4766]: I1126 02:11:00.870483 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9d5ffd2-0ec1-4162-b59a-bac4c8265789-catalog-content\") pod \"redhat-operators-d4pbk\" (UID: \"d9d5ffd2-0ec1-4162-b59a-bac4c8265789\") " pod="openshift-marketplace/redhat-operators-d4pbk" Nov 26 02:11:00 crc kubenswrapper[4766]: I1126 02:11:00.870601 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j9nk\" (UniqueName: \"kubernetes.io/projected/d9d5ffd2-0ec1-4162-b59a-bac4c8265789-kube-api-access-4j9nk\") pod \"redhat-operators-d4pbk\" (UID: \"d9d5ffd2-0ec1-4162-b59a-bac4c8265789\") " pod="openshift-marketplace/redhat-operators-d4pbk" Nov 26 02:11:00 crc kubenswrapper[4766]: I1126 02:11:00.872721 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9d5ffd2-0ec1-4162-b59a-bac4c8265789-utilities\") pod \"redhat-operators-d4pbk\" (UID: \"d9d5ffd2-0ec1-4162-b59a-bac4c8265789\") " pod="openshift-marketplace/redhat-operators-d4pbk" Nov 26 02:11:00 crc kubenswrapper[4766]: I1126 02:11:00.872759 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9d5ffd2-0ec1-4162-b59a-bac4c8265789-catalog-content\") pod \"redhat-operators-d4pbk\" (UID: \"d9d5ffd2-0ec1-4162-b59a-bac4c8265789\") " pod="openshift-marketplace/redhat-operators-d4pbk" Nov 26 02:11:00 crc kubenswrapper[4766]: I1126 02:11:00.895708 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j9nk\" (UniqueName: \"kubernetes.io/projected/d9d5ffd2-0ec1-4162-b59a-bac4c8265789-kube-api-access-4j9nk\") pod \"redhat-operators-d4pbk\" (UID: \"d9d5ffd2-0ec1-4162-b59a-bac4c8265789\") " pod="openshift-marketplace/redhat-operators-d4pbk" Nov 26 02:11:01 crc kubenswrapper[4766]: I1126 02:11:01.024105 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d4pbk" Nov 26 02:11:01 crc kubenswrapper[4766]: I1126 02:11:01.663273 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d4pbk"] Nov 26 02:11:02 crc kubenswrapper[4766]: I1126 02:11:02.367895 4766 generic.go:334] "Generic (PLEG): container finished" podID="d9d5ffd2-0ec1-4162-b59a-bac4c8265789" containerID="4cfa46dcf4bbc7d4b8e2112debe6e10abd35b2c1b7443a63568a2001300a62f3" exitCode=0 Nov 26 02:11:02 crc kubenswrapper[4766]: I1126 02:11:02.368180 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4pbk" event={"ID":"d9d5ffd2-0ec1-4162-b59a-bac4c8265789","Type":"ContainerDied","Data":"4cfa46dcf4bbc7d4b8e2112debe6e10abd35b2c1b7443a63568a2001300a62f3"} Nov 26 02:11:02 crc kubenswrapper[4766]: I1126 02:11:02.368208 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4pbk" event={"ID":"d9d5ffd2-0ec1-4162-b59a-bac4c8265789","Type":"ContainerStarted","Data":"fd7804f19e94bae00f6f9d30543b7e05725e564a22dfdcbe4ee9087fe9a4f349"} Nov 26 02:11:03 crc kubenswrapper[4766]: I1126 02:11:03.379025 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4pbk" event={"ID":"d9d5ffd2-0ec1-4162-b59a-bac4c8265789","Type":"ContainerStarted","Data":"2d0c98d1898485f1c70f5dc66ddd984ce89cc76fa32bcd9481b465b887f8b773"} Nov 26 02:11:04 crc kubenswrapper[4766]: I1126 02:11:04.827926 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:11:04 crc kubenswrapper[4766]: E1126 02:11:04.828542 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:11:07 crc kubenswrapper[4766]: I1126 02:11:07.422992 4766 generic.go:334] "Generic (PLEG): container finished" podID="d9d5ffd2-0ec1-4162-b59a-bac4c8265789" containerID="2d0c98d1898485f1c70f5dc66ddd984ce89cc76fa32bcd9481b465b887f8b773" exitCode=0 Nov 26 02:11:07 crc kubenswrapper[4766]: I1126 02:11:07.423148 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4pbk" event={"ID":"d9d5ffd2-0ec1-4162-b59a-bac4c8265789","Type":"ContainerDied","Data":"2d0c98d1898485f1c70f5dc66ddd984ce89cc76fa32bcd9481b465b887f8b773"} Nov 26 02:11:08 crc kubenswrapper[4766]: I1126 02:11:08.436759 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4pbk" event={"ID":"d9d5ffd2-0ec1-4162-b59a-bac4c8265789","Type":"ContainerStarted","Data":"3f9f7a180297055cbb897d5c83c45a79892a14e9f6b958a2ac1e4cec6bbf9e7f"} Nov 26 02:11:11 crc kubenswrapper[4766]: I1126 02:11:11.024596 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d4pbk" Nov 26 02:11:11 crc kubenswrapper[4766]: I1126 02:11:11.025423 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d4pbk" Nov 26 02:11:12 crc kubenswrapper[4766]: I1126 02:11:12.081041 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d4pbk" podUID="d9d5ffd2-0ec1-4162-b59a-bac4c8265789" containerName="registry-server" probeResult="failure" output=< Nov 26 02:11:12 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 02:11:12 crc kubenswrapper[4766]: > Nov 26 02:11:16 crc kubenswrapper[4766]: I1126 02:11:16.827027 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:11:16 crc kubenswrapper[4766]: E1126 02:11:16.828011 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:11:22 crc kubenswrapper[4766]: I1126 02:11:22.081710 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d4pbk" podUID="d9d5ffd2-0ec1-4162-b59a-bac4c8265789" containerName="registry-server" probeResult="failure" output=< Nov 26 02:11:22 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 02:11:22 crc kubenswrapper[4766]: > Nov 26 02:11:28 crc kubenswrapper[4766]: I1126 02:11:28.826949 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:11:28 crc kubenswrapper[4766]: E1126 02:11:28.827773 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:11:32 crc kubenswrapper[4766]: I1126 02:11:32.085812 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d4pbk" podUID="d9d5ffd2-0ec1-4162-b59a-bac4c8265789" containerName="registry-server" probeResult="failure" output=< Nov 26 02:11:32 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 02:11:32 crc kubenswrapper[4766]: > Nov 26 02:11:41 crc kubenswrapper[4766]: I1126 02:11:41.091671 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d4pbk" Nov 26 02:11:41 crc kubenswrapper[4766]: I1126 02:11:41.122910 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d4pbk" podStartSLOduration=35.414269904 podStartE2EDuration="41.122887843s" podCreationTimestamp="2025-11-26 02:11:00 +0000 UTC" firstStartedPulling="2025-11-26 02:11:02.370587949 +0000 UTC m=+6443.219358389" lastFinishedPulling="2025-11-26 02:11:08.079205828 +0000 UTC m=+6448.927976328" observedRunningTime="2025-11-26 02:11:08.468951179 +0000 UTC m=+6449.317721619" watchObservedRunningTime="2025-11-26 02:11:41.122887843 +0000 UTC m=+6481.971658283" Nov 26 02:11:41 crc kubenswrapper[4766]: I1126 02:11:41.149634 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d4pbk" Nov 26 02:11:41 crc kubenswrapper[4766]: I1126 02:11:41.331722 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d4pbk"] Nov 26 02:11:42 crc kubenswrapper[4766]: I1126 02:11:42.824391 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d4pbk" podUID="d9d5ffd2-0ec1-4162-b59a-bac4c8265789" containerName="registry-server" containerID="cri-o://3f9f7a180297055cbb897d5c83c45a79892a14e9f6b958a2ac1e4cec6bbf9e7f" gracePeriod=2 Nov 26 02:11:43 crc kubenswrapper[4766]: I1126 02:11:43.829581 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:11:43 crc kubenswrapper[4766]: E1126 02:11:43.830160 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:11:43 crc kubenswrapper[4766]: I1126 02:11:43.840145 4766 generic.go:334] "Generic (PLEG): container finished" podID="d9d5ffd2-0ec1-4162-b59a-bac4c8265789" containerID="3f9f7a180297055cbb897d5c83c45a79892a14e9f6b958a2ac1e4cec6bbf9e7f" exitCode=0 Nov 26 02:11:43 crc kubenswrapper[4766]: I1126 02:11:43.840986 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4pbk" event={"ID":"d9d5ffd2-0ec1-4162-b59a-bac4c8265789","Type":"ContainerDied","Data":"3f9f7a180297055cbb897d5c83c45a79892a14e9f6b958a2ac1e4cec6bbf9e7f"} Nov 26 02:11:44 crc kubenswrapper[4766]: I1126 02:11:44.107369 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d4pbk" Nov 26 02:11:44 crc kubenswrapper[4766]: I1126 02:11:44.210095 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9d5ffd2-0ec1-4162-b59a-bac4c8265789-catalog-content\") pod \"d9d5ffd2-0ec1-4162-b59a-bac4c8265789\" (UID: \"d9d5ffd2-0ec1-4162-b59a-bac4c8265789\") " Nov 26 02:11:44 crc kubenswrapper[4766]: I1126 02:11:44.210328 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9d5ffd2-0ec1-4162-b59a-bac4c8265789-utilities\") pod \"d9d5ffd2-0ec1-4162-b59a-bac4c8265789\" (UID: \"d9d5ffd2-0ec1-4162-b59a-bac4c8265789\") " Nov 26 02:11:44 crc kubenswrapper[4766]: I1126 02:11:44.210469 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4j9nk\" (UniqueName: \"kubernetes.io/projected/d9d5ffd2-0ec1-4162-b59a-bac4c8265789-kube-api-access-4j9nk\") pod \"d9d5ffd2-0ec1-4162-b59a-bac4c8265789\" (UID: \"d9d5ffd2-0ec1-4162-b59a-bac4c8265789\") " Nov 26 02:11:44 crc kubenswrapper[4766]: I1126 02:11:44.212130 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9d5ffd2-0ec1-4162-b59a-bac4c8265789-utilities" (OuterVolumeSpecName: "utilities") pod "d9d5ffd2-0ec1-4162-b59a-bac4c8265789" (UID: "d9d5ffd2-0ec1-4162-b59a-bac4c8265789"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:11:44 crc kubenswrapper[4766]: I1126 02:11:44.225908 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9d5ffd2-0ec1-4162-b59a-bac4c8265789-kube-api-access-4j9nk" (OuterVolumeSpecName: "kube-api-access-4j9nk") pod "d9d5ffd2-0ec1-4162-b59a-bac4c8265789" (UID: "d9d5ffd2-0ec1-4162-b59a-bac4c8265789"). InnerVolumeSpecName "kube-api-access-4j9nk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:11:44 crc kubenswrapper[4766]: I1126 02:11:44.286528 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9d5ffd2-0ec1-4162-b59a-bac4c8265789-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d9d5ffd2-0ec1-4162-b59a-bac4c8265789" (UID: "d9d5ffd2-0ec1-4162-b59a-bac4c8265789"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:11:44 crc kubenswrapper[4766]: I1126 02:11:44.315585 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9d5ffd2-0ec1-4162-b59a-bac4c8265789-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 02:11:44 crc kubenswrapper[4766]: I1126 02:11:44.315815 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4j9nk\" (UniqueName: \"kubernetes.io/projected/d9d5ffd2-0ec1-4162-b59a-bac4c8265789-kube-api-access-4j9nk\") on node \"crc\" DevicePath \"\"" Nov 26 02:11:44 crc kubenswrapper[4766]: I1126 02:11:44.316238 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9d5ffd2-0ec1-4162-b59a-bac4c8265789-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 02:11:44 crc kubenswrapper[4766]: I1126 02:11:44.862014 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4pbk" event={"ID":"d9d5ffd2-0ec1-4162-b59a-bac4c8265789","Type":"ContainerDied","Data":"fd7804f19e94bae00f6f9d30543b7e05725e564a22dfdcbe4ee9087fe9a4f349"} Nov 26 02:11:44 crc kubenswrapper[4766]: I1126 02:11:44.862522 4766 scope.go:117] "RemoveContainer" containerID="3f9f7a180297055cbb897d5c83c45a79892a14e9f6b958a2ac1e4cec6bbf9e7f" Nov 26 02:11:44 crc kubenswrapper[4766]: I1126 02:11:44.862075 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d4pbk" Nov 26 02:11:44 crc kubenswrapper[4766]: I1126 02:11:44.910923 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d4pbk"] Nov 26 02:11:44 crc kubenswrapper[4766]: I1126 02:11:44.911339 4766 scope.go:117] "RemoveContainer" containerID="2d0c98d1898485f1c70f5dc66ddd984ce89cc76fa32bcd9481b465b887f8b773" Nov 26 02:11:44 crc kubenswrapper[4766]: I1126 02:11:44.935062 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d4pbk"] Nov 26 02:11:44 crc kubenswrapper[4766]: I1126 02:11:44.940802 4766 scope.go:117] "RemoveContainer" containerID="4cfa46dcf4bbc7d4b8e2112debe6e10abd35b2c1b7443a63568a2001300a62f3" Nov 26 02:11:45 crc kubenswrapper[4766]: I1126 02:11:45.841211 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9d5ffd2-0ec1-4162-b59a-bac4c8265789" path="/var/lib/kubelet/pods/d9d5ffd2-0ec1-4162-b59a-bac4c8265789/volumes" Nov 26 02:11:57 crc kubenswrapper[4766]: I1126 02:11:57.827519 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:11:57 crc kubenswrapper[4766]: E1126 02:11:57.828231 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:12:11 crc kubenswrapper[4766]: I1126 02:12:11.830019 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:12:11 crc kubenswrapper[4766]: E1126 02:12:11.831986 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:12:23 crc kubenswrapper[4766]: I1126 02:12:23.827457 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:12:23 crc kubenswrapper[4766]: E1126 02:12:23.828417 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:12:37 crc kubenswrapper[4766]: I1126 02:12:37.828264 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:12:37 crc kubenswrapper[4766]: E1126 02:12:37.829198 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:12:49 crc kubenswrapper[4766]: I1126 02:12:49.845968 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:12:49 crc kubenswrapper[4766]: E1126 02:12:49.847276 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:13:00 crc kubenswrapper[4766]: I1126 02:13:00.828478 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:13:00 crc kubenswrapper[4766]: E1126 02:13:00.829542 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:13:14 crc kubenswrapper[4766]: I1126 02:13:14.826776 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:13:14 crc kubenswrapper[4766]: E1126 02:13:14.827591 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:13:26 crc kubenswrapper[4766]: I1126 02:13:26.826809 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:13:26 crc kubenswrapper[4766]: E1126 02:13:26.828063 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:13:35 crc kubenswrapper[4766]: I1126 02:13:35.370842 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kwmp6"] Nov 26 02:13:35 crc kubenswrapper[4766]: E1126 02:13:35.371875 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9d5ffd2-0ec1-4162-b59a-bac4c8265789" containerName="extract-utilities" Nov 26 02:13:35 crc kubenswrapper[4766]: I1126 02:13:35.371890 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d5ffd2-0ec1-4162-b59a-bac4c8265789" containerName="extract-utilities" Nov 26 02:13:35 crc kubenswrapper[4766]: E1126 02:13:35.371906 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9d5ffd2-0ec1-4162-b59a-bac4c8265789" containerName="registry-server" Nov 26 02:13:35 crc kubenswrapper[4766]: I1126 02:13:35.371912 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d5ffd2-0ec1-4162-b59a-bac4c8265789" containerName="registry-server" Nov 26 02:13:35 crc kubenswrapper[4766]: E1126 02:13:35.371938 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9d5ffd2-0ec1-4162-b59a-bac4c8265789" containerName="extract-content" Nov 26 02:13:35 crc kubenswrapper[4766]: I1126 02:13:35.371945 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d5ffd2-0ec1-4162-b59a-bac4c8265789" containerName="extract-content" Nov 26 02:13:35 crc kubenswrapper[4766]: I1126 02:13:35.372164 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9d5ffd2-0ec1-4162-b59a-bac4c8265789" containerName="registry-server" Nov 26 02:13:35 crc kubenswrapper[4766]: I1126 02:13:35.373745 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kwmp6" Nov 26 02:13:35 crc kubenswrapper[4766]: I1126 02:13:35.389204 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kwmp6"] Nov 26 02:13:35 crc kubenswrapper[4766]: I1126 02:13:35.519914 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ee59fe9-db71-4771-9ca6-58e16ffbdb5d-utilities\") pod \"certified-operators-kwmp6\" (UID: \"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d\") " pod="openshift-marketplace/certified-operators-kwmp6" Nov 26 02:13:35 crc kubenswrapper[4766]: I1126 02:13:35.519998 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ee59fe9-db71-4771-9ca6-58e16ffbdb5d-catalog-content\") pod \"certified-operators-kwmp6\" (UID: \"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d\") " pod="openshift-marketplace/certified-operators-kwmp6" Nov 26 02:13:35 crc kubenswrapper[4766]: I1126 02:13:35.520102 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6n56\" (UniqueName: \"kubernetes.io/projected/2ee59fe9-db71-4771-9ca6-58e16ffbdb5d-kube-api-access-h6n56\") pod \"certified-operators-kwmp6\" (UID: \"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d\") " pod="openshift-marketplace/certified-operators-kwmp6" Nov 26 02:13:35 crc kubenswrapper[4766]: I1126 02:13:35.621480 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ee59fe9-db71-4771-9ca6-58e16ffbdb5d-utilities\") pod \"certified-operators-kwmp6\" (UID: \"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d\") " pod="openshift-marketplace/certified-operators-kwmp6" Nov 26 02:13:35 crc kubenswrapper[4766]: I1126 02:13:35.621580 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ee59fe9-db71-4771-9ca6-58e16ffbdb5d-catalog-content\") pod \"certified-operators-kwmp6\" (UID: \"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d\") " pod="openshift-marketplace/certified-operators-kwmp6" Nov 26 02:13:35 crc kubenswrapper[4766]: I1126 02:13:35.621695 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6n56\" (UniqueName: \"kubernetes.io/projected/2ee59fe9-db71-4771-9ca6-58e16ffbdb5d-kube-api-access-h6n56\") pod \"certified-operators-kwmp6\" (UID: \"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d\") " pod="openshift-marketplace/certified-operators-kwmp6" Nov 26 02:13:35 crc kubenswrapper[4766]: I1126 02:13:35.622195 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ee59fe9-db71-4771-9ca6-58e16ffbdb5d-utilities\") pod \"certified-operators-kwmp6\" (UID: \"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d\") " pod="openshift-marketplace/certified-operators-kwmp6" Nov 26 02:13:35 crc kubenswrapper[4766]: I1126 02:13:35.622212 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ee59fe9-db71-4771-9ca6-58e16ffbdb5d-catalog-content\") pod \"certified-operators-kwmp6\" (UID: \"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d\") " pod="openshift-marketplace/certified-operators-kwmp6" Nov 26 02:13:35 crc kubenswrapper[4766]: I1126 02:13:35.645578 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6n56\" (UniqueName: \"kubernetes.io/projected/2ee59fe9-db71-4771-9ca6-58e16ffbdb5d-kube-api-access-h6n56\") pod \"certified-operators-kwmp6\" (UID: \"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d\") " pod="openshift-marketplace/certified-operators-kwmp6" Nov 26 02:13:35 crc kubenswrapper[4766]: I1126 02:13:35.702766 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kwmp6" Nov 26 02:13:36 crc kubenswrapper[4766]: I1126 02:13:36.290047 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kwmp6"] Nov 26 02:13:37 crc kubenswrapper[4766]: I1126 02:13:37.259803 4766 generic.go:334] "Generic (PLEG): container finished" podID="2ee59fe9-db71-4771-9ca6-58e16ffbdb5d" containerID="70dc10cd62e2b646de789e104f03f1c92f71c10117f1022ae6f98d73fa4d0b2b" exitCode=0 Nov 26 02:13:37 crc kubenswrapper[4766]: I1126 02:13:37.260231 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kwmp6" event={"ID":"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d","Type":"ContainerDied","Data":"70dc10cd62e2b646de789e104f03f1c92f71c10117f1022ae6f98d73fa4d0b2b"} Nov 26 02:13:37 crc kubenswrapper[4766]: I1126 02:13:37.260256 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kwmp6" event={"ID":"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d","Type":"ContainerStarted","Data":"6a6056d9d17567d0fe6629ff9fbb0ec00d733641f7df2a000b935ad54b876ac8"} Nov 26 02:13:37 crc kubenswrapper[4766]: I1126 02:13:37.263488 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 02:13:38 crc kubenswrapper[4766]: I1126 02:13:38.271927 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kwmp6" event={"ID":"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d","Type":"ContainerStarted","Data":"b0cef7707416a0d05eef4e3507d659437398def515b10002e0bbe1230e8d0c74"} Nov 26 02:13:40 crc kubenswrapper[4766]: I1126 02:13:40.309971 4766 generic.go:334] "Generic (PLEG): container finished" podID="2ee59fe9-db71-4771-9ca6-58e16ffbdb5d" containerID="b0cef7707416a0d05eef4e3507d659437398def515b10002e0bbe1230e8d0c74" exitCode=0 Nov 26 02:13:40 crc kubenswrapper[4766]: I1126 02:13:40.310084 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kwmp6" event={"ID":"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d","Type":"ContainerDied","Data":"b0cef7707416a0d05eef4e3507d659437398def515b10002e0bbe1230e8d0c74"} Nov 26 02:13:40 crc kubenswrapper[4766]: I1126 02:13:40.826680 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:13:40 crc kubenswrapper[4766]: E1126 02:13:40.827340 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:13:41 crc kubenswrapper[4766]: I1126 02:13:41.321723 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kwmp6" event={"ID":"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d","Type":"ContainerStarted","Data":"5e7c44ec5796703a9ce317033be633f9b8a68baf7eb28330066f63bc6ba28902"} Nov 26 02:13:41 crc kubenswrapper[4766]: I1126 02:13:41.343772 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kwmp6" podStartSLOduration=2.878376989 podStartE2EDuration="6.343753598s" podCreationTimestamp="2025-11-26 02:13:35 +0000 UTC" firstStartedPulling="2025-11-26 02:13:37.262000298 +0000 UTC m=+6598.110770728" lastFinishedPulling="2025-11-26 02:13:40.727376907 +0000 UTC m=+6601.576147337" observedRunningTime="2025-11-26 02:13:41.341680807 +0000 UTC m=+6602.190451247" watchObservedRunningTime="2025-11-26 02:13:41.343753598 +0000 UTC m=+6602.192524038" Nov 26 02:13:45 crc kubenswrapper[4766]: I1126 02:13:45.702946 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kwmp6" Nov 26 02:13:45 crc kubenswrapper[4766]: I1126 02:13:45.703388 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kwmp6" Nov 26 02:13:45 crc kubenswrapper[4766]: I1126 02:13:45.781645 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kwmp6" Nov 26 02:13:46 crc kubenswrapper[4766]: I1126 02:13:46.430630 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kwmp6" Nov 26 02:13:46 crc kubenswrapper[4766]: I1126 02:13:46.519220 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kwmp6"] Nov 26 02:13:48 crc kubenswrapper[4766]: I1126 02:13:48.406150 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kwmp6" podUID="2ee59fe9-db71-4771-9ca6-58e16ffbdb5d" containerName="registry-server" containerID="cri-o://5e7c44ec5796703a9ce317033be633f9b8a68baf7eb28330066f63bc6ba28902" gracePeriod=2 Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.119606 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kwmp6" Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.242452 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ee59fe9-db71-4771-9ca6-58e16ffbdb5d-catalog-content\") pod \"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d\" (UID: \"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d\") " Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.242553 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ee59fe9-db71-4771-9ca6-58e16ffbdb5d-utilities\") pod \"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d\" (UID: \"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d\") " Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.242760 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6n56\" (UniqueName: \"kubernetes.io/projected/2ee59fe9-db71-4771-9ca6-58e16ffbdb5d-kube-api-access-h6n56\") pod \"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d\" (UID: \"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d\") " Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.265058 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ee59fe9-db71-4771-9ca6-58e16ffbdb5d-utilities" (OuterVolumeSpecName: "utilities") pod "2ee59fe9-db71-4771-9ca6-58e16ffbdb5d" (UID: "2ee59fe9-db71-4771-9ca6-58e16ffbdb5d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.265900 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ee59fe9-db71-4771-9ca6-58e16ffbdb5d-kube-api-access-h6n56" (OuterVolumeSpecName: "kube-api-access-h6n56") pod "2ee59fe9-db71-4771-9ca6-58e16ffbdb5d" (UID: "2ee59fe9-db71-4771-9ca6-58e16ffbdb5d"). InnerVolumeSpecName "kube-api-access-h6n56". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.344789 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6n56\" (UniqueName: \"kubernetes.io/projected/2ee59fe9-db71-4771-9ca6-58e16ffbdb5d-kube-api-access-h6n56\") on node \"crc\" DevicePath \"\"" Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.345015 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ee59fe9-db71-4771-9ca6-58e16ffbdb5d-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.362506 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ee59fe9-db71-4771-9ca6-58e16ffbdb5d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ee59fe9-db71-4771-9ca6-58e16ffbdb5d" (UID: "2ee59fe9-db71-4771-9ca6-58e16ffbdb5d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.421489 4766 generic.go:334] "Generic (PLEG): container finished" podID="2ee59fe9-db71-4771-9ca6-58e16ffbdb5d" containerID="5e7c44ec5796703a9ce317033be633f9b8a68baf7eb28330066f63bc6ba28902" exitCode=0 Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.421533 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kwmp6" event={"ID":"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d","Type":"ContainerDied","Data":"5e7c44ec5796703a9ce317033be633f9b8a68baf7eb28330066f63bc6ba28902"} Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.421559 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kwmp6" Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.421583 4766 scope.go:117] "RemoveContainer" containerID="5e7c44ec5796703a9ce317033be633f9b8a68baf7eb28330066f63bc6ba28902" Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.421570 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kwmp6" event={"ID":"2ee59fe9-db71-4771-9ca6-58e16ffbdb5d","Type":"ContainerDied","Data":"6a6056d9d17567d0fe6629ff9fbb0ec00d733641f7df2a000b935ad54b876ac8"} Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.443082 4766 scope.go:117] "RemoveContainer" containerID="b0cef7707416a0d05eef4e3507d659437398def515b10002e0bbe1230e8d0c74" Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.447958 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ee59fe9-db71-4771-9ca6-58e16ffbdb5d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.457602 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kwmp6"] Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.467810 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kwmp6"] Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.473213 4766 scope.go:117] "RemoveContainer" containerID="70dc10cd62e2b646de789e104f03f1c92f71c10117f1022ae6f98d73fa4d0b2b" Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.531187 4766 scope.go:117] "RemoveContainer" containerID="5e7c44ec5796703a9ce317033be633f9b8a68baf7eb28330066f63bc6ba28902" Nov 26 02:13:49 crc kubenswrapper[4766]: E1126 02:13:49.535262 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e7c44ec5796703a9ce317033be633f9b8a68baf7eb28330066f63bc6ba28902\": container with ID starting with 5e7c44ec5796703a9ce317033be633f9b8a68baf7eb28330066f63bc6ba28902 not found: ID does not exist" containerID="5e7c44ec5796703a9ce317033be633f9b8a68baf7eb28330066f63bc6ba28902" Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.535304 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e7c44ec5796703a9ce317033be633f9b8a68baf7eb28330066f63bc6ba28902"} err="failed to get container status \"5e7c44ec5796703a9ce317033be633f9b8a68baf7eb28330066f63bc6ba28902\": rpc error: code = NotFound desc = could not find container \"5e7c44ec5796703a9ce317033be633f9b8a68baf7eb28330066f63bc6ba28902\": container with ID starting with 5e7c44ec5796703a9ce317033be633f9b8a68baf7eb28330066f63bc6ba28902 not found: ID does not exist" Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.535329 4766 scope.go:117] "RemoveContainer" containerID="b0cef7707416a0d05eef4e3507d659437398def515b10002e0bbe1230e8d0c74" Nov 26 02:13:49 crc kubenswrapper[4766]: E1126 02:13:49.538470 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0cef7707416a0d05eef4e3507d659437398def515b10002e0bbe1230e8d0c74\": container with ID starting with b0cef7707416a0d05eef4e3507d659437398def515b10002e0bbe1230e8d0c74 not found: ID does not exist" containerID="b0cef7707416a0d05eef4e3507d659437398def515b10002e0bbe1230e8d0c74" Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.538496 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0cef7707416a0d05eef4e3507d659437398def515b10002e0bbe1230e8d0c74"} err="failed to get container status \"b0cef7707416a0d05eef4e3507d659437398def515b10002e0bbe1230e8d0c74\": rpc error: code = NotFound desc = could not find container \"b0cef7707416a0d05eef4e3507d659437398def515b10002e0bbe1230e8d0c74\": container with ID starting with b0cef7707416a0d05eef4e3507d659437398def515b10002e0bbe1230e8d0c74 not found: ID does not exist" Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.538510 4766 scope.go:117] "RemoveContainer" containerID="70dc10cd62e2b646de789e104f03f1c92f71c10117f1022ae6f98d73fa4d0b2b" Nov 26 02:13:49 crc kubenswrapper[4766]: E1126 02:13:49.541732 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70dc10cd62e2b646de789e104f03f1c92f71c10117f1022ae6f98d73fa4d0b2b\": container with ID starting with 70dc10cd62e2b646de789e104f03f1c92f71c10117f1022ae6f98d73fa4d0b2b not found: ID does not exist" containerID="70dc10cd62e2b646de789e104f03f1c92f71c10117f1022ae6f98d73fa4d0b2b" Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.541949 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70dc10cd62e2b646de789e104f03f1c92f71c10117f1022ae6f98d73fa4d0b2b"} err="failed to get container status \"70dc10cd62e2b646de789e104f03f1c92f71c10117f1022ae6f98d73fa4d0b2b\": rpc error: code = NotFound desc = could not find container \"70dc10cd62e2b646de789e104f03f1c92f71c10117f1022ae6f98d73fa4d0b2b\": container with ID starting with 70dc10cd62e2b646de789e104f03f1c92f71c10117f1022ae6f98d73fa4d0b2b not found: ID does not exist" Nov 26 02:13:49 crc kubenswrapper[4766]: I1126 02:13:49.839946 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ee59fe9-db71-4771-9ca6-58e16ffbdb5d" path="/var/lib/kubelet/pods/2ee59fe9-db71-4771-9ca6-58e16ffbdb5d/volumes" Nov 26 02:13:54 crc kubenswrapper[4766]: I1126 02:13:54.826852 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:13:54 crc kubenswrapper[4766]: E1126 02:13:54.828006 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:14:08 crc kubenswrapper[4766]: I1126 02:14:08.828171 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:14:08 crc kubenswrapper[4766]: E1126 02:14:08.829373 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:14:19 crc kubenswrapper[4766]: I1126 02:14:19.837697 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:14:19 crc kubenswrapper[4766]: E1126 02:14:19.838601 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:14:32 crc kubenswrapper[4766]: I1126 02:14:32.827273 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:14:32 crc kubenswrapper[4766]: E1126 02:14:32.828524 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:14:44 crc kubenswrapper[4766]: I1126 02:14:44.827767 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:14:44 crc kubenswrapper[4766]: E1126 02:14:44.828912 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:14:59 crc kubenswrapper[4766]: I1126 02:14:59.843419 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:14:59 crc kubenswrapper[4766]: E1126 02:14:59.844394 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:15:00 crc kubenswrapper[4766]: I1126 02:15:00.194571 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402055-74p2c"] Nov 26 02:15:00 crc kubenswrapper[4766]: E1126 02:15:00.195140 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee59fe9-db71-4771-9ca6-58e16ffbdb5d" containerName="registry-server" Nov 26 02:15:00 crc kubenswrapper[4766]: I1126 02:15:00.195162 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee59fe9-db71-4771-9ca6-58e16ffbdb5d" containerName="registry-server" Nov 26 02:15:00 crc kubenswrapper[4766]: E1126 02:15:00.195212 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee59fe9-db71-4771-9ca6-58e16ffbdb5d" containerName="extract-content" Nov 26 02:15:00 crc kubenswrapper[4766]: I1126 02:15:00.195222 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee59fe9-db71-4771-9ca6-58e16ffbdb5d" containerName="extract-content" Nov 26 02:15:00 crc kubenswrapper[4766]: E1126 02:15:00.195234 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee59fe9-db71-4771-9ca6-58e16ffbdb5d" containerName="extract-utilities" Nov 26 02:15:00 crc kubenswrapper[4766]: I1126 02:15:00.195243 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee59fe9-db71-4771-9ca6-58e16ffbdb5d" containerName="extract-utilities" Nov 26 02:15:00 crc kubenswrapper[4766]: I1126 02:15:00.195521 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ee59fe9-db71-4771-9ca6-58e16ffbdb5d" containerName="registry-server" Nov 26 02:15:00 crc kubenswrapper[4766]: I1126 02:15:00.196464 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402055-74p2c" Nov 26 02:15:00 crc kubenswrapper[4766]: I1126 02:15:00.210834 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 26 02:15:00 crc kubenswrapper[4766]: I1126 02:15:00.211938 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 26 02:15:00 crc kubenswrapper[4766]: I1126 02:15:00.225285 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402055-74p2c"] Nov 26 02:15:00 crc kubenswrapper[4766]: I1126 02:15:00.280944 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7-config-volume\") pod \"collect-profiles-29402055-74p2c\" (UID: \"bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402055-74p2c" Nov 26 02:15:00 crc kubenswrapper[4766]: I1126 02:15:00.281004 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwvjr\" (UniqueName: \"kubernetes.io/projected/bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7-kube-api-access-qwvjr\") pod \"collect-profiles-29402055-74p2c\" (UID: \"bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402055-74p2c" Nov 26 02:15:00 crc kubenswrapper[4766]: I1126 02:15:00.281140 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7-secret-volume\") pod \"collect-profiles-29402055-74p2c\" (UID: \"bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402055-74p2c" Nov 26 02:15:00 crc kubenswrapper[4766]: I1126 02:15:00.383252 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7-config-volume\") pod \"collect-profiles-29402055-74p2c\" (UID: \"bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402055-74p2c" Nov 26 02:15:00 crc kubenswrapper[4766]: I1126 02:15:00.383316 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwvjr\" (UniqueName: \"kubernetes.io/projected/bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7-kube-api-access-qwvjr\") pod \"collect-profiles-29402055-74p2c\" (UID: \"bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402055-74p2c" Nov 26 02:15:00 crc kubenswrapper[4766]: I1126 02:15:00.383441 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7-secret-volume\") pod \"collect-profiles-29402055-74p2c\" (UID: \"bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402055-74p2c" Nov 26 02:15:00 crc kubenswrapper[4766]: I1126 02:15:00.384625 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7-config-volume\") pod \"collect-profiles-29402055-74p2c\" (UID: \"bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402055-74p2c" Nov 26 02:15:00 crc kubenswrapper[4766]: I1126 02:15:00.399446 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7-secret-volume\") pod \"collect-profiles-29402055-74p2c\" (UID: \"bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402055-74p2c" Nov 26 02:15:00 crc kubenswrapper[4766]: I1126 02:15:00.401098 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwvjr\" (UniqueName: \"kubernetes.io/projected/bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7-kube-api-access-qwvjr\") pod \"collect-profiles-29402055-74p2c\" (UID: \"bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402055-74p2c" Nov 26 02:15:00 crc kubenswrapper[4766]: I1126 02:15:00.530071 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402055-74p2c" Nov 26 02:15:01 crc kubenswrapper[4766]: I1126 02:15:01.026466 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402055-74p2c"] Nov 26 02:15:01 crc kubenswrapper[4766]: I1126 02:15:01.329792 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402055-74p2c" event={"ID":"bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7","Type":"ContainerStarted","Data":"0c24cd9fa00c83ec5509ab75d53fc6fde56407696e54d9b33355aed4b7b70f5f"} Nov 26 02:15:01 crc kubenswrapper[4766]: I1126 02:15:01.330043 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402055-74p2c" event={"ID":"bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7","Type":"ContainerStarted","Data":"c70eeb85fc05f96e3dd2000f6ba05fb36869c25fce872fdbdb58e9e295c5bd93"} Nov 26 02:15:01 crc kubenswrapper[4766]: I1126 02:15:01.349287 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29402055-74p2c" podStartSLOduration=1.3492697009999999 podStartE2EDuration="1.349269701s" podCreationTimestamp="2025-11-26 02:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 02:15:01.345886709 +0000 UTC m=+6682.194657149" watchObservedRunningTime="2025-11-26 02:15:01.349269701 +0000 UTC m=+6682.198040131" Nov 26 02:15:02 crc kubenswrapper[4766]: I1126 02:15:02.346264 4766 generic.go:334] "Generic (PLEG): container finished" podID="bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7" containerID="0c24cd9fa00c83ec5509ab75d53fc6fde56407696e54d9b33355aed4b7b70f5f" exitCode=0 Nov 26 02:15:02 crc kubenswrapper[4766]: I1126 02:15:02.346577 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402055-74p2c" event={"ID":"bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7","Type":"ContainerDied","Data":"0c24cd9fa00c83ec5509ab75d53fc6fde56407696e54d9b33355aed4b7b70f5f"} Nov 26 02:15:03 crc kubenswrapper[4766]: I1126 02:15:03.859102 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402055-74p2c" Nov 26 02:15:03 crc kubenswrapper[4766]: I1126 02:15:03.980492 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7-config-volume\") pod \"bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7\" (UID: \"bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7\") " Nov 26 02:15:03 crc kubenswrapper[4766]: I1126 02:15:03.980926 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7-secret-volume\") pod \"bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7\" (UID: \"bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7\") " Nov 26 02:15:03 crc kubenswrapper[4766]: I1126 02:15:03.981191 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwvjr\" (UniqueName: \"kubernetes.io/projected/bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7-kube-api-access-qwvjr\") pod \"bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7\" (UID: \"bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7\") " Nov 26 02:15:03 crc kubenswrapper[4766]: I1126 02:15:03.981475 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7-config-volume" (OuterVolumeSpecName: "config-volume") pod "bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7" (UID: "bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 02:15:03 crc kubenswrapper[4766]: I1126 02:15:03.981869 4766 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 02:15:03 crc kubenswrapper[4766]: I1126 02:15:03.987233 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7-kube-api-access-qwvjr" (OuterVolumeSpecName: "kube-api-access-qwvjr") pod "bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7" (UID: "bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7"). InnerVolumeSpecName "kube-api-access-qwvjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:15:03 crc kubenswrapper[4766]: I1126 02:15:03.987859 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7" (UID: "bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 02:15:04 crc kubenswrapper[4766]: I1126 02:15:04.083904 4766 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 26 02:15:04 crc kubenswrapper[4766]: I1126 02:15:04.084116 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwvjr\" (UniqueName: \"kubernetes.io/projected/bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7-kube-api-access-qwvjr\") on node \"crc\" DevicePath \"\"" Nov 26 02:15:04 crc kubenswrapper[4766]: I1126 02:15:04.371971 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402055-74p2c" event={"ID":"bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7","Type":"ContainerDied","Data":"c70eeb85fc05f96e3dd2000f6ba05fb36869c25fce872fdbdb58e9e295c5bd93"} Nov 26 02:15:04 crc kubenswrapper[4766]: I1126 02:15:04.372025 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c70eeb85fc05f96e3dd2000f6ba05fb36869c25fce872fdbdb58e9e295c5bd93" Nov 26 02:15:04 crc kubenswrapper[4766]: I1126 02:15:04.372732 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402055-74p2c" Nov 26 02:15:04 crc kubenswrapper[4766]: I1126 02:15:04.480629 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw"] Nov 26 02:15:04 crc kubenswrapper[4766]: I1126 02:15:04.491380 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402010-j96pw"] Nov 26 02:15:05 crc kubenswrapper[4766]: I1126 02:15:05.842242 4766 scope.go:117] "RemoveContainer" containerID="26e6f8f0c2f7670c708b926520f3d7313f72a5d8dd65b7856b61162c5f475dd5" Nov 26 02:15:05 crc kubenswrapper[4766]: I1126 02:15:05.849204 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccb5619d-03b6-4fae-9e5d-950952eb0afe" path="/var/lib/kubelet/pods/ccb5619d-03b6-4fae-9e5d-950952eb0afe/volumes" Nov 26 02:15:13 crc kubenswrapper[4766]: I1126 02:15:13.827681 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:15:14 crc kubenswrapper[4766]: I1126 02:15:14.509757 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"f3b79bb97115423a9ddb10f5e7bc376bff4f35b58a4fe41d6b2ad3dfaa0c3a8c"} Nov 26 02:15:29 crc kubenswrapper[4766]: I1126 02:15:29.952519 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s6c7v"] Nov 26 02:15:29 crc kubenswrapper[4766]: E1126 02:15:29.953743 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7" containerName="collect-profiles" Nov 26 02:15:29 crc kubenswrapper[4766]: I1126 02:15:29.953762 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7" containerName="collect-profiles" Nov 26 02:15:29 crc kubenswrapper[4766]: I1126 02:15:29.954091 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdbfa1cc-596b-4c1f-a797-5e5e3fc1aba7" containerName="collect-profiles" Nov 26 02:15:29 crc kubenswrapper[4766]: I1126 02:15:29.956736 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s6c7v" Nov 26 02:15:29 crc kubenswrapper[4766]: I1126 02:15:29.979877 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s6c7v"] Nov 26 02:15:30 crc kubenswrapper[4766]: I1126 02:15:30.019282 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36a6d272-cf25-45fe-93e6-527f2d5c4391-catalog-content\") pod \"community-operators-s6c7v\" (UID: \"36a6d272-cf25-45fe-93e6-527f2d5c4391\") " pod="openshift-marketplace/community-operators-s6c7v" Nov 26 02:15:30 crc kubenswrapper[4766]: I1126 02:15:30.019340 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl9bp\" (UniqueName: \"kubernetes.io/projected/36a6d272-cf25-45fe-93e6-527f2d5c4391-kube-api-access-zl9bp\") pod \"community-operators-s6c7v\" (UID: \"36a6d272-cf25-45fe-93e6-527f2d5c4391\") " pod="openshift-marketplace/community-operators-s6c7v" Nov 26 02:15:30 crc kubenswrapper[4766]: I1126 02:15:30.019977 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36a6d272-cf25-45fe-93e6-527f2d5c4391-utilities\") pod \"community-operators-s6c7v\" (UID: \"36a6d272-cf25-45fe-93e6-527f2d5c4391\") " pod="openshift-marketplace/community-operators-s6c7v" Nov 26 02:15:30 crc kubenswrapper[4766]: I1126 02:15:30.121576 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36a6d272-cf25-45fe-93e6-527f2d5c4391-catalog-content\") pod \"community-operators-s6c7v\" (UID: \"36a6d272-cf25-45fe-93e6-527f2d5c4391\") " pod="openshift-marketplace/community-operators-s6c7v" Nov 26 02:15:30 crc kubenswrapper[4766]: I1126 02:15:30.121620 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl9bp\" (UniqueName: \"kubernetes.io/projected/36a6d272-cf25-45fe-93e6-527f2d5c4391-kube-api-access-zl9bp\") pod \"community-operators-s6c7v\" (UID: \"36a6d272-cf25-45fe-93e6-527f2d5c4391\") " pod="openshift-marketplace/community-operators-s6c7v" Nov 26 02:15:30 crc kubenswrapper[4766]: I1126 02:15:30.121719 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36a6d272-cf25-45fe-93e6-527f2d5c4391-utilities\") pod \"community-operators-s6c7v\" (UID: \"36a6d272-cf25-45fe-93e6-527f2d5c4391\") " pod="openshift-marketplace/community-operators-s6c7v" Nov 26 02:15:30 crc kubenswrapper[4766]: I1126 02:15:30.122233 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36a6d272-cf25-45fe-93e6-527f2d5c4391-utilities\") pod \"community-operators-s6c7v\" (UID: \"36a6d272-cf25-45fe-93e6-527f2d5c4391\") " pod="openshift-marketplace/community-operators-s6c7v" Nov 26 02:15:30 crc kubenswrapper[4766]: I1126 02:15:30.122345 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36a6d272-cf25-45fe-93e6-527f2d5c4391-catalog-content\") pod \"community-operators-s6c7v\" (UID: \"36a6d272-cf25-45fe-93e6-527f2d5c4391\") " pod="openshift-marketplace/community-operators-s6c7v" Nov 26 02:15:30 crc kubenswrapper[4766]: I1126 02:15:30.149974 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl9bp\" (UniqueName: \"kubernetes.io/projected/36a6d272-cf25-45fe-93e6-527f2d5c4391-kube-api-access-zl9bp\") pod \"community-operators-s6c7v\" (UID: \"36a6d272-cf25-45fe-93e6-527f2d5c4391\") " pod="openshift-marketplace/community-operators-s6c7v" Nov 26 02:15:30 crc kubenswrapper[4766]: I1126 02:15:30.290086 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s6c7v" Nov 26 02:15:30 crc kubenswrapper[4766]: I1126 02:15:30.805606 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s6c7v"] Nov 26 02:15:31 crc kubenswrapper[4766]: I1126 02:15:31.702352 4766 generic.go:334] "Generic (PLEG): container finished" podID="36a6d272-cf25-45fe-93e6-527f2d5c4391" containerID="fb7fcc747ac987f0024f00deee4a053f4f7f4545a6bae81745e69342403fa50c" exitCode=0 Nov 26 02:15:31 crc kubenswrapper[4766]: I1126 02:15:31.702463 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s6c7v" event={"ID":"36a6d272-cf25-45fe-93e6-527f2d5c4391","Type":"ContainerDied","Data":"fb7fcc747ac987f0024f00deee4a053f4f7f4545a6bae81745e69342403fa50c"} Nov 26 02:15:31 crc kubenswrapper[4766]: I1126 02:15:31.702903 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s6c7v" event={"ID":"36a6d272-cf25-45fe-93e6-527f2d5c4391","Type":"ContainerStarted","Data":"d0914ebb6aa4871196f49f099fa3c32e5aee3ae570eeedd0b59206aed4a2f22b"} Nov 26 02:15:32 crc kubenswrapper[4766]: I1126 02:15:32.721280 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s6c7v" event={"ID":"36a6d272-cf25-45fe-93e6-527f2d5c4391","Type":"ContainerStarted","Data":"3b61354044617993fcf280f63d196cbdb9b56edb821d3b2646e91e2f742b3e99"} Nov 26 02:15:34 crc kubenswrapper[4766]: I1126 02:15:34.747623 4766 generic.go:334] "Generic (PLEG): container finished" podID="36a6d272-cf25-45fe-93e6-527f2d5c4391" containerID="3b61354044617993fcf280f63d196cbdb9b56edb821d3b2646e91e2f742b3e99" exitCode=0 Nov 26 02:15:34 crc kubenswrapper[4766]: I1126 02:15:34.747691 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s6c7v" event={"ID":"36a6d272-cf25-45fe-93e6-527f2d5c4391","Type":"ContainerDied","Data":"3b61354044617993fcf280f63d196cbdb9b56edb821d3b2646e91e2f742b3e99"} Nov 26 02:15:35 crc kubenswrapper[4766]: I1126 02:15:35.761911 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s6c7v" event={"ID":"36a6d272-cf25-45fe-93e6-527f2d5c4391","Type":"ContainerStarted","Data":"0869fc1507a82dc644cd5ba7c54e77606527a2e9711cfcc9700da270b0e7cfa0"} Nov 26 02:15:35 crc kubenswrapper[4766]: I1126 02:15:35.790738 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s6c7v" podStartSLOduration=3.363573622 podStartE2EDuration="6.790720222s" podCreationTimestamp="2025-11-26 02:15:29 +0000 UTC" firstStartedPulling="2025-11-26 02:15:31.705025588 +0000 UTC m=+6712.553796048" lastFinishedPulling="2025-11-26 02:15:35.132172208 +0000 UTC m=+6715.980942648" observedRunningTime="2025-11-26 02:15:35.785059625 +0000 UTC m=+6716.633830055" watchObservedRunningTime="2025-11-26 02:15:35.790720222 +0000 UTC m=+6716.639490652" Nov 26 02:15:40 crc kubenswrapper[4766]: I1126 02:15:40.290257 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s6c7v" Nov 26 02:15:40 crc kubenswrapper[4766]: I1126 02:15:40.290831 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s6c7v" Nov 26 02:15:41 crc kubenswrapper[4766]: I1126 02:15:41.353581 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-s6c7v" podUID="36a6d272-cf25-45fe-93e6-527f2d5c4391" containerName="registry-server" probeResult="failure" output=< Nov 26 02:15:41 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 02:15:41 crc kubenswrapper[4766]: > Nov 26 02:15:50 crc kubenswrapper[4766]: I1126 02:15:50.364376 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s6c7v" Nov 26 02:15:50 crc kubenswrapper[4766]: I1126 02:15:50.422603 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s6c7v" Nov 26 02:15:50 crc kubenswrapper[4766]: I1126 02:15:50.604137 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s6c7v"] Nov 26 02:15:51 crc kubenswrapper[4766]: I1126 02:15:51.986114 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-s6c7v" podUID="36a6d272-cf25-45fe-93e6-527f2d5c4391" containerName="registry-server" containerID="cri-o://0869fc1507a82dc644cd5ba7c54e77606527a2e9711cfcc9700da270b0e7cfa0" gracePeriod=2 Nov 26 02:15:52 crc kubenswrapper[4766]: I1126 02:15:52.702598 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s6c7v" Nov 26 02:15:52 crc kubenswrapper[4766]: I1126 02:15:52.793769 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36a6d272-cf25-45fe-93e6-527f2d5c4391-catalog-content\") pod \"36a6d272-cf25-45fe-93e6-527f2d5c4391\" (UID: \"36a6d272-cf25-45fe-93e6-527f2d5c4391\") " Nov 26 02:15:52 crc kubenswrapper[4766]: I1126 02:15:52.793820 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36a6d272-cf25-45fe-93e6-527f2d5c4391-utilities\") pod \"36a6d272-cf25-45fe-93e6-527f2d5c4391\" (UID: \"36a6d272-cf25-45fe-93e6-527f2d5c4391\") " Nov 26 02:15:52 crc kubenswrapper[4766]: I1126 02:15:52.793881 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zl9bp\" (UniqueName: \"kubernetes.io/projected/36a6d272-cf25-45fe-93e6-527f2d5c4391-kube-api-access-zl9bp\") pod \"36a6d272-cf25-45fe-93e6-527f2d5c4391\" (UID: \"36a6d272-cf25-45fe-93e6-527f2d5c4391\") " Nov 26 02:15:52 crc kubenswrapper[4766]: I1126 02:15:52.795544 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36a6d272-cf25-45fe-93e6-527f2d5c4391-utilities" (OuterVolumeSpecName: "utilities") pod "36a6d272-cf25-45fe-93e6-527f2d5c4391" (UID: "36a6d272-cf25-45fe-93e6-527f2d5c4391"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:15:52 crc kubenswrapper[4766]: I1126 02:15:52.800107 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36a6d272-cf25-45fe-93e6-527f2d5c4391-kube-api-access-zl9bp" (OuterVolumeSpecName: "kube-api-access-zl9bp") pod "36a6d272-cf25-45fe-93e6-527f2d5c4391" (UID: "36a6d272-cf25-45fe-93e6-527f2d5c4391"). InnerVolumeSpecName "kube-api-access-zl9bp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:15:52 crc kubenswrapper[4766]: I1126 02:15:52.838549 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36a6d272-cf25-45fe-93e6-527f2d5c4391-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "36a6d272-cf25-45fe-93e6-527f2d5c4391" (UID: "36a6d272-cf25-45fe-93e6-527f2d5c4391"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:15:52 crc kubenswrapper[4766]: I1126 02:15:52.897285 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36a6d272-cf25-45fe-93e6-527f2d5c4391-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 02:15:52 crc kubenswrapper[4766]: I1126 02:15:52.897332 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36a6d272-cf25-45fe-93e6-527f2d5c4391-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 02:15:52 crc kubenswrapper[4766]: I1126 02:15:52.897353 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zl9bp\" (UniqueName: \"kubernetes.io/projected/36a6d272-cf25-45fe-93e6-527f2d5c4391-kube-api-access-zl9bp\") on node \"crc\" DevicePath \"\"" Nov 26 02:15:53 crc kubenswrapper[4766]: I1126 02:15:53.002210 4766 generic.go:334] "Generic (PLEG): container finished" podID="36a6d272-cf25-45fe-93e6-527f2d5c4391" containerID="0869fc1507a82dc644cd5ba7c54e77606527a2e9711cfcc9700da270b0e7cfa0" exitCode=0 Nov 26 02:15:53 crc kubenswrapper[4766]: I1126 02:15:53.002263 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s6c7v" event={"ID":"36a6d272-cf25-45fe-93e6-527f2d5c4391","Type":"ContainerDied","Data":"0869fc1507a82dc644cd5ba7c54e77606527a2e9711cfcc9700da270b0e7cfa0"} Nov 26 02:15:53 crc kubenswrapper[4766]: I1126 02:15:53.002282 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s6c7v" Nov 26 02:15:53 crc kubenswrapper[4766]: I1126 02:15:53.002302 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s6c7v" event={"ID":"36a6d272-cf25-45fe-93e6-527f2d5c4391","Type":"ContainerDied","Data":"d0914ebb6aa4871196f49f099fa3c32e5aee3ae570eeedd0b59206aed4a2f22b"} Nov 26 02:15:53 crc kubenswrapper[4766]: I1126 02:15:53.002326 4766 scope.go:117] "RemoveContainer" containerID="0869fc1507a82dc644cd5ba7c54e77606527a2e9711cfcc9700da270b0e7cfa0" Nov 26 02:15:53 crc kubenswrapper[4766]: I1126 02:15:53.054551 4766 scope.go:117] "RemoveContainer" containerID="3b61354044617993fcf280f63d196cbdb9b56edb821d3b2646e91e2f742b3e99" Nov 26 02:15:53 crc kubenswrapper[4766]: I1126 02:15:53.055569 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s6c7v"] Nov 26 02:15:53 crc kubenswrapper[4766]: I1126 02:15:53.070824 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-s6c7v"] Nov 26 02:15:53 crc kubenswrapper[4766]: I1126 02:15:53.088541 4766 scope.go:117] "RemoveContainer" containerID="fb7fcc747ac987f0024f00deee4a053f4f7f4545a6bae81745e69342403fa50c" Nov 26 02:15:53 crc kubenswrapper[4766]: I1126 02:15:53.140553 4766 scope.go:117] "RemoveContainer" containerID="0869fc1507a82dc644cd5ba7c54e77606527a2e9711cfcc9700da270b0e7cfa0" Nov 26 02:15:53 crc kubenswrapper[4766]: E1126 02:15:53.141147 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0869fc1507a82dc644cd5ba7c54e77606527a2e9711cfcc9700da270b0e7cfa0\": container with ID starting with 0869fc1507a82dc644cd5ba7c54e77606527a2e9711cfcc9700da270b0e7cfa0 not found: ID does not exist" containerID="0869fc1507a82dc644cd5ba7c54e77606527a2e9711cfcc9700da270b0e7cfa0" Nov 26 02:15:53 crc kubenswrapper[4766]: I1126 02:15:53.141200 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0869fc1507a82dc644cd5ba7c54e77606527a2e9711cfcc9700da270b0e7cfa0"} err="failed to get container status \"0869fc1507a82dc644cd5ba7c54e77606527a2e9711cfcc9700da270b0e7cfa0\": rpc error: code = NotFound desc = could not find container \"0869fc1507a82dc644cd5ba7c54e77606527a2e9711cfcc9700da270b0e7cfa0\": container with ID starting with 0869fc1507a82dc644cd5ba7c54e77606527a2e9711cfcc9700da270b0e7cfa0 not found: ID does not exist" Nov 26 02:15:53 crc kubenswrapper[4766]: I1126 02:15:53.141231 4766 scope.go:117] "RemoveContainer" containerID="3b61354044617993fcf280f63d196cbdb9b56edb821d3b2646e91e2f742b3e99" Nov 26 02:15:53 crc kubenswrapper[4766]: E1126 02:15:53.141634 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b61354044617993fcf280f63d196cbdb9b56edb821d3b2646e91e2f742b3e99\": container with ID starting with 3b61354044617993fcf280f63d196cbdb9b56edb821d3b2646e91e2f742b3e99 not found: ID does not exist" containerID="3b61354044617993fcf280f63d196cbdb9b56edb821d3b2646e91e2f742b3e99" Nov 26 02:15:53 crc kubenswrapper[4766]: I1126 02:15:53.141686 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b61354044617993fcf280f63d196cbdb9b56edb821d3b2646e91e2f742b3e99"} err="failed to get container status \"3b61354044617993fcf280f63d196cbdb9b56edb821d3b2646e91e2f742b3e99\": rpc error: code = NotFound desc = could not find container \"3b61354044617993fcf280f63d196cbdb9b56edb821d3b2646e91e2f742b3e99\": container with ID starting with 3b61354044617993fcf280f63d196cbdb9b56edb821d3b2646e91e2f742b3e99 not found: ID does not exist" Nov 26 02:15:53 crc kubenswrapper[4766]: I1126 02:15:53.141711 4766 scope.go:117] "RemoveContainer" containerID="fb7fcc747ac987f0024f00deee4a053f4f7f4545a6bae81745e69342403fa50c" Nov 26 02:15:53 crc kubenswrapper[4766]: E1126 02:15:53.142024 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb7fcc747ac987f0024f00deee4a053f4f7f4545a6bae81745e69342403fa50c\": container with ID starting with fb7fcc747ac987f0024f00deee4a053f4f7f4545a6bae81745e69342403fa50c not found: ID does not exist" containerID="fb7fcc747ac987f0024f00deee4a053f4f7f4545a6bae81745e69342403fa50c" Nov 26 02:15:53 crc kubenswrapper[4766]: I1126 02:15:53.142117 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb7fcc747ac987f0024f00deee4a053f4f7f4545a6bae81745e69342403fa50c"} err="failed to get container status \"fb7fcc747ac987f0024f00deee4a053f4f7f4545a6bae81745e69342403fa50c\": rpc error: code = NotFound desc = could not find container \"fb7fcc747ac987f0024f00deee4a053f4f7f4545a6bae81745e69342403fa50c\": container with ID starting with fb7fcc747ac987f0024f00deee4a053f4f7f4545a6bae81745e69342403fa50c not found: ID does not exist" Nov 26 02:15:53 crc kubenswrapper[4766]: I1126 02:15:53.839960 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36a6d272-cf25-45fe-93e6-527f2d5c4391" path="/var/lib/kubelet/pods/36a6d272-cf25-45fe-93e6-527f2d5c4391/volumes" Nov 26 02:17:41 crc kubenswrapper[4766]: I1126 02:17:41.480109 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:17:41 crc kubenswrapper[4766]: I1126 02:17:41.482408 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:18:11 crc kubenswrapper[4766]: I1126 02:18:11.479386 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:18:11 crc kubenswrapper[4766]: I1126 02:18:11.482321 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:18:41 crc kubenswrapper[4766]: I1126 02:18:41.480217 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:18:41 crc kubenswrapper[4766]: I1126 02:18:41.480828 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:18:41 crc kubenswrapper[4766]: I1126 02:18:41.480884 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 02:18:41 crc kubenswrapper[4766]: I1126 02:18:41.481923 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f3b79bb97115423a9ddb10f5e7bc376bff4f35b58a4fe41d6b2ad3dfaa0c3a8c"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 02:18:41 crc kubenswrapper[4766]: I1126 02:18:41.481996 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://f3b79bb97115423a9ddb10f5e7bc376bff4f35b58a4fe41d6b2ad3dfaa0c3a8c" gracePeriod=600 Nov 26 02:18:42 crc kubenswrapper[4766]: I1126 02:18:42.234705 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="f3b79bb97115423a9ddb10f5e7bc376bff4f35b58a4fe41d6b2ad3dfaa0c3a8c" exitCode=0 Nov 26 02:18:42 crc kubenswrapper[4766]: I1126 02:18:42.234776 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"f3b79bb97115423a9ddb10f5e7bc376bff4f35b58a4fe41d6b2ad3dfaa0c3a8c"} Nov 26 02:18:42 crc kubenswrapper[4766]: I1126 02:18:42.235142 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562"} Nov 26 02:18:42 crc kubenswrapper[4766]: I1126 02:18:42.235164 4766 scope.go:117] "RemoveContainer" containerID="a9899df1ab026cd99352a8bed4287c74d2c87e3cf01a89174b93a8129b83eb9b" Nov 26 02:18:49 crc kubenswrapper[4766]: I1126 02:18:49.032434 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p68rz"] Nov 26 02:18:49 crc kubenswrapper[4766]: E1126 02:18:49.033500 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36a6d272-cf25-45fe-93e6-527f2d5c4391" containerName="extract-content" Nov 26 02:18:49 crc kubenswrapper[4766]: I1126 02:18:49.033516 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="36a6d272-cf25-45fe-93e6-527f2d5c4391" containerName="extract-content" Nov 26 02:18:49 crc kubenswrapper[4766]: E1126 02:18:49.033562 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36a6d272-cf25-45fe-93e6-527f2d5c4391" containerName="extract-utilities" Nov 26 02:18:49 crc kubenswrapper[4766]: I1126 02:18:49.033570 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="36a6d272-cf25-45fe-93e6-527f2d5c4391" containerName="extract-utilities" Nov 26 02:18:49 crc kubenswrapper[4766]: E1126 02:18:49.033689 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36a6d272-cf25-45fe-93e6-527f2d5c4391" containerName="registry-server" Nov 26 02:18:49 crc kubenswrapper[4766]: I1126 02:18:49.033700 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="36a6d272-cf25-45fe-93e6-527f2d5c4391" containerName="registry-server" Nov 26 02:18:49 crc kubenswrapper[4766]: I1126 02:18:49.033996 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="36a6d272-cf25-45fe-93e6-527f2d5c4391" containerName="registry-server" Nov 26 02:18:49 crc kubenswrapper[4766]: I1126 02:18:49.036166 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p68rz" Nov 26 02:18:49 crc kubenswrapper[4766]: I1126 02:18:49.046010 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p68rz"] Nov 26 02:18:49 crc kubenswrapper[4766]: I1126 02:18:49.152943 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/074017a4-5c1e-4dbb-b74c-ab60c1dd53b7-catalog-content\") pod \"redhat-marketplace-p68rz\" (UID: \"074017a4-5c1e-4dbb-b74c-ab60c1dd53b7\") " pod="openshift-marketplace/redhat-marketplace-p68rz" Nov 26 02:18:49 crc kubenswrapper[4766]: I1126 02:18:49.152990 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/074017a4-5c1e-4dbb-b74c-ab60c1dd53b7-utilities\") pod \"redhat-marketplace-p68rz\" (UID: \"074017a4-5c1e-4dbb-b74c-ab60c1dd53b7\") " pod="openshift-marketplace/redhat-marketplace-p68rz" Nov 26 02:18:49 crc kubenswrapper[4766]: I1126 02:18:49.153086 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7phm\" (UniqueName: \"kubernetes.io/projected/074017a4-5c1e-4dbb-b74c-ab60c1dd53b7-kube-api-access-r7phm\") pod \"redhat-marketplace-p68rz\" (UID: \"074017a4-5c1e-4dbb-b74c-ab60c1dd53b7\") " pod="openshift-marketplace/redhat-marketplace-p68rz" Nov 26 02:18:49 crc kubenswrapper[4766]: I1126 02:18:49.255223 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/074017a4-5c1e-4dbb-b74c-ab60c1dd53b7-catalog-content\") pod \"redhat-marketplace-p68rz\" (UID: \"074017a4-5c1e-4dbb-b74c-ab60c1dd53b7\") " pod="openshift-marketplace/redhat-marketplace-p68rz" Nov 26 02:18:49 crc kubenswrapper[4766]: I1126 02:18:49.255265 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/074017a4-5c1e-4dbb-b74c-ab60c1dd53b7-utilities\") pod \"redhat-marketplace-p68rz\" (UID: \"074017a4-5c1e-4dbb-b74c-ab60c1dd53b7\") " pod="openshift-marketplace/redhat-marketplace-p68rz" Nov 26 02:18:49 crc kubenswrapper[4766]: I1126 02:18:49.255318 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7phm\" (UniqueName: \"kubernetes.io/projected/074017a4-5c1e-4dbb-b74c-ab60c1dd53b7-kube-api-access-r7phm\") pod \"redhat-marketplace-p68rz\" (UID: \"074017a4-5c1e-4dbb-b74c-ab60c1dd53b7\") " pod="openshift-marketplace/redhat-marketplace-p68rz" Nov 26 02:18:49 crc kubenswrapper[4766]: I1126 02:18:49.256116 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/074017a4-5c1e-4dbb-b74c-ab60c1dd53b7-catalog-content\") pod \"redhat-marketplace-p68rz\" (UID: \"074017a4-5c1e-4dbb-b74c-ab60c1dd53b7\") " pod="openshift-marketplace/redhat-marketplace-p68rz" Nov 26 02:18:49 crc kubenswrapper[4766]: I1126 02:18:49.256127 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/074017a4-5c1e-4dbb-b74c-ab60c1dd53b7-utilities\") pod \"redhat-marketplace-p68rz\" (UID: \"074017a4-5c1e-4dbb-b74c-ab60c1dd53b7\") " pod="openshift-marketplace/redhat-marketplace-p68rz" Nov 26 02:18:49 crc kubenswrapper[4766]: I1126 02:18:49.286251 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7phm\" (UniqueName: \"kubernetes.io/projected/074017a4-5c1e-4dbb-b74c-ab60c1dd53b7-kube-api-access-r7phm\") pod \"redhat-marketplace-p68rz\" (UID: \"074017a4-5c1e-4dbb-b74c-ab60c1dd53b7\") " pod="openshift-marketplace/redhat-marketplace-p68rz" Nov 26 02:18:49 crc kubenswrapper[4766]: I1126 02:18:49.361137 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p68rz" Nov 26 02:18:49 crc kubenswrapper[4766]: I1126 02:18:49.866682 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p68rz"] Nov 26 02:18:49 crc kubenswrapper[4766]: W1126 02:18:49.874911 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod074017a4_5c1e_4dbb_b74c_ab60c1dd53b7.slice/crio-6fee58c12fb8b9f85909781a71d492520982d3e7b2a64547bb2f4897e49e6a31 WatchSource:0}: Error finding container 6fee58c12fb8b9f85909781a71d492520982d3e7b2a64547bb2f4897e49e6a31: Status 404 returned error can't find the container with id 6fee58c12fb8b9f85909781a71d492520982d3e7b2a64547bb2f4897e49e6a31 Nov 26 02:18:50 crc kubenswrapper[4766]: I1126 02:18:50.333382 4766 generic.go:334] "Generic (PLEG): container finished" podID="074017a4-5c1e-4dbb-b74c-ab60c1dd53b7" containerID="d0855c1f2774ccd5e1b07d9ae9de53a2b593160952ae65ecf23b7be6784fbc5f" exitCode=0 Nov 26 02:18:50 crc kubenswrapper[4766]: I1126 02:18:50.333479 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p68rz" event={"ID":"074017a4-5c1e-4dbb-b74c-ab60c1dd53b7","Type":"ContainerDied","Data":"d0855c1f2774ccd5e1b07d9ae9de53a2b593160952ae65ecf23b7be6784fbc5f"} Nov 26 02:18:50 crc kubenswrapper[4766]: I1126 02:18:50.333928 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p68rz" event={"ID":"074017a4-5c1e-4dbb-b74c-ab60c1dd53b7","Type":"ContainerStarted","Data":"6fee58c12fb8b9f85909781a71d492520982d3e7b2a64547bb2f4897e49e6a31"} Nov 26 02:18:50 crc kubenswrapper[4766]: I1126 02:18:50.336808 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 02:18:52 crc kubenswrapper[4766]: I1126 02:18:52.360122 4766 generic.go:334] "Generic (PLEG): container finished" podID="074017a4-5c1e-4dbb-b74c-ab60c1dd53b7" containerID="fa7adb003f89e075f7440e274cb60399f5804d489b6eb603e329635509842ec0" exitCode=0 Nov 26 02:18:52 crc kubenswrapper[4766]: I1126 02:18:52.360183 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p68rz" event={"ID":"074017a4-5c1e-4dbb-b74c-ab60c1dd53b7","Type":"ContainerDied","Data":"fa7adb003f89e075f7440e274cb60399f5804d489b6eb603e329635509842ec0"} Nov 26 02:18:53 crc kubenswrapper[4766]: I1126 02:18:53.376140 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p68rz" event={"ID":"074017a4-5c1e-4dbb-b74c-ab60c1dd53b7","Type":"ContainerStarted","Data":"a4b9e5a7b14cda603580cb252d9656888507bccbbef71ba77f104468e3d2258e"} Nov 26 02:18:53 crc kubenswrapper[4766]: I1126 02:18:53.401862 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p68rz" podStartSLOduration=1.9244751610000002 podStartE2EDuration="4.401842626s" podCreationTimestamp="2025-11-26 02:18:49 +0000 UTC" firstStartedPulling="2025-11-26 02:18:50.336258683 +0000 UTC m=+6911.185029153" lastFinishedPulling="2025-11-26 02:18:52.813626188 +0000 UTC m=+6913.662396618" observedRunningTime="2025-11-26 02:18:53.392617232 +0000 UTC m=+6914.241387672" watchObservedRunningTime="2025-11-26 02:18:53.401842626 +0000 UTC m=+6914.250613056" Nov 26 02:18:59 crc kubenswrapper[4766]: I1126 02:18:59.361554 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p68rz" Nov 26 02:18:59 crc kubenswrapper[4766]: I1126 02:18:59.362310 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p68rz" Nov 26 02:18:59 crc kubenswrapper[4766]: I1126 02:18:59.434739 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p68rz" Nov 26 02:18:59 crc kubenswrapper[4766]: I1126 02:18:59.517385 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p68rz" Nov 26 02:18:59 crc kubenswrapper[4766]: I1126 02:18:59.683605 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p68rz"] Nov 26 02:19:01 crc kubenswrapper[4766]: I1126 02:19:01.466409 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-p68rz" podUID="074017a4-5c1e-4dbb-b74c-ab60c1dd53b7" containerName="registry-server" containerID="cri-o://a4b9e5a7b14cda603580cb252d9656888507bccbbef71ba77f104468e3d2258e" gracePeriod=2 Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.126134 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p68rz" Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.211301 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7phm\" (UniqueName: \"kubernetes.io/projected/074017a4-5c1e-4dbb-b74c-ab60c1dd53b7-kube-api-access-r7phm\") pod \"074017a4-5c1e-4dbb-b74c-ab60c1dd53b7\" (UID: \"074017a4-5c1e-4dbb-b74c-ab60c1dd53b7\") " Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.211541 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/074017a4-5c1e-4dbb-b74c-ab60c1dd53b7-catalog-content\") pod \"074017a4-5c1e-4dbb-b74c-ab60c1dd53b7\" (UID: \"074017a4-5c1e-4dbb-b74c-ab60c1dd53b7\") " Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.211622 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/074017a4-5c1e-4dbb-b74c-ab60c1dd53b7-utilities\") pod \"074017a4-5c1e-4dbb-b74c-ab60c1dd53b7\" (UID: \"074017a4-5c1e-4dbb-b74c-ab60c1dd53b7\") " Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.212374 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/074017a4-5c1e-4dbb-b74c-ab60c1dd53b7-utilities" (OuterVolumeSpecName: "utilities") pod "074017a4-5c1e-4dbb-b74c-ab60c1dd53b7" (UID: "074017a4-5c1e-4dbb-b74c-ab60c1dd53b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.217787 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/074017a4-5c1e-4dbb-b74c-ab60c1dd53b7-kube-api-access-r7phm" (OuterVolumeSpecName: "kube-api-access-r7phm") pod "074017a4-5c1e-4dbb-b74c-ab60c1dd53b7" (UID: "074017a4-5c1e-4dbb-b74c-ab60c1dd53b7"). InnerVolumeSpecName "kube-api-access-r7phm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.232627 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/074017a4-5c1e-4dbb-b74c-ab60c1dd53b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "074017a4-5c1e-4dbb-b74c-ab60c1dd53b7" (UID: "074017a4-5c1e-4dbb-b74c-ab60c1dd53b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.314091 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/074017a4-5c1e-4dbb-b74c-ab60c1dd53b7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.314120 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/074017a4-5c1e-4dbb-b74c-ab60c1dd53b7-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.314131 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7phm\" (UniqueName: \"kubernetes.io/projected/074017a4-5c1e-4dbb-b74c-ab60c1dd53b7-kube-api-access-r7phm\") on node \"crc\" DevicePath \"\"" Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.483386 4766 generic.go:334] "Generic (PLEG): container finished" podID="074017a4-5c1e-4dbb-b74c-ab60c1dd53b7" containerID="a4b9e5a7b14cda603580cb252d9656888507bccbbef71ba77f104468e3d2258e" exitCode=0 Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.483432 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p68rz" event={"ID":"074017a4-5c1e-4dbb-b74c-ab60c1dd53b7","Type":"ContainerDied","Data":"a4b9e5a7b14cda603580cb252d9656888507bccbbef71ba77f104468e3d2258e"} Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.483465 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p68rz" event={"ID":"074017a4-5c1e-4dbb-b74c-ab60c1dd53b7","Type":"ContainerDied","Data":"6fee58c12fb8b9f85909781a71d492520982d3e7b2a64547bb2f4897e49e6a31"} Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.483481 4766 scope.go:117] "RemoveContainer" containerID="a4b9e5a7b14cda603580cb252d9656888507bccbbef71ba77f104468e3d2258e" Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.483476 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p68rz" Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.547520 4766 scope.go:117] "RemoveContainer" containerID="fa7adb003f89e075f7440e274cb60399f5804d489b6eb603e329635509842ec0" Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.547770 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p68rz"] Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.566109 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-p68rz"] Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.590475 4766 scope.go:117] "RemoveContainer" containerID="d0855c1f2774ccd5e1b07d9ae9de53a2b593160952ae65ecf23b7be6784fbc5f" Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.649023 4766 scope.go:117] "RemoveContainer" containerID="a4b9e5a7b14cda603580cb252d9656888507bccbbef71ba77f104468e3d2258e" Nov 26 02:19:02 crc kubenswrapper[4766]: E1126 02:19:02.649580 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4b9e5a7b14cda603580cb252d9656888507bccbbef71ba77f104468e3d2258e\": container with ID starting with a4b9e5a7b14cda603580cb252d9656888507bccbbef71ba77f104468e3d2258e not found: ID does not exist" containerID="a4b9e5a7b14cda603580cb252d9656888507bccbbef71ba77f104468e3d2258e" Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.649636 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4b9e5a7b14cda603580cb252d9656888507bccbbef71ba77f104468e3d2258e"} err="failed to get container status \"a4b9e5a7b14cda603580cb252d9656888507bccbbef71ba77f104468e3d2258e\": rpc error: code = NotFound desc = could not find container \"a4b9e5a7b14cda603580cb252d9656888507bccbbef71ba77f104468e3d2258e\": container with ID starting with a4b9e5a7b14cda603580cb252d9656888507bccbbef71ba77f104468e3d2258e not found: ID does not exist" Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.649751 4766 scope.go:117] "RemoveContainer" containerID="fa7adb003f89e075f7440e274cb60399f5804d489b6eb603e329635509842ec0" Nov 26 02:19:02 crc kubenswrapper[4766]: E1126 02:19:02.650028 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa7adb003f89e075f7440e274cb60399f5804d489b6eb603e329635509842ec0\": container with ID starting with fa7adb003f89e075f7440e274cb60399f5804d489b6eb603e329635509842ec0 not found: ID does not exist" containerID="fa7adb003f89e075f7440e274cb60399f5804d489b6eb603e329635509842ec0" Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.650054 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa7adb003f89e075f7440e274cb60399f5804d489b6eb603e329635509842ec0"} err="failed to get container status \"fa7adb003f89e075f7440e274cb60399f5804d489b6eb603e329635509842ec0\": rpc error: code = NotFound desc = could not find container \"fa7adb003f89e075f7440e274cb60399f5804d489b6eb603e329635509842ec0\": container with ID starting with fa7adb003f89e075f7440e274cb60399f5804d489b6eb603e329635509842ec0 not found: ID does not exist" Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.650066 4766 scope.go:117] "RemoveContainer" containerID="d0855c1f2774ccd5e1b07d9ae9de53a2b593160952ae65ecf23b7be6784fbc5f" Nov 26 02:19:02 crc kubenswrapper[4766]: E1126 02:19:02.650647 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0855c1f2774ccd5e1b07d9ae9de53a2b593160952ae65ecf23b7be6784fbc5f\": container with ID starting with d0855c1f2774ccd5e1b07d9ae9de53a2b593160952ae65ecf23b7be6784fbc5f not found: ID does not exist" containerID="d0855c1f2774ccd5e1b07d9ae9de53a2b593160952ae65ecf23b7be6784fbc5f" Nov 26 02:19:02 crc kubenswrapper[4766]: I1126 02:19:02.650708 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0855c1f2774ccd5e1b07d9ae9de53a2b593160952ae65ecf23b7be6784fbc5f"} err="failed to get container status \"d0855c1f2774ccd5e1b07d9ae9de53a2b593160952ae65ecf23b7be6784fbc5f\": rpc error: code = NotFound desc = could not find container \"d0855c1f2774ccd5e1b07d9ae9de53a2b593160952ae65ecf23b7be6784fbc5f\": container with ID starting with d0855c1f2774ccd5e1b07d9ae9de53a2b593160952ae65ecf23b7be6784fbc5f not found: ID does not exist" Nov 26 02:19:03 crc kubenswrapper[4766]: I1126 02:19:03.845199 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="074017a4-5c1e-4dbb-b74c-ab60c1dd53b7" path="/var/lib/kubelet/pods/074017a4-5c1e-4dbb-b74c-ab60c1dd53b7/volumes" Nov 26 02:20:41 crc kubenswrapper[4766]: I1126 02:20:41.479917 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:20:41 crc kubenswrapper[4766]: I1126 02:20:41.480693 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:21:11 crc kubenswrapper[4766]: I1126 02:21:11.479992 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:21:11 crc kubenswrapper[4766]: I1126 02:21:11.480734 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:21:34 crc kubenswrapper[4766]: I1126 02:21:34.994890 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cgf98"] Nov 26 02:21:34 crc kubenswrapper[4766]: E1126 02:21:34.995987 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="074017a4-5c1e-4dbb-b74c-ab60c1dd53b7" containerName="extract-utilities" Nov 26 02:21:34 crc kubenswrapper[4766]: I1126 02:21:34.996002 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="074017a4-5c1e-4dbb-b74c-ab60c1dd53b7" containerName="extract-utilities" Nov 26 02:21:34 crc kubenswrapper[4766]: E1126 02:21:34.996072 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="074017a4-5c1e-4dbb-b74c-ab60c1dd53b7" containerName="extract-content" Nov 26 02:21:34 crc kubenswrapper[4766]: I1126 02:21:34.996081 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="074017a4-5c1e-4dbb-b74c-ab60c1dd53b7" containerName="extract-content" Nov 26 02:21:34 crc kubenswrapper[4766]: E1126 02:21:34.996123 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="074017a4-5c1e-4dbb-b74c-ab60c1dd53b7" containerName="registry-server" Nov 26 02:21:34 crc kubenswrapper[4766]: I1126 02:21:34.996132 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="074017a4-5c1e-4dbb-b74c-ab60c1dd53b7" containerName="registry-server" Nov 26 02:21:34 crc kubenswrapper[4766]: I1126 02:21:34.996458 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="074017a4-5c1e-4dbb-b74c-ab60c1dd53b7" containerName="registry-server" Nov 26 02:21:34 crc kubenswrapper[4766]: I1126 02:21:34.998848 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cgf98" Nov 26 02:21:35 crc kubenswrapper[4766]: I1126 02:21:35.033345 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cgf98"] Nov 26 02:21:35 crc kubenswrapper[4766]: I1126 02:21:35.047394 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c406a7e-97b7-49d2-a9f1-527007552ee7-catalog-content\") pod \"redhat-operators-cgf98\" (UID: \"2c406a7e-97b7-49d2-a9f1-527007552ee7\") " pod="openshift-marketplace/redhat-operators-cgf98" Nov 26 02:21:35 crc kubenswrapper[4766]: I1126 02:21:35.047619 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c406a7e-97b7-49d2-a9f1-527007552ee7-utilities\") pod \"redhat-operators-cgf98\" (UID: \"2c406a7e-97b7-49d2-a9f1-527007552ee7\") " pod="openshift-marketplace/redhat-operators-cgf98" Nov 26 02:21:35 crc kubenswrapper[4766]: I1126 02:21:35.047725 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqpgp\" (UniqueName: \"kubernetes.io/projected/2c406a7e-97b7-49d2-a9f1-527007552ee7-kube-api-access-lqpgp\") pod \"redhat-operators-cgf98\" (UID: \"2c406a7e-97b7-49d2-a9f1-527007552ee7\") " pod="openshift-marketplace/redhat-operators-cgf98" Nov 26 02:21:35 crc kubenswrapper[4766]: I1126 02:21:35.150415 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c406a7e-97b7-49d2-a9f1-527007552ee7-utilities\") pod \"redhat-operators-cgf98\" (UID: \"2c406a7e-97b7-49d2-a9f1-527007552ee7\") " pod="openshift-marketplace/redhat-operators-cgf98" Nov 26 02:21:35 crc kubenswrapper[4766]: I1126 02:21:35.150795 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqpgp\" (UniqueName: \"kubernetes.io/projected/2c406a7e-97b7-49d2-a9f1-527007552ee7-kube-api-access-lqpgp\") pod \"redhat-operators-cgf98\" (UID: \"2c406a7e-97b7-49d2-a9f1-527007552ee7\") " pod="openshift-marketplace/redhat-operators-cgf98" Nov 26 02:21:35 crc kubenswrapper[4766]: I1126 02:21:35.150838 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c406a7e-97b7-49d2-a9f1-527007552ee7-catalog-content\") pod \"redhat-operators-cgf98\" (UID: \"2c406a7e-97b7-49d2-a9f1-527007552ee7\") " pod="openshift-marketplace/redhat-operators-cgf98" Nov 26 02:21:35 crc kubenswrapper[4766]: I1126 02:21:35.151182 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c406a7e-97b7-49d2-a9f1-527007552ee7-utilities\") pod \"redhat-operators-cgf98\" (UID: \"2c406a7e-97b7-49d2-a9f1-527007552ee7\") " pod="openshift-marketplace/redhat-operators-cgf98" Nov 26 02:21:35 crc kubenswrapper[4766]: I1126 02:21:35.151482 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c406a7e-97b7-49d2-a9f1-527007552ee7-catalog-content\") pod \"redhat-operators-cgf98\" (UID: \"2c406a7e-97b7-49d2-a9f1-527007552ee7\") " pod="openshift-marketplace/redhat-operators-cgf98" Nov 26 02:21:35 crc kubenswrapper[4766]: I1126 02:21:35.170345 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqpgp\" (UniqueName: \"kubernetes.io/projected/2c406a7e-97b7-49d2-a9f1-527007552ee7-kube-api-access-lqpgp\") pod \"redhat-operators-cgf98\" (UID: \"2c406a7e-97b7-49d2-a9f1-527007552ee7\") " pod="openshift-marketplace/redhat-operators-cgf98" Nov 26 02:21:35 crc kubenswrapper[4766]: I1126 02:21:35.324595 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cgf98" Nov 26 02:21:35 crc kubenswrapper[4766]: I1126 02:21:35.825421 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cgf98"] Nov 26 02:21:35 crc kubenswrapper[4766]: I1126 02:21:35.911439 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cgf98" event={"ID":"2c406a7e-97b7-49d2-a9f1-527007552ee7","Type":"ContainerStarted","Data":"9f754789652cb4e2c655c8f54f05888bcdfb402875657b41c743b6a0b8d87ba0"} Nov 26 02:21:36 crc kubenswrapper[4766]: I1126 02:21:36.921470 4766 generic.go:334] "Generic (PLEG): container finished" podID="2c406a7e-97b7-49d2-a9f1-527007552ee7" containerID="ffc8b1b334e9e58b7f6e9a565e61d2becbad3618d7a7673a627ee28ca49524bb" exitCode=0 Nov 26 02:21:36 crc kubenswrapper[4766]: I1126 02:21:36.921557 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cgf98" event={"ID":"2c406a7e-97b7-49d2-a9f1-527007552ee7","Type":"ContainerDied","Data":"ffc8b1b334e9e58b7f6e9a565e61d2becbad3618d7a7673a627ee28ca49524bb"} Nov 26 02:21:37 crc kubenswrapper[4766]: I1126 02:21:37.936092 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cgf98" event={"ID":"2c406a7e-97b7-49d2-a9f1-527007552ee7","Type":"ContainerStarted","Data":"3dbb86447a76e557287ec362e48edba7a154fbe8cc705f7c0df5cb53cf76d614"} Nov 26 02:21:41 crc kubenswrapper[4766]: I1126 02:21:41.479534 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:21:41 crc kubenswrapper[4766]: I1126 02:21:41.480208 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:21:41 crc kubenswrapper[4766]: I1126 02:21:41.480288 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 02:21:41 crc kubenswrapper[4766]: I1126 02:21:41.481584 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 02:21:41 crc kubenswrapper[4766]: I1126 02:21:41.481771 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" gracePeriod=600 Nov 26 02:21:42 crc kubenswrapper[4766]: I1126 02:21:42.005532 4766 generic.go:334] "Generic (PLEG): container finished" podID="2c406a7e-97b7-49d2-a9f1-527007552ee7" containerID="3dbb86447a76e557287ec362e48edba7a154fbe8cc705f7c0df5cb53cf76d614" exitCode=0 Nov 26 02:21:42 crc kubenswrapper[4766]: I1126 02:21:42.005604 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cgf98" event={"ID":"2c406a7e-97b7-49d2-a9f1-527007552ee7","Type":"ContainerDied","Data":"3dbb86447a76e557287ec362e48edba7a154fbe8cc705f7c0df5cb53cf76d614"} Nov 26 02:21:42 crc kubenswrapper[4766]: I1126 02:21:42.012880 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" exitCode=0 Nov 26 02:21:42 crc kubenswrapper[4766]: I1126 02:21:42.012926 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562"} Nov 26 02:21:42 crc kubenswrapper[4766]: I1126 02:21:42.012974 4766 scope.go:117] "RemoveContainer" containerID="f3b79bb97115423a9ddb10f5e7bc376bff4f35b58a4fe41d6b2ad3dfaa0c3a8c" Nov 26 02:21:42 crc kubenswrapper[4766]: E1126 02:21:42.143825 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:21:43 crc kubenswrapper[4766]: I1126 02:21:43.027856 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:21:43 crc kubenswrapper[4766]: E1126 02:21:43.028693 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:21:43 crc kubenswrapper[4766]: I1126 02:21:43.030233 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cgf98" event={"ID":"2c406a7e-97b7-49d2-a9f1-527007552ee7","Type":"ContainerStarted","Data":"c79b623e8093f6e9bd56c322f35c219c83fd03572ab408b1ff283f57ebe25625"} Nov 26 02:21:43 crc kubenswrapper[4766]: I1126 02:21:43.110164 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cgf98" podStartSLOduration=3.618727312 podStartE2EDuration="9.11014365s" podCreationTimestamp="2025-11-26 02:21:34 +0000 UTC" firstStartedPulling="2025-11-26 02:21:36.923826284 +0000 UTC m=+7077.772596714" lastFinishedPulling="2025-11-26 02:21:42.415242622 +0000 UTC m=+7083.264013052" observedRunningTime="2025-11-26 02:21:43.079910016 +0000 UTC m=+7083.928680456" watchObservedRunningTime="2025-11-26 02:21:43.11014365 +0000 UTC m=+7083.958914090" Nov 26 02:21:45 crc kubenswrapper[4766]: I1126 02:21:45.325583 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cgf98" Nov 26 02:21:45 crc kubenswrapper[4766]: I1126 02:21:45.326035 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cgf98" Nov 26 02:21:46 crc kubenswrapper[4766]: I1126 02:21:46.394718 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cgf98" podUID="2c406a7e-97b7-49d2-a9f1-527007552ee7" containerName="registry-server" probeResult="failure" output=< Nov 26 02:21:46 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 02:21:46 crc kubenswrapper[4766]: > Nov 26 02:21:55 crc kubenswrapper[4766]: I1126 02:21:55.404566 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cgf98" Nov 26 02:21:55 crc kubenswrapper[4766]: I1126 02:21:55.485048 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cgf98" Nov 26 02:21:55 crc kubenswrapper[4766]: I1126 02:21:55.664937 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cgf98"] Nov 26 02:21:56 crc kubenswrapper[4766]: I1126 02:21:56.827523 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:21:56 crc kubenswrapper[4766]: E1126 02:21:56.829314 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:21:57 crc kubenswrapper[4766]: I1126 02:21:57.196216 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cgf98" podUID="2c406a7e-97b7-49d2-a9f1-527007552ee7" containerName="registry-server" containerID="cri-o://c79b623e8093f6e9bd56c322f35c219c83fd03572ab408b1ff283f57ebe25625" gracePeriod=2 Nov 26 02:21:57 crc kubenswrapper[4766]: I1126 02:21:57.798943 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cgf98" Nov 26 02:21:57 crc kubenswrapper[4766]: I1126 02:21:57.904956 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c406a7e-97b7-49d2-a9f1-527007552ee7-utilities\") pod \"2c406a7e-97b7-49d2-a9f1-527007552ee7\" (UID: \"2c406a7e-97b7-49d2-a9f1-527007552ee7\") " Nov 26 02:21:57 crc kubenswrapper[4766]: I1126 02:21:57.905180 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c406a7e-97b7-49d2-a9f1-527007552ee7-catalog-content\") pod \"2c406a7e-97b7-49d2-a9f1-527007552ee7\" (UID: \"2c406a7e-97b7-49d2-a9f1-527007552ee7\") " Nov 26 02:21:57 crc kubenswrapper[4766]: I1126 02:21:57.905338 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqpgp\" (UniqueName: \"kubernetes.io/projected/2c406a7e-97b7-49d2-a9f1-527007552ee7-kube-api-access-lqpgp\") pod \"2c406a7e-97b7-49d2-a9f1-527007552ee7\" (UID: \"2c406a7e-97b7-49d2-a9f1-527007552ee7\") " Nov 26 02:21:57 crc kubenswrapper[4766]: I1126 02:21:57.905943 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c406a7e-97b7-49d2-a9f1-527007552ee7-utilities" (OuterVolumeSpecName: "utilities") pod "2c406a7e-97b7-49d2-a9f1-527007552ee7" (UID: "2c406a7e-97b7-49d2-a9f1-527007552ee7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:21:57 crc kubenswrapper[4766]: I1126 02:21:57.914845 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c406a7e-97b7-49d2-a9f1-527007552ee7-kube-api-access-lqpgp" (OuterVolumeSpecName: "kube-api-access-lqpgp") pod "2c406a7e-97b7-49d2-a9f1-527007552ee7" (UID: "2c406a7e-97b7-49d2-a9f1-527007552ee7"). InnerVolumeSpecName "kube-api-access-lqpgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:21:58 crc kubenswrapper[4766]: I1126 02:21:58.008191 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqpgp\" (UniqueName: \"kubernetes.io/projected/2c406a7e-97b7-49d2-a9f1-527007552ee7-kube-api-access-lqpgp\") on node \"crc\" DevicePath \"\"" Nov 26 02:21:58 crc kubenswrapper[4766]: I1126 02:21:58.008466 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c406a7e-97b7-49d2-a9f1-527007552ee7-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 02:21:58 crc kubenswrapper[4766]: I1126 02:21:58.010881 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c406a7e-97b7-49d2-a9f1-527007552ee7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2c406a7e-97b7-49d2-a9f1-527007552ee7" (UID: "2c406a7e-97b7-49d2-a9f1-527007552ee7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:21:58 crc kubenswrapper[4766]: I1126 02:21:58.110867 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c406a7e-97b7-49d2-a9f1-527007552ee7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 02:21:58 crc kubenswrapper[4766]: I1126 02:21:58.234886 4766 generic.go:334] "Generic (PLEG): container finished" podID="2c406a7e-97b7-49d2-a9f1-527007552ee7" containerID="c79b623e8093f6e9bd56c322f35c219c83fd03572ab408b1ff283f57ebe25625" exitCode=0 Nov 26 02:21:58 crc kubenswrapper[4766]: I1126 02:21:58.234943 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cgf98" event={"ID":"2c406a7e-97b7-49d2-a9f1-527007552ee7","Type":"ContainerDied","Data":"c79b623e8093f6e9bd56c322f35c219c83fd03572ab408b1ff283f57ebe25625"} Nov 26 02:21:58 crc kubenswrapper[4766]: I1126 02:21:58.234968 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cgf98" event={"ID":"2c406a7e-97b7-49d2-a9f1-527007552ee7","Type":"ContainerDied","Data":"9f754789652cb4e2c655c8f54f05888bcdfb402875657b41c743b6a0b8d87ba0"} Nov 26 02:21:58 crc kubenswrapper[4766]: I1126 02:21:58.235050 4766 scope.go:117] "RemoveContainer" containerID="c79b623e8093f6e9bd56c322f35c219c83fd03572ab408b1ff283f57ebe25625" Nov 26 02:21:58 crc kubenswrapper[4766]: I1126 02:21:58.235175 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cgf98" Nov 26 02:21:58 crc kubenswrapper[4766]: I1126 02:21:58.268388 4766 scope.go:117] "RemoveContainer" containerID="3dbb86447a76e557287ec362e48edba7a154fbe8cc705f7c0df5cb53cf76d614" Nov 26 02:21:58 crc kubenswrapper[4766]: I1126 02:21:58.286381 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cgf98"] Nov 26 02:21:58 crc kubenswrapper[4766]: I1126 02:21:58.303525 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cgf98"] Nov 26 02:21:58 crc kubenswrapper[4766]: I1126 02:21:58.307761 4766 scope.go:117] "RemoveContainer" containerID="ffc8b1b334e9e58b7f6e9a565e61d2becbad3618d7a7673a627ee28ca49524bb" Nov 26 02:21:58 crc kubenswrapper[4766]: I1126 02:21:58.350719 4766 scope.go:117] "RemoveContainer" containerID="c79b623e8093f6e9bd56c322f35c219c83fd03572ab408b1ff283f57ebe25625" Nov 26 02:21:58 crc kubenswrapper[4766]: E1126 02:21:58.351220 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c79b623e8093f6e9bd56c322f35c219c83fd03572ab408b1ff283f57ebe25625\": container with ID starting with c79b623e8093f6e9bd56c322f35c219c83fd03572ab408b1ff283f57ebe25625 not found: ID does not exist" containerID="c79b623e8093f6e9bd56c322f35c219c83fd03572ab408b1ff283f57ebe25625" Nov 26 02:21:58 crc kubenswrapper[4766]: I1126 02:21:58.351270 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c79b623e8093f6e9bd56c322f35c219c83fd03572ab408b1ff283f57ebe25625"} err="failed to get container status \"c79b623e8093f6e9bd56c322f35c219c83fd03572ab408b1ff283f57ebe25625\": rpc error: code = NotFound desc = could not find container \"c79b623e8093f6e9bd56c322f35c219c83fd03572ab408b1ff283f57ebe25625\": container with ID starting with c79b623e8093f6e9bd56c322f35c219c83fd03572ab408b1ff283f57ebe25625 not found: ID does not exist" Nov 26 02:21:58 crc kubenswrapper[4766]: I1126 02:21:58.351296 4766 scope.go:117] "RemoveContainer" containerID="3dbb86447a76e557287ec362e48edba7a154fbe8cc705f7c0df5cb53cf76d614" Nov 26 02:21:58 crc kubenswrapper[4766]: E1126 02:21:58.351753 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dbb86447a76e557287ec362e48edba7a154fbe8cc705f7c0df5cb53cf76d614\": container with ID starting with 3dbb86447a76e557287ec362e48edba7a154fbe8cc705f7c0df5cb53cf76d614 not found: ID does not exist" containerID="3dbb86447a76e557287ec362e48edba7a154fbe8cc705f7c0df5cb53cf76d614" Nov 26 02:21:58 crc kubenswrapper[4766]: I1126 02:21:58.351793 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dbb86447a76e557287ec362e48edba7a154fbe8cc705f7c0df5cb53cf76d614"} err="failed to get container status \"3dbb86447a76e557287ec362e48edba7a154fbe8cc705f7c0df5cb53cf76d614\": rpc error: code = NotFound desc = could not find container \"3dbb86447a76e557287ec362e48edba7a154fbe8cc705f7c0df5cb53cf76d614\": container with ID starting with 3dbb86447a76e557287ec362e48edba7a154fbe8cc705f7c0df5cb53cf76d614 not found: ID does not exist" Nov 26 02:21:58 crc kubenswrapper[4766]: I1126 02:21:58.351819 4766 scope.go:117] "RemoveContainer" containerID="ffc8b1b334e9e58b7f6e9a565e61d2becbad3618d7a7673a627ee28ca49524bb" Nov 26 02:21:58 crc kubenswrapper[4766]: E1126 02:21:58.352254 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffc8b1b334e9e58b7f6e9a565e61d2becbad3618d7a7673a627ee28ca49524bb\": container with ID starting with ffc8b1b334e9e58b7f6e9a565e61d2becbad3618d7a7673a627ee28ca49524bb not found: ID does not exist" containerID="ffc8b1b334e9e58b7f6e9a565e61d2becbad3618d7a7673a627ee28ca49524bb" Nov 26 02:21:58 crc kubenswrapper[4766]: I1126 02:21:58.352279 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffc8b1b334e9e58b7f6e9a565e61d2becbad3618d7a7673a627ee28ca49524bb"} err="failed to get container status \"ffc8b1b334e9e58b7f6e9a565e61d2becbad3618d7a7673a627ee28ca49524bb\": rpc error: code = NotFound desc = could not find container \"ffc8b1b334e9e58b7f6e9a565e61d2becbad3618d7a7673a627ee28ca49524bb\": container with ID starting with ffc8b1b334e9e58b7f6e9a565e61d2becbad3618d7a7673a627ee28ca49524bb not found: ID does not exist" Nov 26 02:21:59 crc kubenswrapper[4766]: I1126 02:21:59.845603 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c406a7e-97b7-49d2-a9f1-527007552ee7" path="/var/lib/kubelet/pods/2c406a7e-97b7-49d2-a9f1-527007552ee7/volumes" Nov 26 02:22:11 crc kubenswrapper[4766]: I1126 02:22:11.827155 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:22:11 crc kubenswrapper[4766]: E1126 02:22:11.828315 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:22:26 crc kubenswrapper[4766]: I1126 02:22:26.830782 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:22:26 crc kubenswrapper[4766]: E1126 02:22:26.833682 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:22:38 crc kubenswrapper[4766]: I1126 02:22:38.827575 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:22:38 crc kubenswrapper[4766]: E1126 02:22:38.829001 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:22:49 crc kubenswrapper[4766]: I1126 02:22:49.847338 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:22:49 crc kubenswrapper[4766]: E1126 02:22:49.850590 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:23:00 crc kubenswrapper[4766]: I1126 02:23:00.827115 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:23:00 crc kubenswrapper[4766]: E1126 02:23:00.829965 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:23:15 crc kubenswrapper[4766]: I1126 02:23:15.828439 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:23:15 crc kubenswrapper[4766]: E1126 02:23:15.829575 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:23:26 crc kubenswrapper[4766]: I1126 02:23:26.827058 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:23:26 crc kubenswrapper[4766]: E1126 02:23:26.828357 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:23:38 crc kubenswrapper[4766]: I1126 02:23:38.826965 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:23:38 crc kubenswrapper[4766]: E1126 02:23:38.827888 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:23:52 crc kubenswrapper[4766]: I1126 02:23:52.827320 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:23:52 crc kubenswrapper[4766]: E1126 02:23:52.833246 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:24:03 crc kubenswrapper[4766]: I1126 02:24:03.580351 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k27p4"] Nov 26 02:24:03 crc kubenswrapper[4766]: E1126 02:24:03.584564 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c406a7e-97b7-49d2-a9f1-527007552ee7" containerName="registry-server" Nov 26 02:24:03 crc kubenswrapper[4766]: I1126 02:24:03.584615 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c406a7e-97b7-49d2-a9f1-527007552ee7" containerName="registry-server" Nov 26 02:24:03 crc kubenswrapper[4766]: E1126 02:24:03.584729 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c406a7e-97b7-49d2-a9f1-527007552ee7" containerName="extract-utilities" Nov 26 02:24:03 crc kubenswrapper[4766]: I1126 02:24:03.584744 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c406a7e-97b7-49d2-a9f1-527007552ee7" containerName="extract-utilities" Nov 26 02:24:03 crc kubenswrapper[4766]: E1126 02:24:03.584813 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c406a7e-97b7-49d2-a9f1-527007552ee7" containerName="extract-content" Nov 26 02:24:03 crc kubenswrapper[4766]: I1126 02:24:03.584825 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c406a7e-97b7-49d2-a9f1-527007552ee7" containerName="extract-content" Nov 26 02:24:03 crc kubenswrapper[4766]: I1126 02:24:03.585237 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c406a7e-97b7-49d2-a9f1-527007552ee7" containerName="registry-server" Nov 26 02:24:03 crc kubenswrapper[4766]: I1126 02:24:03.588229 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k27p4" Nov 26 02:24:03 crc kubenswrapper[4766]: I1126 02:24:03.601085 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k27p4"] Nov 26 02:24:03 crc kubenswrapper[4766]: I1126 02:24:03.716251 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad06a4c9-3c79-4a6f-beea-65ab4b974047-catalog-content\") pod \"certified-operators-k27p4\" (UID: \"ad06a4c9-3c79-4a6f-beea-65ab4b974047\") " pod="openshift-marketplace/certified-operators-k27p4" Nov 26 02:24:03 crc kubenswrapper[4766]: I1126 02:24:03.716632 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad06a4c9-3c79-4a6f-beea-65ab4b974047-utilities\") pod \"certified-operators-k27p4\" (UID: \"ad06a4c9-3c79-4a6f-beea-65ab4b974047\") " pod="openshift-marketplace/certified-operators-k27p4" Nov 26 02:24:03 crc kubenswrapper[4766]: I1126 02:24:03.716881 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ltm8\" (UniqueName: \"kubernetes.io/projected/ad06a4c9-3c79-4a6f-beea-65ab4b974047-kube-api-access-8ltm8\") pod \"certified-operators-k27p4\" (UID: \"ad06a4c9-3c79-4a6f-beea-65ab4b974047\") " pod="openshift-marketplace/certified-operators-k27p4" Nov 26 02:24:03 crc kubenswrapper[4766]: I1126 02:24:03.819093 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad06a4c9-3c79-4a6f-beea-65ab4b974047-catalog-content\") pod \"certified-operators-k27p4\" (UID: \"ad06a4c9-3c79-4a6f-beea-65ab4b974047\") " pod="openshift-marketplace/certified-operators-k27p4" Nov 26 02:24:03 crc kubenswrapper[4766]: I1126 02:24:03.819541 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad06a4c9-3c79-4a6f-beea-65ab4b974047-utilities\") pod \"certified-operators-k27p4\" (UID: \"ad06a4c9-3c79-4a6f-beea-65ab4b974047\") " pod="openshift-marketplace/certified-operators-k27p4" Nov 26 02:24:03 crc kubenswrapper[4766]: I1126 02:24:03.819854 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad06a4c9-3c79-4a6f-beea-65ab4b974047-utilities\") pod \"certified-operators-k27p4\" (UID: \"ad06a4c9-3c79-4a6f-beea-65ab4b974047\") " pod="openshift-marketplace/certified-operators-k27p4" Nov 26 02:24:03 crc kubenswrapper[4766]: I1126 02:24:03.819858 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ltm8\" (UniqueName: \"kubernetes.io/projected/ad06a4c9-3c79-4a6f-beea-65ab4b974047-kube-api-access-8ltm8\") pod \"certified-operators-k27p4\" (UID: \"ad06a4c9-3c79-4a6f-beea-65ab4b974047\") " pod="openshift-marketplace/certified-operators-k27p4" Nov 26 02:24:03 crc kubenswrapper[4766]: I1126 02:24:03.820202 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad06a4c9-3c79-4a6f-beea-65ab4b974047-catalog-content\") pod \"certified-operators-k27p4\" (UID: \"ad06a4c9-3c79-4a6f-beea-65ab4b974047\") " pod="openshift-marketplace/certified-operators-k27p4" Nov 26 02:24:03 crc kubenswrapper[4766]: I1126 02:24:03.847580 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ltm8\" (UniqueName: \"kubernetes.io/projected/ad06a4c9-3c79-4a6f-beea-65ab4b974047-kube-api-access-8ltm8\") pod \"certified-operators-k27p4\" (UID: \"ad06a4c9-3c79-4a6f-beea-65ab4b974047\") " pod="openshift-marketplace/certified-operators-k27p4" Nov 26 02:24:03 crc kubenswrapper[4766]: I1126 02:24:03.926387 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k27p4" Nov 26 02:24:04 crc kubenswrapper[4766]: I1126 02:24:04.495446 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k27p4"] Nov 26 02:24:04 crc kubenswrapper[4766]: I1126 02:24:04.922705 4766 generic.go:334] "Generic (PLEG): container finished" podID="ad06a4c9-3c79-4a6f-beea-65ab4b974047" containerID="b303947420b5551892475bcbfeaf8431273653a987c377b93427a3a7bc024cc9" exitCode=0 Nov 26 02:24:04 crc kubenswrapper[4766]: I1126 02:24:04.922802 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k27p4" event={"ID":"ad06a4c9-3c79-4a6f-beea-65ab4b974047","Type":"ContainerDied","Data":"b303947420b5551892475bcbfeaf8431273653a987c377b93427a3a7bc024cc9"} Nov 26 02:24:04 crc kubenswrapper[4766]: I1126 02:24:04.922977 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k27p4" event={"ID":"ad06a4c9-3c79-4a6f-beea-65ab4b974047","Type":"ContainerStarted","Data":"8c02d1031b9ad07678144b4c38c3e1c4d8137696f47ee3943c97b8171f647dce"} Nov 26 02:24:04 crc kubenswrapper[4766]: I1126 02:24:04.924624 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 02:24:05 crc kubenswrapper[4766]: I1126 02:24:05.827172 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:24:05 crc kubenswrapper[4766]: E1126 02:24:05.827957 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:24:05 crc kubenswrapper[4766]: I1126 02:24:05.938769 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k27p4" event={"ID":"ad06a4c9-3c79-4a6f-beea-65ab4b974047","Type":"ContainerStarted","Data":"ff923fa0e0cdfe5a3525bcbd06dfa4a1564743be563fec635417116814fbc917"} Nov 26 02:24:07 crc kubenswrapper[4766]: I1126 02:24:07.967126 4766 generic.go:334] "Generic (PLEG): container finished" podID="ad06a4c9-3c79-4a6f-beea-65ab4b974047" containerID="ff923fa0e0cdfe5a3525bcbd06dfa4a1564743be563fec635417116814fbc917" exitCode=0 Nov 26 02:24:07 crc kubenswrapper[4766]: I1126 02:24:07.967432 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k27p4" event={"ID":"ad06a4c9-3c79-4a6f-beea-65ab4b974047","Type":"ContainerDied","Data":"ff923fa0e0cdfe5a3525bcbd06dfa4a1564743be563fec635417116814fbc917"} Nov 26 02:24:08 crc kubenswrapper[4766]: I1126 02:24:08.980101 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k27p4" event={"ID":"ad06a4c9-3c79-4a6f-beea-65ab4b974047","Type":"ContainerStarted","Data":"2c59af8f92fcca5aac46787346a10e67664e629bf42e727d60ee84d68ffcb8d8"} Nov 26 02:24:08 crc kubenswrapper[4766]: I1126 02:24:08.996177 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k27p4" podStartSLOduration=2.445053259 podStartE2EDuration="5.996157977s" podCreationTimestamp="2025-11-26 02:24:03 +0000 UTC" firstStartedPulling="2025-11-26 02:24:04.92438363 +0000 UTC m=+7225.773154060" lastFinishedPulling="2025-11-26 02:24:08.475488338 +0000 UTC m=+7229.324258778" observedRunningTime="2025-11-26 02:24:08.99427568 +0000 UTC m=+7229.843046130" watchObservedRunningTime="2025-11-26 02:24:08.996157977 +0000 UTC m=+7229.844928407" Nov 26 02:24:13 crc kubenswrapper[4766]: I1126 02:24:13.926877 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k27p4" Nov 26 02:24:13 crc kubenswrapper[4766]: I1126 02:24:13.927452 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k27p4" Nov 26 02:24:13 crc kubenswrapper[4766]: I1126 02:24:13.981042 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k27p4" Nov 26 02:24:14 crc kubenswrapper[4766]: I1126 02:24:14.097503 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k27p4" Nov 26 02:24:14 crc kubenswrapper[4766]: I1126 02:24:14.224144 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k27p4"] Nov 26 02:24:16 crc kubenswrapper[4766]: I1126 02:24:16.060454 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k27p4" podUID="ad06a4c9-3c79-4a6f-beea-65ab4b974047" containerName="registry-server" containerID="cri-o://2c59af8f92fcca5aac46787346a10e67664e629bf42e727d60ee84d68ffcb8d8" gracePeriod=2 Nov 26 02:24:16 crc kubenswrapper[4766]: I1126 02:24:16.677608 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k27p4" Nov 26 02:24:16 crc kubenswrapper[4766]: I1126 02:24:16.828194 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad06a4c9-3c79-4a6f-beea-65ab4b974047-catalog-content\") pod \"ad06a4c9-3c79-4a6f-beea-65ab4b974047\" (UID: \"ad06a4c9-3c79-4a6f-beea-65ab4b974047\") " Nov 26 02:24:16 crc kubenswrapper[4766]: I1126 02:24:16.828317 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ltm8\" (UniqueName: \"kubernetes.io/projected/ad06a4c9-3c79-4a6f-beea-65ab4b974047-kube-api-access-8ltm8\") pod \"ad06a4c9-3c79-4a6f-beea-65ab4b974047\" (UID: \"ad06a4c9-3c79-4a6f-beea-65ab4b974047\") " Nov 26 02:24:16 crc kubenswrapper[4766]: I1126 02:24:16.828338 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad06a4c9-3c79-4a6f-beea-65ab4b974047-utilities\") pod \"ad06a4c9-3c79-4a6f-beea-65ab4b974047\" (UID: \"ad06a4c9-3c79-4a6f-beea-65ab4b974047\") " Nov 26 02:24:16 crc kubenswrapper[4766]: I1126 02:24:16.829345 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad06a4c9-3c79-4a6f-beea-65ab4b974047-utilities" (OuterVolumeSpecName: "utilities") pod "ad06a4c9-3c79-4a6f-beea-65ab4b974047" (UID: "ad06a4c9-3c79-4a6f-beea-65ab4b974047"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:24:16 crc kubenswrapper[4766]: I1126 02:24:16.833840 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad06a4c9-3c79-4a6f-beea-65ab4b974047-kube-api-access-8ltm8" (OuterVolumeSpecName: "kube-api-access-8ltm8") pod "ad06a4c9-3c79-4a6f-beea-65ab4b974047" (UID: "ad06a4c9-3c79-4a6f-beea-65ab4b974047"). InnerVolumeSpecName "kube-api-access-8ltm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:24:16 crc kubenswrapper[4766]: I1126 02:24:16.872399 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad06a4c9-3c79-4a6f-beea-65ab4b974047-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad06a4c9-3c79-4a6f-beea-65ab4b974047" (UID: "ad06a4c9-3c79-4a6f-beea-65ab4b974047"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:24:16 crc kubenswrapper[4766]: I1126 02:24:16.931242 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad06a4c9-3c79-4a6f-beea-65ab4b974047-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 02:24:16 crc kubenswrapper[4766]: I1126 02:24:16.931276 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ltm8\" (UniqueName: \"kubernetes.io/projected/ad06a4c9-3c79-4a6f-beea-65ab4b974047-kube-api-access-8ltm8\") on node \"crc\" DevicePath \"\"" Nov 26 02:24:16 crc kubenswrapper[4766]: I1126 02:24:16.931288 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad06a4c9-3c79-4a6f-beea-65ab4b974047-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 02:24:17 crc kubenswrapper[4766]: I1126 02:24:17.070321 4766 generic.go:334] "Generic (PLEG): container finished" podID="ad06a4c9-3c79-4a6f-beea-65ab4b974047" containerID="2c59af8f92fcca5aac46787346a10e67664e629bf42e727d60ee84d68ffcb8d8" exitCode=0 Nov 26 02:24:17 crc kubenswrapper[4766]: I1126 02:24:17.070391 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k27p4" event={"ID":"ad06a4c9-3c79-4a6f-beea-65ab4b974047","Type":"ContainerDied","Data":"2c59af8f92fcca5aac46787346a10e67664e629bf42e727d60ee84d68ffcb8d8"} Nov 26 02:24:17 crc kubenswrapper[4766]: I1126 02:24:17.070443 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k27p4" event={"ID":"ad06a4c9-3c79-4a6f-beea-65ab4b974047","Type":"ContainerDied","Data":"8c02d1031b9ad07678144b4c38c3e1c4d8137696f47ee3943c97b8171f647dce"} Nov 26 02:24:17 crc kubenswrapper[4766]: I1126 02:24:17.070472 4766 scope.go:117] "RemoveContainer" containerID="2c59af8f92fcca5aac46787346a10e67664e629bf42e727d60ee84d68ffcb8d8" Nov 26 02:24:17 crc kubenswrapper[4766]: I1126 02:24:17.070358 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k27p4" Nov 26 02:24:17 crc kubenswrapper[4766]: I1126 02:24:17.109780 4766 scope.go:117] "RemoveContainer" containerID="ff923fa0e0cdfe5a3525bcbd06dfa4a1564743be563fec635417116814fbc917" Nov 26 02:24:17 crc kubenswrapper[4766]: I1126 02:24:17.110778 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k27p4"] Nov 26 02:24:17 crc kubenswrapper[4766]: I1126 02:24:17.124634 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k27p4"] Nov 26 02:24:17 crc kubenswrapper[4766]: I1126 02:24:17.142898 4766 scope.go:117] "RemoveContainer" containerID="b303947420b5551892475bcbfeaf8431273653a987c377b93427a3a7bc024cc9" Nov 26 02:24:17 crc kubenswrapper[4766]: E1126 02:24:17.200714 4766 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad06a4c9_3c79_4a6f_beea_65ab4b974047.slice/crio-8c02d1031b9ad07678144b4c38c3e1c4d8137696f47ee3943c97b8171f647dce\": RecentStats: unable to find data in memory cache]" Nov 26 02:24:17 crc kubenswrapper[4766]: I1126 02:24:17.222360 4766 scope.go:117] "RemoveContainer" containerID="2c59af8f92fcca5aac46787346a10e67664e629bf42e727d60ee84d68ffcb8d8" Nov 26 02:24:17 crc kubenswrapper[4766]: E1126 02:24:17.222858 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c59af8f92fcca5aac46787346a10e67664e629bf42e727d60ee84d68ffcb8d8\": container with ID starting with 2c59af8f92fcca5aac46787346a10e67664e629bf42e727d60ee84d68ffcb8d8 not found: ID does not exist" containerID="2c59af8f92fcca5aac46787346a10e67664e629bf42e727d60ee84d68ffcb8d8" Nov 26 02:24:17 crc kubenswrapper[4766]: I1126 02:24:17.222958 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c59af8f92fcca5aac46787346a10e67664e629bf42e727d60ee84d68ffcb8d8"} err="failed to get container status \"2c59af8f92fcca5aac46787346a10e67664e629bf42e727d60ee84d68ffcb8d8\": rpc error: code = NotFound desc = could not find container \"2c59af8f92fcca5aac46787346a10e67664e629bf42e727d60ee84d68ffcb8d8\": container with ID starting with 2c59af8f92fcca5aac46787346a10e67664e629bf42e727d60ee84d68ffcb8d8 not found: ID does not exist" Nov 26 02:24:17 crc kubenswrapper[4766]: I1126 02:24:17.223039 4766 scope.go:117] "RemoveContainer" containerID="ff923fa0e0cdfe5a3525bcbd06dfa4a1564743be563fec635417116814fbc917" Nov 26 02:24:17 crc kubenswrapper[4766]: E1126 02:24:17.223554 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff923fa0e0cdfe5a3525bcbd06dfa4a1564743be563fec635417116814fbc917\": container with ID starting with ff923fa0e0cdfe5a3525bcbd06dfa4a1564743be563fec635417116814fbc917 not found: ID does not exist" containerID="ff923fa0e0cdfe5a3525bcbd06dfa4a1564743be563fec635417116814fbc917" Nov 26 02:24:17 crc kubenswrapper[4766]: I1126 02:24:17.223613 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff923fa0e0cdfe5a3525bcbd06dfa4a1564743be563fec635417116814fbc917"} err="failed to get container status \"ff923fa0e0cdfe5a3525bcbd06dfa4a1564743be563fec635417116814fbc917\": rpc error: code = NotFound desc = could not find container \"ff923fa0e0cdfe5a3525bcbd06dfa4a1564743be563fec635417116814fbc917\": container with ID starting with ff923fa0e0cdfe5a3525bcbd06dfa4a1564743be563fec635417116814fbc917 not found: ID does not exist" Nov 26 02:24:17 crc kubenswrapper[4766]: I1126 02:24:17.223644 4766 scope.go:117] "RemoveContainer" containerID="b303947420b5551892475bcbfeaf8431273653a987c377b93427a3a7bc024cc9" Nov 26 02:24:17 crc kubenswrapper[4766]: E1126 02:24:17.224095 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b303947420b5551892475bcbfeaf8431273653a987c377b93427a3a7bc024cc9\": container with ID starting with b303947420b5551892475bcbfeaf8431273653a987c377b93427a3a7bc024cc9 not found: ID does not exist" containerID="b303947420b5551892475bcbfeaf8431273653a987c377b93427a3a7bc024cc9" Nov 26 02:24:17 crc kubenswrapper[4766]: I1126 02:24:17.224137 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b303947420b5551892475bcbfeaf8431273653a987c377b93427a3a7bc024cc9"} err="failed to get container status \"b303947420b5551892475bcbfeaf8431273653a987c377b93427a3a7bc024cc9\": rpc error: code = NotFound desc = could not find container \"b303947420b5551892475bcbfeaf8431273653a987c377b93427a3a7bc024cc9\": container with ID starting with b303947420b5551892475bcbfeaf8431273653a987c377b93427a3a7bc024cc9 not found: ID does not exist" Nov 26 02:24:17 crc kubenswrapper[4766]: I1126 02:24:17.828924 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:24:17 crc kubenswrapper[4766]: E1126 02:24:17.829193 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:24:17 crc kubenswrapper[4766]: I1126 02:24:17.858349 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad06a4c9-3c79-4a6f-beea-65ab4b974047" path="/var/lib/kubelet/pods/ad06a4c9-3c79-4a6f-beea-65ab4b974047/volumes" Nov 26 02:24:28 crc kubenswrapper[4766]: I1126 02:24:28.826968 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:24:28 crc kubenswrapper[4766]: E1126 02:24:28.828275 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:24:41 crc kubenswrapper[4766]: I1126 02:24:41.828303 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:24:41 crc kubenswrapper[4766]: E1126 02:24:41.829413 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:24:54 crc kubenswrapper[4766]: I1126 02:24:54.827810 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:24:54 crc kubenswrapper[4766]: E1126 02:24:54.828795 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:25:08 crc kubenswrapper[4766]: I1126 02:25:08.826630 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:25:08 crc kubenswrapper[4766]: E1126 02:25:08.827455 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:25:21 crc kubenswrapper[4766]: I1126 02:25:21.828108 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:25:21 crc kubenswrapper[4766]: E1126 02:25:21.830018 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:25:36 crc kubenswrapper[4766]: I1126 02:25:36.451783 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vc92q"] Nov 26 02:25:36 crc kubenswrapper[4766]: E1126 02:25:36.452724 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad06a4c9-3c79-4a6f-beea-65ab4b974047" containerName="registry-server" Nov 26 02:25:36 crc kubenswrapper[4766]: I1126 02:25:36.452737 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad06a4c9-3c79-4a6f-beea-65ab4b974047" containerName="registry-server" Nov 26 02:25:36 crc kubenswrapper[4766]: E1126 02:25:36.452776 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad06a4c9-3c79-4a6f-beea-65ab4b974047" containerName="extract-content" Nov 26 02:25:36 crc kubenswrapper[4766]: I1126 02:25:36.452783 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad06a4c9-3c79-4a6f-beea-65ab4b974047" containerName="extract-content" Nov 26 02:25:36 crc kubenswrapper[4766]: E1126 02:25:36.452817 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad06a4c9-3c79-4a6f-beea-65ab4b974047" containerName="extract-utilities" Nov 26 02:25:36 crc kubenswrapper[4766]: I1126 02:25:36.452825 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad06a4c9-3c79-4a6f-beea-65ab4b974047" containerName="extract-utilities" Nov 26 02:25:36 crc kubenswrapper[4766]: I1126 02:25:36.453045 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad06a4c9-3c79-4a6f-beea-65ab4b974047" containerName="registry-server" Nov 26 02:25:36 crc kubenswrapper[4766]: I1126 02:25:36.454598 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vc92q" Nov 26 02:25:36 crc kubenswrapper[4766]: I1126 02:25:36.470881 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vc92q"] Nov 26 02:25:36 crc kubenswrapper[4766]: I1126 02:25:36.546808 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42be4498-8361-407b-926f-e2c64e9b7f66-catalog-content\") pod \"community-operators-vc92q\" (UID: \"42be4498-8361-407b-926f-e2c64e9b7f66\") " pod="openshift-marketplace/community-operators-vc92q" Nov 26 02:25:36 crc kubenswrapper[4766]: I1126 02:25:36.546899 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkr88\" (UniqueName: \"kubernetes.io/projected/42be4498-8361-407b-926f-e2c64e9b7f66-kube-api-access-jkr88\") pod \"community-operators-vc92q\" (UID: \"42be4498-8361-407b-926f-e2c64e9b7f66\") " pod="openshift-marketplace/community-operators-vc92q" Nov 26 02:25:36 crc kubenswrapper[4766]: I1126 02:25:36.546965 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42be4498-8361-407b-926f-e2c64e9b7f66-utilities\") pod \"community-operators-vc92q\" (UID: \"42be4498-8361-407b-926f-e2c64e9b7f66\") " pod="openshift-marketplace/community-operators-vc92q" Nov 26 02:25:36 crc kubenswrapper[4766]: I1126 02:25:36.648548 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42be4498-8361-407b-926f-e2c64e9b7f66-catalog-content\") pod \"community-operators-vc92q\" (UID: \"42be4498-8361-407b-926f-e2c64e9b7f66\") " pod="openshift-marketplace/community-operators-vc92q" Nov 26 02:25:36 crc kubenswrapper[4766]: I1126 02:25:36.648913 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkr88\" (UniqueName: \"kubernetes.io/projected/42be4498-8361-407b-926f-e2c64e9b7f66-kube-api-access-jkr88\") pod \"community-operators-vc92q\" (UID: \"42be4498-8361-407b-926f-e2c64e9b7f66\") " pod="openshift-marketplace/community-operators-vc92q" Nov 26 02:25:36 crc kubenswrapper[4766]: I1126 02:25:36.648985 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42be4498-8361-407b-926f-e2c64e9b7f66-utilities\") pod \"community-operators-vc92q\" (UID: \"42be4498-8361-407b-926f-e2c64e9b7f66\") " pod="openshift-marketplace/community-operators-vc92q" Nov 26 02:25:36 crc kubenswrapper[4766]: I1126 02:25:36.649044 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42be4498-8361-407b-926f-e2c64e9b7f66-catalog-content\") pod \"community-operators-vc92q\" (UID: \"42be4498-8361-407b-926f-e2c64e9b7f66\") " pod="openshift-marketplace/community-operators-vc92q" Nov 26 02:25:36 crc kubenswrapper[4766]: I1126 02:25:36.649448 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42be4498-8361-407b-926f-e2c64e9b7f66-utilities\") pod \"community-operators-vc92q\" (UID: \"42be4498-8361-407b-926f-e2c64e9b7f66\") " pod="openshift-marketplace/community-operators-vc92q" Nov 26 02:25:36 crc kubenswrapper[4766]: I1126 02:25:36.667723 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkr88\" (UniqueName: \"kubernetes.io/projected/42be4498-8361-407b-926f-e2c64e9b7f66-kube-api-access-jkr88\") pod \"community-operators-vc92q\" (UID: \"42be4498-8361-407b-926f-e2c64e9b7f66\") " pod="openshift-marketplace/community-operators-vc92q" Nov 26 02:25:36 crc kubenswrapper[4766]: I1126 02:25:36.801239 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vc92q" Nov 26 02:25:36 crc kubenswrapper[4766]: I1126 02:25:36.826641 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:25:36 crc kubenswrapper[4766]: E1126 02:25:36.826937 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:25:37 crc kubenswrapper[4766]: I1126 02:25:37.341339 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vc92q"] Nov 26 02:25:38 crc kubenswrapper[4766]: I1126 02:25:38.121517 4766 generic.go:334] "Generic (PLEG): container finished" podID="42be4498-8361-407b-926f-e2c64e9b7f66" containerID="de1771f8a825ee6bbf9acc1adc7b85f221791588ba409b9dcfbfe439883bbf45" exitCode=0 Nov 26 02:25:38 crc kubenswrapper[4766]: I1126 02:25:38.121609 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vc92q" event={"ID":"42be4498-8361-407b-926f-e2c64e9b7f66","Type":"ContainerDied","Data":"de1771f8a825ee6bbf9acc1adc7b85f221791588ba409b9dcfbfe439883bbf45"} Nov 26 02:25:38 crc kubenswrapper[4766]: I1126 02:25:38.121778 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vc92q" event={"ID":"42be4498-8361-407b-926f-e2c64e9b7f66","Type":"ContainerStarted","Data":"78faeac6fa4199909fa75121464ebf97108b003398d5b707cfa30a6f95947f94"} Nov 26 02:25:39 crc kubenswrapper[4766]: I1126 02:25:39.140690 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vc92q" event={"ID":"42be4498-8361-407b-926f-e2c64e9b7f66","Type":"ContainerStarted","Data":"d0a1428cacd0c3949205603f541dae217c8cf78f262a152cea7434c30f0afaeb"} Nov 26 02:25:41 crc kubenswrapper[4766]: I1126 02:25:41.165324 4766 generic.go:334] "Generic (PLEG): container finished" podID="42be4498-8361-407b-926f-e2c64e9b7f66" containerID="d0a1428cacd0c3949205603f541dae217c8cf78f262a152cea7434c30f0afaeb" exitCode=0 Nov 26 02:25:41 crc kubenswrapper[4766]: I1126 02:25:41.165374 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vc92q" event={"ID":"42be4498-8361-407b-926f-e2c64e9b7f66","Type":"ContainerDied","Data":"d0a1428cacd0c3949205603f541dae217c8cf78f262a152cea7434c30f0afaeb"} Nov 26 02:25:42 crc kubenswrapper[4766]: I1126 02:25:42.189726 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vc92q" event={"ID":"42be4498-8361-407b-926f-e2c64e9b7f66","Type":"ContainerStarted","Data":"6e014b74901594d82ded215a537e40ea022e302ab9ebca1e14249b7df564a0dc"} Nov 26 02:25:42 crc kubenswrapper[4766]: I1126 02:25:42.230031 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vc92q" podStartSLOduration=2.782477382 podStartE2EDuration="6.230013638s" podCreationTimestamp="2025-11-26 02:25:36 +0000 UTC" firstStartedPulling="2025-11-26 02:25:38.123615952 +0000 UTC m=+7318.972386382" lastFinishedPulling="2025-11-26 02:25:41.571152208 +0000 UTC m=+7322.419922638" observedRunningTime="2025-11-26 02:25:42.213944335 +0000 UTC m=+7323.062714815" watchObservedRunningTime="2025-11-26 02:25:42.230013638 +0000 UTC m=+7323.078784068" Nov 26 02:25:43 crc kubenswrapper[4766]: I1126 02:25:43.200377 4766 generic.go:334] "Generic (PLEG): container finished" podID="edfc5f7f-1ff2-4147-b68e-b9be01ff6a99" containerID="d6cb0b3456d7352838582860f7ebfad091a9a9316ec28d54973e7e546b75c0e6" exitCode=0 Nov 26 02:25:43 crc kubenswrapper[4766]: I1126 02:25:43.200423 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99","Type":"ContainerDied","Data":"d6cb0b3456d7352838582860f7ebfad091a9a9316ec28d54973e7e546b75c0e6"} Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.677441 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.748016 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vcrc\" (UniqueName: \"kubernetes.io/projected/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-kube-api-access-6vcrc\") pod \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.748149 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-ca-certs\") pod \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.748209 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-test-operator-ephemeral-temporary\") pod \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.748231 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.748260 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-ssh-key\") pod \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.748367 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-config-data\") pod \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.748392 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-openstack-config-secret\") pod \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.748414 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-openstack-config\") pod \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.748481 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-test-operator-ephemeral-workdir\") pod \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\" (UID: \"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99\") " Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.755586 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "edfc5f7f-1ff2-4147-b68e-b9be01ff6a99" (UID: "edfc5f7f-1ff2-4147-b68e-b9be01ff6a99"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.757966 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-config-data" (OuterVolumeSpecName: "config-data") pod "edfc5f7f-1ff2-4147-b68e-b9be01ff6a99" (UID: "edfc5f7f-1ff2-4147-b68e-b9be01ff6a99"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.758779 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "edfc5f7f-1ff2-4147-b68e-b9be01ff6a99" (UID: "edfc5f7f-1ff2-4147-b68e-b9be01ff6a99"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.759199 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "edfc5f7f-1ff2-4147-b68e-b9be01ff6a99" (UID: "edfc5f7f-1ff2-4147-b68e-b9be01ff6a99"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.762400 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-kube-api-access-6vcrc" (OuterVolumeSpecName: "kube-api-access-6vcrc") pod "edfc5f7f-1ff2-4147-b68e-b9be01ff6a99" (UID: "edfc5f7f-1ff2-4147-b68e-b9be01ff6a99"). InnerVolumeSpecName "kube-api-access-6vcrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.789758 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "edfc5f7f-1ff2-4147-b68e-b9be01ff6a99" (UID: "edfc5f7f-1ff2-4147-b68e-b9be01ff6a99"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.799919 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "edfc5f7f-1ff2-4147-b68e-b9be01ff6a99" (UID: "edfc5f7f-1ff2-4147-b68e-b9be01ff6a99"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.833816 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "edfc5f7f-1ff2-4147-b68e-b9be01ff6a99" (UID: "edfc5f7f-1ff2-4147-b68e-b9be01ff6a99"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.847454 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "edfc5f7f-1ff2-4147-b68e-b9be01ff6a99" (UID: "edfc5f7f-1ff2-4147-b68e-b9be01ff6a99"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.851311 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vcrc\" (UniqueName: \"kubernetes.io/projected/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-kube-api-access-6vcrc\") on node \"crc\" DevicePath \"\"" Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.851349 4766 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.851366 4766 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.852149 4766 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.852176 4766 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.852189 4766 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.852204 4766 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.852219 4766 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.852232 4766 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/edfc5f7f-1ff2-4147-b68e-b9be01ff6a99-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.881188 4766 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 26 02:25:44 crc kubenswrapper[4766]: I1126 02:25:44.953871 4766 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 26 02:25:45 crc kubenswrapper[4766]: I1126 02:25:45.236638 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"edfc5f7f-1ff2-4147-b68e-b9be01ff6a99","Type":"ContainerDied","Data":"ef1d76ebb4a4de9034c8eb4c573b72645fa07647927c25ac5c62f0025758a60f"} Nov 26 02:25:45 crc kubenswrapper[4766]: I1126 02:25:45.236968 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef1d76ebb4a4de9034c8eb4c573b72645fa07647927c25ac5c62f0025758a60f" Nov 26 02:25:45 crc kubenswrapper[4766]: I1126 02:25:45.237058 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 26 02:25:46 crc kubenswrapper[4766]: I1126 02:25:46.801748 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vc92q" Nov 26 02:25:46 crc kubenswrapper[4766]: I1126 02:25:46.802047 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vc92q" Nov 26 02:25:46 crc kubenswrapper[4766]: I1126 02:25:46.866324 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vc92q" Nov 26 02:25:47 crc kubenswrapper[4766]: I1126 02:25:47.334558 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vc92q" Nov 26 02:25:47 crc kubenswrapper[4766]: I1126 02:25:47.401991 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vc92q"] Nov 26 02:25:49 crc kubenswrapper[4766]: I1126 02:25:49.288028 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vc92q" podUID="42be4498-8361-407b-926f-e2c64e9b7f66" containerName="registry-server" containerID="cri-o://6e014b74901594d82ded215a537e40ea022e302ab9ebca1e14249b7df564a0dc" gracePeriod=2 Nov 26 02:25:49 crc kubenswrapper[4766]: I1126 02:25:49.887347 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vc92q" Nov 26 02:25:49 crc kubenswrapper[4766]: I1126 02:25:49.981637 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkr88\" (UniqueName: \"kubernetes.io/projected/42be4498-8361-407b-926f-e2c64e9b7f66-kube-api-access-jkr88\") pod \"42be4498-8361-407b-926f-e2c64e9b7f66\" (UID: \"42be4498-8361-407b-926f-e2c64e9b7f66\") " Nov 26 02:25:49 crc kubenswrapper[4766]: I1126 02:25:49.982396 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42be4498-8361-407b-926f-e2c64e9b7f66-catalog-content\") pod \"42be4498-8361-407b-926f-e2c64e9b7f66\" (UID: \"42be4498-8361-407b-926f-e2c64e9b7f66\") " Nov 26 02:25:49 crc kubenswrapper[4766]: I1126 02:25:49.982479 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42be4498-8361-407b-926f-e2c64e9b7f66-utilities\") pod \"42be4498-8361-407b-926f-e2c64e9b7f66\" (UID: \"42be4498-8361-407b-926f-e2c64e9b7f66\") " Nov 26 02:25:49 crc kubenswrapper[4766]: I1126 02:25:49.983804 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42be4498-8361-407b-926f-e2c64e9b7f66-utilities" (OuterVolumeSpecName: "utilities") pod "42be4498-8361-407b-926f-e2c64e9b7f66" (UID: "42be4498-8361-407b-926f-e2c64e9b7f66"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:25:50 crc kubenswrapper[4766]: I1126 02:25:50.019524 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42be4498-8361-407b-926f-e2c64e9b7f66-kube-api-access-jkr88" (OuterVolumeSpecName: "kube-api-access-jkr88") pod "42be4498-8361-407b-926f-e2c64e9b7f66" (UID: "42be4498-8361-407b-926f-e2c64e9b7f66"). InnerVolumeSpecName "kube-api-access-jkr88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:25:50 crc kubenswrapper[4766]: I1126 02:25:50.078319 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42be4498-8361-407b-926f-e2c64e9b7f66-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "42be4498-8361-407b-926f-e2c64e9b7f66" (UID: "42be4498-8361-407b-926f-e2c64e9b7f66"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:25:50 crc kubenswrapper[4766]: I1126 02:25:50.086754 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42be4498-8361-407b-926f-e2c64e9b7f66-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 02:25:50 crc kubenswrapper[4766]: I1126 02:25:50.086810 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42be4498-8361-407b-926f-e2c64e9b7f66-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 02:25:50 crc kubenswrapper[4766]: I1126 02:25:50.086830 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkr88\" (UniqueName: \"kubernetes.io/projected/42be4498-8361-407b-926f-e2c64e9b7f66-kube-api-access-jkr88\") on node \"crc\" DevicePath \"\"" Nov 26 02:25:50 crc kubenswrapper[4766]: I1126 02:25:50.304232 4766 generic.go:334] "Generic (PLEG): container finished" podID="42be4498-8361-407b-926f-e2c64e9b7f66" containerID="6e014b74901594d82ded215a537e40ea022e302ab9ebca1e14249b7df564a0dc" exitCode=0 Nov 26 02:25:50 crc kubenswrapper[4766]: I1126 02:25:50.304306 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vc92q" event={"ID":"42be4498-8361-407b-926f-e2c64e9b7f66","Type":"ContainerDied","Data":"6e014b74901594d82ded215a537e40ea022e302ab9ebca1e14249b7df564a0dc"} Nov 26 02:25:50 crc kubenswrapper[4766]: I1126 02:25:50.304377 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vc92q" event={"ID":"42be4498-8361-407b-926f-e2c64e9b7f66","Type":"ContainerDied","Data":"78faeac6fa4199909fa75121464ebf97108b003398d5b707cfa30a6f95947f94"} Nov 26 02:25:50 crc kubenswrapper[4766]: I1126 02:25:50.304415 4766 scope.go:117] "RemoveContainer" containerID="6e014b74901594d82ded215a537e40ea022e302ab9ebca1e14249b7df564a0dc" Nov 26 02:25:50 crc kubenswrapper[4766]: I1126 02:25:50.304331 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vc92q" Nov 26 02:25:50 crc kubenswrapper[4766]: I1126 02:25:50.331940 4766 scope.go:117] "RemoveContainer" containerID="d0a1428cacd0c3949205603f541dae217c8cf78f262a152cea7434c30f0afaeb" Nov 26 02:25:50 crc kubenswrapper[4766]: I1126 02:25:50.347134 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vc92q"] Nov 26 02:25:50 crc kubenswrapper[4766]: I1126 02:25:50.365983 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vc92q"] Nov 26 02:25:50 crc kubenswrapper[4766]: I1126 02:25:50.375557 4766 scope.go:117] "RemoveContainer" containerID="de1771f8a825ee6bbf9acc1adc7b85f221791588ba409b9dcfbfe439883bbf45" Nov 26 02:25:50 crc kubenswrapper[4766]: I1126 02:25:50.447065 4766 scope.go:117] "RemoveContainer" containerID="6e014b74901594d82ded215a537e40ea022e302ab9ebca1e14249b7df564a0dc" Nov 26 02:25:50 crc kubenswrapper[4766]: E1126 02:25:50.447834 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e014b74901594d82ded215a537e40ea022e302ab9ebca1e14249b7df564a0dc\": container with ID starting with 6e014b74901594d82ded215a537e40ea022e302ab9ebca1e14249b7df564a0dc not found: ID does not exist" containerID="6e014b74901594d82ded215a537e40ea022e302ab9ebca1e14249b7df564a0dc" Nov 26 02:25:50 crc kubenswrapper[4766]: I1126 02:25:50.447874 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e014b74901594d82ded215a537e40ea022e302ab9ebca1e14249b7df564a0dc"} err="failed to get container status \"6e014b74901594d82ded215a537e40ea022e302ab9ebca1e14249b7df564a0dc\": rpc error: code = NotFound desc = could not find container \"6e014b74901594d82ded215a537e40ea022e302ab9ebca1e14249b7df564a0dc\": container with ID starting with 6e014b74901594d82ded215a537e40ea022e302ab9ebca1e14249b7df564a0dc not found: ID does not exist" Nov 26 02:25:50 crc kubenswrapper[4766]: I1126 02:25:50.447910 4766 scope.go:117] "RemoveContainer" containerID="d0a1428cacd0c3949205603f541dae217c8cf78f262a152cea7434c30f0afaeb" Nov 26 02:25:50 crc kubenswrapper[4766]: E1126 02:25:50.448240 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0a1428cacd0c3949205603f541dae217c8cf78f262a152cea7434c30f0afaeb\": container with ID starting with d0a1428cacd0c3949205603f541dae217c8cf78f262a152cea7434c30f0afaeb not found: ID does not exist" containerID="d0a1428cacd0c3949205603f541dae217c8cf78f262a152cea7434c30f0afaeb" Nov 26 02:25:50 crc kubenswrapper[4766]: I1126 02:25:50.448285 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0a1428cacd0c3949205603f541dae217c8cf78f262a152cea7434c30f0afaeb"} err="failed to get container status \"d0a1428cacd0c3949205603f541dae217c8cf78f262a152cea7434c30f0afaeb\": rpc error: code = NotFound desc = could not find container \"d0a1428cacd0c3949205603f541dae217c8cf78f262a152cea7434c30f0afaeb\": container with ID starting with d0a1428cacd0c3949205603f541dae217c8cf78f262a152cea7434c30f0afaeb not found: ID does not exist" Nov 26 02:25:50 crc kubenswrapper[4766]: I1126 02:25:50.448316 4766 scope.go:117] "RemoveContainer" containerID="de1771f8a825ee6bbf9acc1adc7b85f221791588ba409b9dcfbfe439883bbf45" Nov 26 02:25:50 crc kubenswrapper[4766]: E1126 02:25:50.448601 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de1771f8a825ee6bbf9acc1adc7b85f221791588ba409b9dcfbfe439883bbf45\": container with ID starting with de1771f8a825ee6bbf9acc1adc7b85f221791588ba409b9dcfbfe439883bbf45 not found: ID does not exist" containerID="de1771f8a825ee6bbf9acc1adc7b85f221791588ba409b9dcfbfe439883bbf45" Nov 26 02:25:50 crc kubenswrapper[4766]: I1126 02:25:50.448731 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de1771f8a825ee6bbf9acc1adc7b85f221791588ba409b9dcfbfe439883bbf45"} err="failed to get container status \"de1771f8a825ee6bbf9acc1adc7b85f221791588ba409b9dcfbfe439883bbf45\": rpc error: code = NotFound desc = could not find container \"de1771f8a825ee6bbf9acc1adc7b85f221791588ba409b9dcfbfe439883bbf45\": container with ID starting with de1771f8a825ee6bbf9acc1adc7b85f221791588ba409b9dcfbfe439883bbf45 not found: ID does not exist" Nov 26 02:25:51 crc kubenswrapper[4766]: I1126 02:25:51.827792 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:25:51 crc kubenswrapper[4766]: E1126 02:25:51.828448 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:25:51 crc kubenswrapper[4766]: I1126 02:25:51.843908 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42be4498-8361-407b-926f-e2c64e9b7f66" path="/var/lib/kubelet/pods/42be4498-8361-407b-926f-e2c64e9b7f66/volumes" Nov 26 02:25:56 crc kubenswrapper[4766]: I1126 02:25:56.694277 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 26 02:25:56 crc kubenswrapper[4766]: E1126 02:25:56.695376 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42be4498-8361-407b-926f-e2c64e9b7f66" containerName="extract-utilities" Nov 26 02:25:56 crc kubenswrapper[4766]: I1126 02:25:56.695392 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="42be4498-8361-407b-926f-e2c64e9b7f66" containerName="extract-utilities" Nov 26 02:25:56 crc kubenswrapper[4766]: E1126 02:25:56.695434 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42be4498-8361-407b-926f-e2c64e9b7f66" containerName="extract-content" Nov 26 02:25:56 crc kubenswrapper[4766]: I1126 02:25:56.695440 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="42be4498-8361-407b-926f-e2c64e9b7f66" containerName="extract-content" Nov 26 02:25:56 crc kubenswrapper[4766]: E1126 02:25:56.695477 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42be4498-8361-407b-926f-e2c64e9b7f66" containerName="registry-server" Nov 26 02:25:56 crc kubenswrapper[4766]: I1126 02:25:56.695483 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="42be4498-8361-407b-926f-e2c64e9b7f66" containerName="registry-server" Nov 26 02:25:56 crc kubenswrapper[4766]: E1126 02:25:56.695499 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edfc5f7f-1ff2-4147-b68e-b9be01ff6a99" containerName="tempest-tests-tempest-tests-runner" Nov 26 02:25:56 crc kubenswrapper[4766]: I1126 02:25:56.695506 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="edfc5f7f-1ff2-4147-b68e-b9be01ff6a99" containerName="tempest-tests-tempest-tests-runner" Nov 26 02:25:56 crc kubenswrapper[4766]: I1126 02:25:56.695734 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="edfc5f7f-1ff2-4147-b68e-b9be01ff6a99" containerName="tempest-tests-tempest-tests-runner" Nov 26 02:25:56 crc kubenswrapper[4766]: I1126 02:25:56.695753 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="42be4498-8361-407b-926f-e2c64e9b7f66" containerName="registry-server" Nov 26 02:25:56 crc kubenswrapper[4766]: I1126 02:25:56.696800 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 26 02:25:56 crc kubenswrapper[4766]: I1126 02:25:56.701159 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-kq4g7" Nov 26 02:25:56 crc kubenswrapper[4766]: I1126 02:25:56.713388 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 26 02:25:56 crc kubenswrapper[4766]: I1126 02:25:56.753240 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f4a47960-5a63-4a92-a4e1-89d48aaa6515\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 26 02:25:56 crc kubenswrapper[4766]: I1126 02:25:56.753328 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bkk6\" (UniqueName: \"kubernetes.io/projected/f4a47960-5a63-4a92-a4e1-89d48aaa6515-kube-api-access-4bkk6\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f4a47960-5a63-4a92-a4e1-89d48aaa6515\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 26 02:25:56 crc kubenswrapper[4766]: I1126 02:25:56.855232 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f4a47960-5a63-4a92-a4e1-89d48aaa6515\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 26 02:25:56 crc kubenswrapper[4766]: I1126 02:25:56.855335 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bkk6\" (UniqueName: \"kubernetes.io/projected/f4a47960-5a63-4a92-a4e1-89d48aaa6515-kube-api-access-4bkk6\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f4a47960-5a63-4a92-a4e1-89d48aaa6515\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 26 02:25:56 crc kubenswrapper[4766]: I1126 02:25:56.855793 4766 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f4a47960-5a63-4a92-a4e1-89d48aaa6515\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 26 02:25:56 crc kubenswrapper[4766]: I1126 02:25:56.882490 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bkk6\" (UniqueName: \"kubernetes.io/projected/f4a47960-5a63-4a92-a4e1-89d48aaa6515-kube-api-access-4bkk6\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f4a47960-5a63-4a92-a4e1-89d48aaa6515\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 26 02:25:56 crc kubenswrapper[4766]: I1126 02:25:56.891627 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f4a47960-5a63-4a92-a4e1-89d48aaa6515\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 26 02:25:57 crc kubenswrapper[4766]: I1126 02:25:57.019914 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 26 02:25:57 crc kubenswrapper[4766]: I1126 02:25:57.560246 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 26 02:25:58 crc kubenswrapper[4766]: I1126 02:25:58.403575 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"f4a47960-5a63-4a92-a4e1-89d48aaa6515","Type":"ContainerStarted","Data":"c27d3847f7bb0419cc85d17249d119747cfdd1814cf41ee4ae939484e642bbaa"} Nov 26 02:25:59 crc kubenswrapper[4766]: I1126 02:25:59.413444 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"f4a47960-5a63-4a92-a4e1-89d48aaa6515","Type":"ContainerStarted","Data":"454b4288ba32db4e71d4c050e0c5736ee8f973cef7a043a6f80b57ab628e16fc"} Nov 26 02:25:59 crc kubenswrapper[4766]: I1126 02:25:59.435448 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.389816041 podStartE2EDuration="3.435421692s" podCreationTimestamp="2025-11-26 02:25:56 +0000 UTC" firstStartedPulling="2025-11-26 02:25:57.568102827 +0000 UTC m=+7338.416873267" lastFinishedPulling="2025-11-26 02:25:58.613708488 +0000 UTC m=+7339.462478918" observedRunningTime="2025-11-26 02:25:59.425640722 +0000 UTC m=+7340.274411152" watchObservedRunningTime="2025-11-26 02:25:59.435421692 +0000 UTC m=+7340.284192152" Nov 26 02:26:06 crc kubenswrapper[4766]: I1126 02:26:06.826421 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:26:06 crc kubenswrapper[4766]: E1126 02:26:06.827277 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:26:18 crc kubenswrapper[4766]: I1126 02:26:18.826974 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:26:18 crc kubenswrapper[4766]: E1126 02:26:18.827691 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:26:30 crc kubenswrapper[4766]: I1126 02:26:30.438790 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tv2fj/must-gather-sfxwh"] Nov 26 02:26:30 crc kubenswrapper[4766]: I1126 02:26:30.441363 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv2fj/must-gather-sfxwh" Nov 26 02:26:30 crc kubenswrapper[4766]: I1126 02:26:30.445337 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-tv2fj"/"kube-root-ca.crt" Nov 26 02:26:30 crc kubenswrapper[4766]: I1126 02:26:30.445360 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-tv2fj"/"default-dockercfg-hcmpd" Nov 26 02:26:30 crc kubenswrapper[4766]: I1126 02:26:30.447781 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-tv2fj"/"openshift-service-ca.crt" Nov 26 02:26:30 crc kubenswrapper[4766]: I1126 02:26:30.456706 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-tv2fj/must-gather-sfxwh"] Nov 26 02:26:30 crc kubenswrapper[4766]: I1126 02:26:30.627449 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1e340437-6f94-4ea6-82c5-bed5ac1b9c71-must-gather-output\") pod \"must-gather-sfxwh\" (UID: \"1e340437-6f94-4ea6-82c5-bed5ac1b9c71\") " pod="openshift-must-gather-tv2fj/must-gather-sfxwh" Nov 26 02:26:30 crc kubenswrapper[4766]: I1126 02:26:30.627882 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbg78\" (UniqueName: \"kubernetes.io/projected/1e340437-6f94-4ea6-82c5-bed5ac1b9c71-kube-api-access-gbg78\") pod \"must-gather-sfxwh\" (UID: \"1e340437-6f94-4ea6-82c5-bed5ac1b9c71\") " pod="openshift-must-gather-tv2fj/must-gather-sfxwh" Nov 26 02:26:30 crc kubenswrapper[4766]: I1126 02:26:30.729825 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbg78\" (UniqueName: \"kubernetes.io/projected/1e340437-6f94-4ea6-82c5-bed5ac1b9c71-kube-api-access-gbg78\") pod \"must-gather-sfxwh\" (UID: \"1e340437-6f94-4ea6-82c5-bed5ac1b9c71\") " pod="openshift-must-gather-tv2fj/must-gather-sfxwh" Nov 26 02:26:30 crc kubenswrapper[4766]: I1126 02:26:30.730056 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1e340437-6f94-4ea6-82c5-bed5ac1b9c71-must-gather-output\") pod \"must-gather-sfxwh\" (UID: \"1e340437-6f94-4ea6-82c5-bed5ac1b9c71\") " pod="openshift-must-gather-tv2fj/must-gather-sfxwh" Nov 26 02:26:30 crc kubenswrapper[4766]: I1126 02:26:30.730619 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1e340437-6f94-4ea6-82c5-bed5ac1b9c71-must-gather-output\") pod \"must-gather-sfxwh\" (UID: \"1e340437-6f94-4ea6-82c5-bed5ac1b9c71\") " pod="openshift-must-gather-tv2fj/must-gather-sfxwh" Nov 26 02:26:30 crc kubenswrapper[4766]: I1126 02:26:30.756276 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbg78\" (UniqueName: \"kubernetes.io/projected/1e340437-6f94-4ea6-82c5-bed5ac1b9c71-kube-api-access-gbg78\") pod \"must-gather-sfxwh\" (UID: \"1e340437-6f94-4ea6-82c5-bed5ac1b9c71\") " pod="openshift-must-gather-tv2fj/must-gather-sfxwh" Nov 26 02:26:30 crc kubenswrapper[4766]: I1126 02:26:30.768112 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv2fj/must-gather-sfxwh" Nov 26 02:26:31 crc kubenswrapper[4766]: I1126 02:26:31.267679 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-tv2fj/must-gather-sfxwh"] Nov 26 02:26:31 crc kubenswrapper[4766]: I1126 02:26:31.826873 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:26:31 crc kubenswrapper[4766]: E1126 02:26:31.827461 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:26:31 crc kubenswrapper[4766]: I1126 02:26:31.858135 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv2fj/must-gather-sfxwh" event={"ID":"1e340437-6f94-4ea6-82c5-bed5ac1b9c71","Type":"ContainerStarted","Data":"b2407d7fb259e0674f72b45674575bd7ce1b83bfacac8db54adb8044fbbf5f12"} Nov 26 02:26:35 crc kubenswrapper[4766]: I1126 02:26:35.902373 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv2fj/must-gather-sfxwh" event={"ID":"1e340437-6f94-4ea6-82c5-bed5ac1b9c71","Type":"ContainerStarted","Data":"ae717b377a1717d9d05706682945c3e64ebb7ca8aecf7b9a4ee91bbb6e288daf"} Nov 26 02:26:36 crc kubenswrapper[4766]: I1126 02:26:36.916756 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv2fj/must-gather-sfxwh" event={"ID":"1e340437-6f94-4ea6-82c5-bed5ac1b9c71","Type":"ContainerStarted","Data":"7836e1c1d70e7eff8c083085cf73a68a29333eb8b481ac700911467164edc9ee"} Nov 26 02:26:36 crc kubenswrapper[4766]: I1126 02:26:36.937573 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-tv2fj/must-gather-sfxwh" podStartSLOduration=2.986888987 podStartE2EDuration="6.937547663s" podCreationTimestamp="2025-11-26 02:26:30 +0000 UTC" firstStartedPulling="2025-11-26 02:26:31.261845742 +0000 UTC m=+7372.110616172" lastFinishedPulling="2025-11-26 02:26:35.212504408 +0000 UTC m=+7376.061274848" observedRunningTime="2025-11-26 02:26:36.935059792 +0000 UTC m=+7377.783830242" watchObservedRunningTime="2025-11-26 02:26:36.937547663 +0000 UTC m=+7377.786318133" Nov 26 02:26:41 crc kubenswrapper[4766]: I1126 02:26:41.642806 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tv2fj/crc-debug-49rpf"] Nov 26 02:26:41 crc kubenswrapper[4766]: I1126 02:26:41.644930 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv2fj/crc-debug-49rpf" Nov 26 02:26:41 crc kubenswrapper[4766]: I1126 02:26:41.830951 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/53360513-f8aa-41d6-94d2-fc348e4f2e16-host\") pod \"crc-debug-49rpf\" (UID: \"53360513-f8aa-41d6-94d2-fc348e4f2e16\") " pod="openshift-must-gather-tv2fj/crc-debug-49rpf" Nov 26 02:26:41 crc kubenswrapper[4766]: I1126 02:26:41.831018 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dn2s\" (UniqueName: \"kubernetes.io/projected/53360513-f8aa-41d6-94d2-fc348e4f2e16-kube-api-access-2dn2s\") pod \"crc-debug-49rpf\" (UID: \"53360513-f8aa-41d6-94d2-fc348e4f2e16\") " pod="openshift-must-gather-tv2fj/crc-debug-49rpf" Nov 26 02:26:41 crc kubenswrapper[4766]: I1126 02:26:41.933564 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dn2s\" (UniqueName: \"kubernetes.io/projected/53360513-f8aa-41d6-94d2-fc348e4f2e16-kube-api-access-2dn2s\") pod \"crc-debug-49rpf\" (UID: \"53360513-f8aa-41d6-94d2-fc348e4f2e16\") " pod="openshift-must-gather-tv2fj/crc-debug-49rpf" Nov 26 02:26:41 crc kubenswrapper[4766]: I1126 02:26:41.933918 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/53360513-f8aa-41d6-94d2-fc348e4f2e16-host\") pod \"crc-debug-49rpf\" (UID: \"53360513-f8aa-41d6-94d2-fc348e4f2e16\") " pod="openshift-must-gather-tv2fj/crc-debug-49rpf" Nov 26 02:26:41 crc kubenswrapper[4766]: I1126 02:26:41.934079 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/53360513-f8aa-41d6-94d2-fc348e4f2e16-host\") pod \"crc-debug-49rpf\" (UID: \"53360513-f8aa-41d6-94d2-fc348e4f2e16\") " pod="openshift-must-gather-tv2fj/crc-debug-49rpf" Nov 26 02:26:41 crc kubenswrapper[4766]: I1126 02:26:41.955353 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dn2s\" (UniqueName: \"kubernetes.io/projected/53360513-f8aa-41d6-94d2-fc348e4f2e16-kube-api-access-2dn2s\") pod \"crc-debug-49rpf\" (UID: \"53360513-f8aa-41d6-94d2-fc348e4f2e16\") " pod="openshift-must-gather-tv2fj/crc-debug-49rpf" Nov 26 02:26:41 crc kubenswrapper[4766]: I1126 02:26:41.964378 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv2fj/crc-debug-49rpf" Nov 26 02:26:42 crc kubenswrapper[4766]: I1126 02:26:42.991308 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv2fj/crc-debug-49rpf" event={"ID":"53360513-f8aa-41d6-94d2-fc348e4f2e16","Type":"ContainerStarted","Data":"ad6d009e218bd30048e002c2706a833c306a9e0430857a9f1e2a05aaeaae9367"} Nov 26 02:26:44 crc kubenswrapper[4766]: I1126 02:26:44.828027 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:26:46 crc kubenswrapper[4766]: I1126 02:26:46.035208 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"e12146d79b16ca01dd3ddaff20fdb9b94b5dd64df0a00c9f78979d3aecbe9802"} Nov 26 02:26:54 crc kubenswrapper[4766]: I1126 02:26:54.140272 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv2fj/crc-debug-49rpf" event={"ID":"53360513-f8aa-41d6-94d2-fc348e4f2e16","Type":"ContainerStarted","Data":"50ec60584e0e9b18b2c90f62dc5ed056feb979841a2b8cbf242dc1da347c1130"} Nov 26 02:26:54 crc kubenswrapper[4766]: I1126 02:26:54.156598 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-tv2fj/crc-debug-49rpf" podStartSLOduration=1.386089856 podStartE2EDuration="13.15657551s" podCreationTimestamp="2025-11-26 02:26:41 +0000 UTC" firstStartedPulling="2025-11-26 02:26:42.041842093 +0000 UTC m=+7382.890612523" lastFinishedPulling="2025-11-26 02:26:53.812327747 +0000 UTC m=+7394.661098177" observedRunningTime="2025-11-26 02:26:54.15497626 +0000 UTC m=+7395.003746690" watchObservedRunningTime="2025-11-26 02:26:54.15657551 +0000 UTC m=+7395.005345940" Nov 26 02:27:46 crc kubenswrapper[4766]: I1126 02:27:46.714111 4766 generic.go:334] "Generic (PLEG): container finished" podID="53360513-f8aa-41d6-94d2-fc348e4f2e16" containerID="50ec60584e0e9b18b2c90f62dc5ed056feb979841a2b8cbf242dc1da347c1130" exitCode=0 Nov 26 02:27:46 crc kubenswrapper[4766]: I1126 02:27:46.714218 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv2fj/crc-debug-49rpf" event={"ID":"53360513-f8aa-41d6-94d2-fc348e4f2e16","Type":"ContainerDied","Data":"50ec60584e0e9b18b2c90f62dc5ed056feb979841a2b8cbf242dc1da347c1130"} Nov 26 02:27:47 crc kubenswrapper[4766]: I1126 02:27:47.865487 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv2fj/crc-debug-49rpf" Nov 26 02:27:47 crc kubenswrapper[4766]: I1126 02:27:47.928624 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tv2fj/crc-debug-49rpf"] Nov 26 02:27:47 crc kubenswrapper[4766]: I1126 02:27:47.942331 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tv2fj/crc-debug-49rpf"] Nov 26 02:27:47 crc kubenswrapper[4766]: I1126 02:27:47.964432 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dn2s\" (UniqueName: \"kubernetes.io/projected/53360513-f8aa-41d6-94d2-fc348e4f2e16-kube-api-access-2dn2s\") pod \"53360513-f8aa-41d6-94d2-fc348e4f2e16\" (UID: \"53360513-f8aa-41d6-94d2-fc348e4f2e16\") " Nov 26 02:27:47 crc kubenswrapper[4766]: I1126 02:27:47.964671 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/53360513-f8aa-41d6-94d2-fc348e4f2e16-host\") pod \"53360513-f8aa-41d6-94d2-fc348e4f2e16\" (UID: \"53360513-f8aa-41d6-94d2-fc348e4f2e16\") " Nov 26 02:27:47 crc kubenswrapper[4766]: I1126 02:27:47.964954 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53360513-f8aa-41d6-94d2-fc348e4f2e16-host" (OuterVolumeSpecName: "host") pod "53360513-f8aa-41d6-94d2-fc348e4f2e16" (UID: "53360513-f8aa-41d6-94d2-fc348e4f2e16"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 02:27:47 crc kubenswrapper[4766]: I1126 02:27:47.965477 4766 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/53360513-f8aa-41d6-94d2-fc348e4f2e16-host\") on node \"crc\" DevicePath \"\"" Nov 26 02:27:47 crc kubenswrapper[4766]: I1126 02:27:47.975928 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53360513-f8aa-41d6-94d2-fc348e4f2e16-kube-api-access-2dn2s" (OuterVolumeSpecName: "kube-api-access-2dn2s") pod "53360513-f8aa-41d6-94d2-fc348e4f2e16" (UID: "53360513-f8aa-41d6-94d2-fc348e4f2e16"). InnerVolumeSpecName "kube-api-access-2dn2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:27:48 crc kubenswrapper[4766]: I1126 02:27:48.067668 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dn2s\" (UniqueName: \"kubernetes.io/projected/53360513-f8aa-41d6-94d2-fc348e4f2e16-kube-api-access-2dn2s\") on node \"crc\" DevicePath \"\"" Nov 26 02:27:48 crc kubenswrapper[4766]: I1126 02:27:48.753424 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad6d009e218bd30048e002c2706a833c306a9e0430857a9f1e2a05aaeaae9367" Nov 26 02:27:48 crc kubenswrapper[4766]: I1126 02:27:48.754706 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv2fj/crc-debug-49rpf" Nov 26 02:27:49 crc kubenswrapper[4766]: I1126 02:27:49.161119 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tv2fj/crc-debug-6gtq2"] Nov 26 02:27:49 crc kubenswrapper[4766]: E1126 02:27:49.162038 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53360513-f8aa-41d6-94d2-fc348e4f2e16" containerName="container-00" Nov 26 02:27:49 crc kubenswrapper[4766]: I1126 02:27:49.162073 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="53360513-f8aa-41d6-94d2-fc348e4f2e16" containerName="container-00" Nov 26 02:27:49 crc kubenswrapper[4766]: I1126 02:27:49.162746 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="53360513-f8aa-41d6-94d2-fc348e4f2e16" containerName="container-00" Nov 26 02:27:49 crc kubenswrapper[4766]: I1126 02:27:49.164320 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv2fj/crc-debug-6gtq2" Nov 26 02:27:49 crc kubenswrapper[4766]: I1126 02:27:49.298692 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee36ddc9-aab0-4c21-bdb8-96ba138ad71f-host\") pod \"crc-debug-6gtq2\" (UID: \"ee36ddc9-aab0-4c21-bdb8-96ba138ad71f\") " pod="openshift-must-gather-tv2fj/crc-debug-6gtq2" Nov 26 02:27:49 crc kubenswrapper[4766]: I1126 02:27:49.298861 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk7kq\" (UniqueName: \"kubernetes.io/projected/ee36ddc9-aab0-4c21-bdb8-96ba138ad71f-kube-api-access-qk7kq\") pod \"crc-debug-6gtq2\" (UID: \"ee36ddc9-aab0-4c21-bdb8-96ba138ad71f\") " pod="openshift-must-gather-tv2fj/crc-debug-6gtq2" Nov 26 02:27:49 crc kubenswrapper[4766]: I1126 02:27:49.401630 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee36ddc9-aab0-4c21-bdb8-96ba138ad71f-host\") pod \"crc-debug-6gtq2\" (UID: \"ee36ddc9-aab0-4c21-bdb8-96ba138ad71f\") " pod="openshift-must-gather-tv2fj/crc-debug-6gtq2" Nov 26 02:27:49 crc kubenswrapper[4766]: I1126 02:27:49.401895 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk7kq\" (UniqueName: \"kubernetes.io/projected/ee36ddc9-aab0-4c21-bdb8-96ba138ad71f-kube-api-access-qk7kq\") pod \"crc-debug-6gtq2\" (UID: \"ee36ddc9-aab0-4c21-bdb8-96ba138ad71f\") " pod="openshift-must-gather-tv2fj/crc-debug-6gtq2" Nov 26 02:27:49 crc kubenswrapper[4766]: I1126 02:27:49.402282 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee36ddc9-aab0-4c21-bdb8-96ba138ad71f-host\") pod \"crc-debug-6gtq2\" (UID: \"ee36ddc9-aab0-4c21-bdb8-96ba138ad71f\") " pod="openshift-must-gather-tv2fj/crc-debug-6gtq2" Nov 26 02:27:49 crc kubenswrapper[4766]: I1126 02:27:49.428077 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk7kq\" (UniqueName: \"kubernetes.io/projected/ee36ddc9-aab0-4c21-bdb8-96ba138ad71f-kube-api-access-qk7kq\") pod \"crc-debug-6gtq2\" (UID: \"ee36ddc9-aab0-4c21-bdb8-96ba138ad71f\") " pod="openshift-must-gather-tv2fj/crc-debug-6gtq2" Nov 26 02:27:49 crc kubenswrapper[4766]: I1126 02:27:49.486050 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv2fj/crc-debug-6gtq2" Nov 26 02:27:49 crc kubenswrapper[4766]: I1126 02:27:49.763455 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv2fj/crc-debug-6gtq2" event={"ID":"ee36ddc9-aab0-4c21-bdb8-96ba138ad71f","Type":"ContainerStarted","Data":"5e16a476a7f6c197d1e26fd028b4d7ac42b37214e6967e76b94a489bfc7e77a4"} Nov 26 02:27:49 crc kubenswrapper[4766]: I1126 02:27:49.843023 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53360513-f8aa-41d6-94d2-fc348e4f2e16" path="/var/lib/kubelet/pods/53360513-f8aa-41d6-94d2-fc348e4f2e16/volumes" Nov 26 02:27:50 crc kubenswrapper[4766]: I1126 02:27:50.779499 4766 generic.go:334] "Generic (PLEG): container finished" podID="ee36ddc9-aab0-4c21-bdb8-96ba138ad71f" containerID="62a84cd8dcfadcb8288b1512069ccb0620c98789b17e318ef701fa58a25fd935" exitCode=0 Nov 26 02:27:50 crc kubenswrapper[4766]: I1126 02:27:50.779558 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv2fj/crc-debug-6gtq2" event={"ID":"ee36ddc9-aab0-4c21-bdb8-96ba138ad71f","Type":"ContainerDied","Data":"62a84cd8dcfadcb8288b1512069ccb0620c98789b17e318ef701fa58a25fd935"} Nov 26 02:27:51 crc kubenswrapper[4766]: I1126 02:27:51.919399 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv2fj/crc-debug-6gtq2" Nov 26 02:27:52 crc kubenswrapper[4766]: I1126 02:27:52.055355 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qk7kq\" (UniqueName: \"kubernetes.io/projected/ee36ddc9-aab0-4c21-bdb8-96ba138ad71f-kube-api-access-qk7kq\") pod \"ee36ddc9-aab0-4c21-bdb8-96ba138ad71f\" (UID: \"ee36ddc9-aab0-4c21-bdb8-96ba138ad71f\") " Nov 26 02:27:52 crc kubenswrapper[4766]: I1126 02:27:52.056022 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee36ddc9-aab0-4c21-bdb8-96ba138ad71f-host\") pod \"ee36ddc9-aab0-4c21-bdb8-96ba138ad71f\" (UID: \"ee36ddc9-aab0-4c21-bdb8-96ba138ad71f\") " Nov 26 02:27:52 crc kubenswrapper[4766]: I1126 02:27:52.056131 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee36ddc9-aab0-4c21-bdb8-96ba138ad71f-host" (OuterVolumeSpecName: "host") pod "ee36ddc9-aab0-4c21-bdb8-96ba138ad71f" (UID: "ee36ddc9-aab0-4c21-bdb8-96ba138ad71f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 02:27:52 crc kubenswrapper[4766]: I1126 02:27:52.056597 4766 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee36ddc9-aab0-4c21-bdb8-96ba138ad71f-host\") on node \"crc\" DevicePath \"\"" Nov 26 02:27:52 crc kubenswrapper[4766]: I1126 02:27:52.061939 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee36ddc9-aab0-4c21-bdb8-96ba138ad71f-kube-api-access-qk7kq" (OuterVolumeSpecName: "kube-api-access-qk7kq") pod "ee36ddc9-aab0-4c21-bdb8-96ba138ad71f" (UID: "ee36ddc9-aab0-4c21-bdb8-96ba138ad71f"). InnerVolumeSpecName "kube-api-access-qk7kq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:27:52 crc kubenswrapper[4766]: I1126 02:27:52.158113 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qk7kq\" (UniqueName: \"kubernetes.io/projected/ee36ddc9-aab0-4c21-bdb8-96ba138ad71f-kube-api-access-qk7kq\") on node \"crc\" DevicePath \"\"" Nov 26 02:27:52 crc kubenswrapper[4766]: I1126 02:27:52.805431 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv2fj/crc-debug-6gtq2" event={"ID":"ee36ddc9-aab0-4c21-bdb8-96ba138ad71f","Type":"ContainerDied","Data":"5e16a476a7f6c197d1e26fd028b4d7ac42b37214e6967e76b94a489bfc7e77a4"} Nov 26 02:27:52 crc kubenswrapper[4766]: I1126 02:27:52.805470 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e16a476a7f6c197d1e26fd028b4d7ac42b37214e6967e76b94a489bfc7e77a4" Nov 26 02:27:52 crc kubenswrapper[4766]: I1126 02:27:52.805506 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv2fj/crc-debug-6gtq2" Nov 26 02:27:53 crc kubenswrapper[4766]: I1126 02:27:53.369149 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tv2fj/crc-debug-6gtq2"] Nov 26 02:27:53 crc kubenswrapper[4766]: I1126 02:27:53.380241 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tv2fj/crc-debug-6gtq2"] Nov 26 02:27:53 crc kubenswrapper[4766]: I1126 02:27:53.841811 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee36ddc9-aab0-4c21-bdb8-96ba138ad71f" path="/var/lib/kubelet/pods/ee36ddc9-aab0-4c21-bdb8-96ba138ad71f/volumes" Nov 26 02:27:54 crc kubenswrapper[4766]: I1126 02:27:54.554384 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tv2fj/crc-debug-tx9k7"] Nov 26 02:27:54 crc kubenswrapper[4766]: E1126 02:27:54.555030 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee36ddc9-aab0-4c21-bdb8-96ba138ad71f" containerName="container-00" Nov 26 02:27:54 crc kubenswrapper[4766]: I1126 02:27:54.555041 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee36ddc9-aab0-4c21-bdb8-96ba138ad71f" containerName="container-00" Nov 26 02:27:54 crc kubenswrapper[4766]: I1126 02:27:54.555286 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee36ddc9-aab0-4c21-bdb8-96ba138ad71f" containerName="container-00" Nov 26 02:27:54 crc kubenswrapper[4766]: I1126 02:27:54.556051 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv2fj/crc-debug-tx9k7" Nov 26 02:27:54 crc kubenswrapper[4766]: I1126 02:27:54.622374 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qngmb\" (UniqueName: \"kubernetes.io/projected/228a1db5-b9d9-414e-a0a5-b9edeb35daa5-kube-api-access-qngmb\") pod \"crc-debug-tx9k7\" (UID: \"228a1db5-b9d9-414e-a0a5-b9edeb35daa5\") " pod="openshift-must-gather-tv2fj/crc-debug-tx9k7" Nov 26 02:27:54 crc kubenswrapper[4766]: I1126 02:27:54.622515 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/228a1db5-b9d9-414e-a0a5-b9edeb35daa5-host\") pod \"crc-debug-tx9k7\" (UID: \"228a1db5-b9d9-414e-a0a5-b9edeb35daa5\") " pod="openshift-must-gather-tv2fj/crc-debug-tx9k7" Nov 26 02:27:54 crc kubenswrapper[4766]: I1126 02:27:54.724554 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qngmb\" (UniqueName: \"kubernetes.io/projected/228a1db5-b9d9-414e-a0a5-b9edeb35daa5-kube-api-access-qngmb\") pod \"crc-debug-tx9k7\" (UID: \"228a1db5-b9d9-414e-a0a5-b9edeb35daa5\") " pod="openshift-must-gather-tv2fj/crc-debug-tx9k7" Nov 26 02:27:54 crc kubenswrapper[4766]: I1126 02:27:54.724628 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/228a1db5-b9d9-414e-a0a5-b9edeb35daa5-host\") pod \"crc-debug-tx9k7\" (UID: \"228a1db5-b9d9-414e-a0a5-b9edeb35daa5\") " pod="openshift-must-gather-tv2fj/crc-debug-tx9k7" Nov 26 02:27:54 crc kubenswrapper[4766]: I1126 02:27:54.724831 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/228a1db5-b9d9-414e-a0a5-b9edeb35daa5-host\") pod \"crc-debug-tx9k7\" (UID: \"228a1db5-b9d9-414e-a0a5-b9edeb35daa5\") " pod="openshift-must-gather-tv2fj/crc-debug-tx9k7" Nov 26 02:27:54 crc kubenswrapper[4766]: I1126 02:27:54.746014 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qngmb\" (UniqueName: \"kubernetes.io/projected/228a1db5-b9d9-414e-a0a5-b9edeb35daa5-kube-api-access-qngmb\") pod \"crc-debug-tx9k7\" (UID: \"228a1db5-b9d9-414e-a0a5-b9edeb35daa5\") " pod="openshift-must-gather-tv2fj/crc-debug-tx9k7" Nov 26 02:27:54 crc kubenswrapper[4766]: I1126 02:27:54.872452 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv2fj/crc-debug-tx9k7" Nov 26 02:27:55 crc kubenswrapper[4766]: I1126 02:27:55.842561 4766 generic.go:334] "Generic (PLEG): container finished" podID="228a1db5-b9d9-414e-a0a5-b9edeb35daa5" containerID="bdea260d06728ecfcd8f5bb69677ad934038f79aadbf507303a1efd28381b3af" exitCode=0 Nov 26 02:27:55 crc kubenswrapper[4766]: I1126 02:27:55.842761 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv2fj/crc-debug-tx9k7" event={"ID":"228a1db5-b9d9-414e-a0a5-b9edeb35daa5","Type":"ContainerDied","Data":"bdea260d06728ecfcd8f5bb69677ad934038f79aadbf507303a1efd28381b3af"} Nov 26 02:27:55 crc kubenswrapper[4766]: I1126 02:27:55.842939 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv2fj/crc-debug-tx9k7" event={"ID":"228a1db5-b9d9-414e-a0a5-b9edeb35daa5","Type":"ContainerStarted","Data":"d5cd5580555aeeb157ed20819483838065f8af0076389705bdc210c4a0667843"} Nov 26 02:27:55 crc kubenswrapper[4766]: I1126 02:27:55.892867 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tv2fj/crc-debug-tx9k7"] Nov 26 02:27:55 crc kubenswrapper[4766]: I1126 02:27:55.903351 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tv2fj/crc-debug-tx9k7"] Nov 26 02:27:56 crc kubenswrapper[4766]: I1126 02:27:56.966792 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv2fj/crc-debug-tx9k7" Nov 26 02:27:56 crc kubenswrapper[4766]: I1126 02:27:56.977285 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/228a1db5-b9d9-414e-a0a5-b9edeb35daa5-host\") pod \"228a1db5-b9d9-414e-a0a5-b9edeb35daa5\" (UID: \"228a1db5-b9d9-414e-a0a5-b9edeb35daa5\") " Nov 26 02:27:56 crc kubenswrapper[4766]: I1126 02:27:56.977394 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/228a1db5-b9d9-414e-a0a5-b9edeb35daa5-host" (OuterVolumeSpecName: "host") pod "228a1db5-b9d9-414e-a0a5-b9edeb35daa5" (UID: "228a1db5-b9d9-414e-a0a5-b9edeb35daa5"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 02:27:56 crc kubenswrapper[4766]: I1126 02:27:56.977458 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qngmb\" (UniqueName: \"kubernetes.io/projected/228a1db5-b9d9-414e-a0a5-b9edeb35daa5-kube-api-access-qngmb\") pod \"228a1db5-b9d9-414e-a0a5-b9edeb35daa5\" (UID: \"228a1db5-b9d9-414e-a0a5-b9edeb35daa5\") " Nov 26 02:27:56 crc kubenswrapper[4766]: I1126 02:27:56.979531 4766 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/228a1db5-b9d9-414e-a0a5-b9edeb35daa5-host\") on node \"crc\" DevicePath \"\"" Nov 26 02:27:56 crc kubenswrapper[4766]: I1126 02:27:56.986957 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/228a1db5-b9d9-414e-a0a5-b9edeb35daa5-kube-api-access-qngmb" (OuterVolumeSpecName: "kube-api-access-qngmb") pod "228a1db5-b9d9-414e-a0a5-b9edeb35daa5" (UID: "228a1db5-b9d9-414e-a0a5-b9edeb35daa5"). InnerVolumeSpecName "kube-api-access-qngmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:27:57 crc kubenswrapper[4766]: I1126 02:27:57.082315 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qngmb\" (UniqueName: \"kubernetes.io/projected/228a1db5-b9d9-414e-a0a5-b9edeb35daa5-kube-api-access-qngmb\") on node \"crc\" DevicePath \"\"" Nov 26 02:27:57 crc kubenswrapper[4766]: I1126 02:27:57.853975 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="228a1db5-b9d9-414e-a0a5-b9edeb35daa5" path="/var/lib/kubelet/pods/228a1db5-b9d9-414e-a0a5-b9edeb35daa5/volumes" Nov 26 02:27:57 crc kubenswrapper[4766]: I1126 02:27:57.872148 4766 scope.go:117] "RemoveContainer" containerID="bdea260d06728ecfcd8f5bb69677ad934038f79aadbf507303a1efd28381b3af" Nov 26 02:27:57 crc kubenswrapper[4766]: I1126 02:27:57.872380 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv2fj/crc-debug-tx9k7" Nov 26 02:28:57 crc kubenswrapper[4766]: I1126 02:28:57.150031 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_54262828-5166-42d6-934e-5eab8683ecbd/aodh-api/0.log" Nov 26 02:28:57 crc kubenswrapper[4766]: I1126 02:28:57.378813 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4snm6"] Nov 26 02:28:57 crc kubenswrapper[4766]: E1126 02:28:57.379362 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="228a1db5-b9d9-414e-a0a5-b9edeb35daa5" containerName="container-00" Nov 26 02:28:57 crc kubenswrapper[4766]: I1126 02:28:57.379379 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="228a1db5-b9d9-414e-a0a5-b9edeb35daa5" containerName="container-00" Nov 26 02:28:57 crc kubenswrapper[4766]: I1126 02:28:57.379598 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="228a1db5-b9d9-414e-a0a5-b9edeb35daa5" containerName="container-00" Nov 26 02:28:57 crc kubenswrapper[4766]: I1126 02:28:57.381275 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4snm6" Nov 26 02:28:57 crc kubenswrapper[4766]: I1126 02:28:57.389641 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_54262828-5166-42d6-934e-5eab8683ecbd/aodh-listener/0.log" Nov 26 02:28:57 crc kubenswrapper[4766]: I1126 02:28:57.389739 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4snm6"] Nov 26 02:28:57 crc kubenswrapper[4766]: I1126 02:28:57.436631 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_54262828-5166-42d6-934e-5eab8683ecbd/aodh-evaluator/0.log" Nov 26 02:28:57 crc kubenswrapper[4766]: I1126 02:28:57.462365 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_54262828-5166-42d6-934e-5eab8683ecbd/aodh-notifier/0.log" Nov 26 02:28:57 crc kubenswrapper[4766]: I1126 02:28:57.473911 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1064657-b06b-4370-928e-5972ff8b46ae-utilities\") pod \"redhat-marketplace-4snm6\" (UID: \"a1064657-b06b-4370-928e-5972ff8b46ae\") " pod="openshift-marketplace/redhat-marketplace-4snm6" Nov 26 02:28:57 crc kubenswrapper[4766]: I1126 02:28:57.473961 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1064657-b06b-4370-928e-5972ff8b46ae-catalog-content\") pod \"redhat-marketplace-4snm6\" (UID: \"a1064657-b06b-4370-928e-5972ff8b46ae\") " pod="openshift-marketplace/redhat-marketplace-4snm6" Nov 26 02:28:57 crc kubenswrapper[4766]: I1126 02:28:57.474122 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t44w\" (UniqueName: \"kubernetes.io/projected/a1064657-b06b-4370-928e-5972ff8b46ae-kube-api-access-6t44w\") pod \"redhat-marketplace-4snm6\" (UID: \"a1064657-b06b-4370-928e-5972ff8b46ae\") " pod="openshift-marketplace/redhat-marketplace-4snm6" Nov 26 02:28:57 crc kubenswrapper[4766]: I1126 02:28:57.576246 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t44w\" (UniqueName: \"kubernetes.io/projected/a1064657-b06b-4370-928e-5972ff8b46ae-kube-api-access-6t44w\") pod \"redhat-marketplace-4snm6\" (UID: \"a1064657-b06b-4370-928e-5972ff8b46ae\") " pod="openshift-marketplace/redhat-marketplace-4snm6" Nov 26 02:28:57 crc kubenswrapper[4766]: I1126 02:28:57.576361 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1064657-b06b-4370-928e-5972ff8b46ae-utilities\") pod \"redhat-marketplace-4snm6\" (UID: \"a1064657-b06b-4370-928e-5972ff8b46ae\") " pod="openshift-marketplace/redhat-marketplace-4snm6" Nov 26 02:28:57 crc kubenswrapper[4766]: I1126 02:28:57.576391 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1064657-b06b-4370-928e-5972ff8b46ae-catalog-content\") pod \"redhat-marketplace-4snm6\" (UID: \"a1064657-b06b-4370-928e-5972ff8b46ae\") " pod="openshift-marketplace/redhat-marketplace-4snm6" Nov 26 02:28:57 crc kubenswrapper[4766]: I1126 02:28:57.576943 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1064657-b06b-4370-928e-5972ff8b46ae-catalog-content\") pod \"redhat-marketplace-4snm6\" (UID: \"a1064657-b06b-4370-928e-5972ff8b46ae\") " pod="openshift-marketplace/redhat-marketplace-4snm6" Nov 26 02:28:57 crc kubenswrapper[4766]: I1126 02:28:57.577013 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1064657-b06b-4370-928e-5972ff8b46ae-utilities\") pod \"redhat-marketplace-4snm6\" (UID: \"a1064657-b06b-4370-928e-5972ff8b46ae\") " pod="openshift-marketplace/redhat-marketplace-4snm6" Nov 26 02:28:57 crc kubenswrapper[4766]: I1126 02:28:57.596385 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t44w\" (UniqueName: \"kubernetes.io/projected/a1064657-b06b-4370-928e-5972ff8b46ae-kube-api-access-6t44w\") pod \"redhat-marketplace-4snm6\" (UID: \"a1064657-b06b-4370-928e-5972ff8b46ae\") " pod="openshift-marketplace/redhat-marketplace-4snm6" Nov 26 02:28:57 crc kubenswrapper[4766]: I1126 02:28:57.633095 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6b7ccbcb5b-rf9v9_5655d266-3ea5-404f-901e-80da609b0319/barbican-api-log/0.log" Nov 26 02:28:57 crc kubenswrapper[4766]: I1126 02:28:57.742135 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4snm6" Nov 26 02:28:57 crc kubenswrapper[4766]: I1126 02:28:57.913084 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7fd5c59d54-6ccpg_81836728-9333-4234-8175-954682305e4e/barbican-keystone-listener/0.log" Nov 26 02:28:58 crc kubenswrapper[4766]: I1126 02:28:58.072419 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6b7ccbcb5b-rf9v9_5655d266-3ea5-404f-901e-80da609b0319/barbican-api/0.log" Nov 26 02:28:58 crc kubenswrapper[4766]: I1126 02:28:58.294212 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-68b9476655-7svbz_01cfcf71-49d4-498f-a9cc-5e9c65c08709/barbican-worker/0.log" Nov 26 02:28:58 crc kubenswrapper[4766]: I1126 02:28:58.351573 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7fd5c59d54-6ccpg_81836728-9333-4234-8175-954682305e4e/barbican-keystone-listener-log/0.log" Nov 26 02:28:58 crc kubenswrapper[4766]: I1126 02:28:58.381254 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4snm6"] Nov 26 02:28:58 crc kubenswrapper[4766]: I1126 02:28:58.483115 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-68b9476655-7svbz_01cfcf71-49d4-498f-a9cc-5e9c65c08709/barbican-worker-log/0.log" Nov 26 02:28:58 crc kubenswrapper[4766]: I1126 02:28:58.673153 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4snm6" event={"ID":"a1064657-b06b-4370-928e-5972ff8b46ae","Type":"ContainerStarted","Data":"de8fa1c2126280db25301dfe7a944455ac09bdd9c065e48ff1f0386146a06a9d"} Nov 26 02:28:58 crc kubenswrapper[4766]: I1126 02:28:58.673190 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4snm6" event={"ID":"a1064657-b06b-4370-928e-5972ff8b46ae","Type":"ContainerStarted","Data":"2dab757e83732522486b228b30eaff79cab6bbe49b1c2a9cc9374b94a88554f5"} Nov 26 02:28:58 crc kubenswrapper[4766]: I1126 02:28:58.674335 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl_973efc5a-bdcf-43af-baf5-11ae375d3494/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:28:58 crc kubenswrapper[4766]: I1126 02:28:58.813857 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2c82c6d8-c395-4177-930f-f840d902df28/ceilometer-central-agent/0.log" Nov 26 02:28:58 crc kubenswrapper[4766]: I1126 02:28:58.902319 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2c82c6d8-c395-4177-930f-f840d902df28/proxy-httpd/0.log" Nov 26 02:28:58 crc kubenswrapper[4766]: I1126 02:28:58.907986 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2c82c6d8-c395-4177-930f-f840d902df28/ceilometer-notification-agent/0.log" Nov 26 02:28:59 crc kubenswrapper[4766]: I1126 02:28:59.050048 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2c82c6d8-c395-4177-930f-f840d902df28/sg-core/0.log" Nov 26 02:28:59 crc kubenswrapper[4766]: I1126 02:28:59.092223 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl_91756acc-c3e0-471b-b79d-4eb1cf2a80df/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:28:59 crc kubenswrapper[4766]: I1126 02:28:59.287432 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5_9f63fa32-bcb4-490f-947b-ec7ab639505c/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:28:59 crc kubenswrapper[4766]: I1126 02:28:59.391701 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_d4e1503f-0e08-4e3a-a578-fcbd3b51989a/cinder-api-log/0.log" Nov 26 02:28:59 crc kubenswrapper[4766]: I1126 02:28:59.463244 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_d4e1503f-0e08-4e3a-a578-fcbd3b51989a/cinder-api/0.log" Nov 26 02:28:59 crc kubenswrapper[4766]: I1126 02:28:59.681388 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_2c89caf0-3bca-4172-9a9b-27657d46a1c5/cinder-backup/0.log" Nov 26 02:28:59 crc kubenswrapper[4766]: I1126 02:28:59.682712 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_2c89caf0-3bca-4172-9a9b-27657d46a1c5/probe/0.log" Nov 26 02:28:59 crc kubenswrapper[4766]: I1126 02:28:59.685177 4766 generic.go:334] "Generic (PLEG): container finished" podID="a1064657-b06b-4370-928e-5972ff8b46ae" containerID="de8fa1c2126280db25301dfe7a944455ac09bdd9c065e48ff1f0386146a06a9d" exitCode=0 Nov 26 02:28:59 crc kubenswrapper[4766]: I1126 02:28:59.685216 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4snm6" event={"ID":"a1064657-b06b-4370-928e-5972ff8b46ae","Type":"ContainerDied","Data":"de8fa1c2126280db25301dfe7a944455ac09bdd9c065e48ff1f0386146a06a9d"} Nov 26 02:28:59 crc kubenswrapper[4766]: I1126 02:28:59.807529 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_92d51b16-e82c-4fec-a095-53e3f0608fad/cinder-scheduler/0.log" Nov 26 02:29:00 crc kubenswrapper[4766]: I1126 02:29:00.001923 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_92d51b16-e82c-4fec-a095-53e3f0608fad/probe/0.log" Nov 26 02:29:00 crc kubenswrapper[4766]: I1126 02:29:00.045648 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_5687a3ac-6850-4533-9c40-b0954d98c628/cinder-volume/0.log" Nov 26 02:29:00 crc kubenswrapper[4766]: I1126 02:29:00.089629 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_5687a3ac-6850-4533-9c40-b0954d98c628/probe/0.log" Nov 26 02:29:00 crc kubenswrapper[4766]: I1126 02:29:00.217920 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-4bmds_a7ae4f62-184e-4093-9a0a-12d6188ee301/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:29:00 crc kubenswrapper[4766]: I1126 02:29:00.315632 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-9t59g_511ab0e6-b14c-4249-92e9-184b57a03147/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:29:00 crc kubenswrapper[4766]: I1126 02:29:00.462852 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-c8d8d886c-9lgvr_316b23d8-2140-4ac5-8e26-d63171814293/init/0.log" Nov 26 02:29:00 crc kubenswrapper[4766]: I1126 02:29:00.696363 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4snm6" event={"ID":"a1064657-b06b-4370-928e-5972ff8b46ae","Type":"ContainerStarted","Data":"bf36d21bce57c1db501d2c69a3916d0708757f005dcb0ce7edc76cb3c42840fa"} Nov 26 02:29:00 crc kubenswrapper[4766]: I1126 02:29:00.745259 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-c8d8d886c-9lgvr_316b23d8-2140-4ac5-8e26-d63171814293/init/0.log" Nov 26 02:29:00 crc kubenswrapper[4766]: I1126 02:29:00.811594 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_5ea31542-fdc7-481a-853a-5935e39fca92/glance-httpd/0.log" Nov 26 02:29:00 crc kubenswrapper[4766]: I1126 02:29:00.830618 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-c8d8d886c-9lgvr_316b23d8-2140-4ac5-8e26-d63171814293/dnsmasq-dns/0.log" Nov 26 02:29:00 crc kubenswrapper[4766]: I1126 02:29:00.973000 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_5ea31542-fdc7-481a-853a-5935e39fca92/glance-log/0.log" Nov 26 02:29:01 crc kubenswrapper[4766]: I1126 02:29:01.106671 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f5f6c81c-04d9-4542-9ab1-b932737f388c/glance-httpd/0.log" Nov 26 02:29:01 crc kubenswrapper[4766]: I1126 02:29:01.352063 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f5f6c81c-04d9-4542-9ab1-b932737f388c/glance-log/0.log" Nov 26 02:29:01 crc kubenswrapper[4766]: I1126 02:29:01.712520 4766 generic.go:334] "Generic (PLEG): container finished" podID="a1064657-b06b-4370-928e-5972ff8b46ae" containerID="bf36d21bce57c1db501d2c69a3916d0708757f005dcb0ce7edc76cb3c42840fa" exitCode=0 Nov 26 02:29:01 crc kubenswrapper[4766]: I1126 02:29:01.712811 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4snm6" event={"ID":"a1064657-b06b-4370-928e-5972ff8b46ae","Type":"ContainerDied","Data":"bf36d21bce57c1db501d2c69a3916d0708757f005dcb0ce7edc76cb3c42840fa"} Nov 26 02:29:01 crc kubenswrapper[4766]: I1126 02:29:01.746966 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-6ff8c446-6wzh8_775687b1-1657-4115-b482-f7e418e77727/heat-engine/0.log" Nov 26 02:29:02 crc kubenswrapper[4766]: I1126 02:29:02.021118 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-cf46f664-8k6s8_778430ca-2e98-4156-b1d9-fdf05b76ef7e/horizon/0.log" Nov 26 02:29:02 crc kubenswrapper[4766]: I1126 02:29:02.373284 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-l279n_7509ec1a-b269-4bbe-bfca-0dcb052360ea/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:29:02 crc kubenswrapper[4766]: I1126 02:29:02.595003 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-hvdtg_5d516c45-cb2e-4920-8e14-3a47063212a1/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:29:02 crc kubenswrapper[4766]: I1126 02:29:02.729038 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4snm6" event={"ID":"a1064657-b06b-4370-928e-5972ff8b46ae","Type":"ContainerStarted","Data":"1b129982706444f72be0a69c788ebe8be308f1f177151a33e61fb8059a8dc993"} Nov 26 02:29:02 crc kubenswrapper[4766]: I1126 02:29:02.756999 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4snm6" podStartSLOduration=3.32526046 podStartE2EDuration="5.756981594s" podCreationTimestamp="2025-11-26 02:28:57 +0000 UTC" firstStartedPulling="2025-11-26 02:28:59.687986039 +0000 UTC m=+7520.536756469" lastFinishedPulling="2025-11-26 02:29:02.119707173 +0000 UTC m=+7522.968477603" observedRunningTime="2025-11-26 02:29:02.75191071 +0000 UTC m=+7523.600681140" watchObservedRunningTime="2025-11-26 02:29:02.756981594 +0000 UTC m=+7523.605752024" Nov 26 02:29:02 crc kubenswrapper[4766]: I1126 02:29:02.791641 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-cf46f664-8k6s8_778430ca-2e98-4156-b1d9-fdf05b76ef7e/horizon-log/0.log" Nov 26 02:29:03 crc kubenswrapper[4766]: I1126 02:29:03.031110 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-5ff994f754-l2ctn_5f65a888-def8-495b-915a-0b6556aff6de/heat-cfnapi/0.log" Nov 26 02:29:03 crc kubenswrapper[4766]: I1126 02:29:03.129807 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29401981-22f5z_5e25beeb-e89d-4aed-9353-3755a0ef2efb/keystone-cron/0.log" Nov 26 02:29:03 crc kubenswrapper[4766]: I1126 02:29:03.133998 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-7f6d8f97f4-szt84_65159db0-64b4-4cce-ba5e-aa6f889f7f58/heat-api/0.log" Nov 26 02:29:03 crc kubenswrapper[4766]: I1126 02:29:03.296584 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29402041-bjg4c_bd82c837-925c-4835-a90f-3d5a873dfbae/keystone-cron/0.log" Nov 26 02:29:03 crc kubenswrapper[4766]: I1126 02:29:03.373540 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6768b6c6f7-lm9pw_3fdd0908-10b2-437d-92bf-ccdc345a0d80/keystone-api/0.log" Nov 26 02:29:03 crc kubenswrapper[4766]: I1126 02:29:03.415545 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a700ac64-3ee9-4e7b-8da8-1cb7537fee78/kube-state-metrics/0.log" Nov 26 02:29:03 crc kubenswrapper[4766]: I1126 02:29:03.553997 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-nswmp_64255763-3ccc-42d2-88ae-3c48ecea8c07/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:29:03 crc kubenswrapper[4766]: I1126 02:29:03.661006 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_logging-edpm-deployment-openstack-edpm-ipam-25nr7_85810b6d-fde1-48a2-b55d-b6c9c72d6b6a/logging-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:29:03 crc kubenswrapper[4766]: I1126 02:29:03.809114 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_8cb6838c-a086-4990-82f0-a3c36675be67/manila-api/0.log" Nov 26 02:29:03 crc kubenswrapper[4766]: I1126 02:29:03.852890 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_8cb6838c-a086-4990-82f0-a3c36675be67/manila-api-log/0.log" Nov 26 02:29:03 crc kubenswrapper[4766]: I1126 02:29:03.906425 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_a487b55f-cdaa-4351-95ed-7fb4a8ce382d/probe/0.log" Nov 26 02:29:04 crc kubenswrapper[4766]: I1126 02:29:04.044941 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_a487b55f-cdaa-4351-95ed-7fb4a8ce382d/manila-scheduler/0.log" Nov 26 02:29:04 crc kubenswrapper[4766]: I1126 02:29:04.102727 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_91ee515c-0d11-436f-b0c9-184f065d52eb/manila-share/0.log" Nov 26 02:29:04 crc kubenswrapper[4766]: I1126 02:29:04.109639 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_91ee515c-0d11-436f-b0c9-184f065d52eb/probe/0.log" Nov 26 02:29:04 crc kubenswrapper[4766]: I1126 02:29:04.534004 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mysqld-exporter-0_51bb050a-0106-4f7c-acf2-8c6c19f93d39/mysqld-exporter/0.log" Nov 26 02:29:04 crc kubenswrapper[4766]: I1126 02:29:04.851013 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h_3efd579d-7c56-4624-a920-1e40cea0b951/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:29:04 crc kubenswrapper[4766]: I1126 02:29:04.888640 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-756fcfc485-ktbfm_8d171003-64ac-4108-957b-70f4225359c5/neutron-api/0.log" Nov 26 02:29:04 crc kubenswrapper[4766]: I1126 02:29:04.917819 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-756fcfc485-ktbfm_8d171003-64ac-4108-957b-70f4225359c5/neutron-httpd/0.log" Nov 26 02:29:05 crc kubenswrapper[4766]: I1126 02:29:05.435731 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_b33aa993-e31f-402f-9574-33bd30a942c4/nova-cell0-conductor-conductor/0.log" Nov 26 02:29:05 crc kubenswrapper[4766]: I1126 02:29:05.797228 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a93e0b93-3cde-4098-a853-ea1b3c86c323/nova-api-log/0.log" Nov 26 02:29:05 crc kubenswrapper[4766]: I1126 02:29:05.807746 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_f2ec2417-bd6f-4166-b42c-407424769274/nova-cell1-conductor-conductor/0.log" Nov 26 02:29:06 crc kubenswrapper[4766]: I1126 02:29:06.083274 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964_8da0c4a5-4c26-46de-8408-a8b562ba8f76/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:29:06 crc kubenswrapper[4766]: I1126 02:29:06.171900 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_4cf2f787-c351-49ce-a347-dc0975ea8bc4/nova-cell1-novncproxy-novncproxy/0.log" Nov 26 02:29:06 crc kubenswrapper[4766]: I1126 02:29:06.367231 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1f172da7-f357-44b5-a4c0-5a2cc493f906/nova-metadata-log/0.log" Nov 26 02:29:06 crc kubenswrapper[4766]: I1126 02:29:06.599387 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a93e0b93-3cde-4098-a853-ea1b3c86c323/nova-api-api/0.log" Nov 26 02:29:06 crc kubenswrapper[4766]: I1126 02:29:06.795133 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_64dcd71b-3e04-470c-ab99-37105bd225dc/nova-scheduler-scheduler/0.log" Nov 26 02:29:06 crc kubenswrapper[4766]: I1126 02:29:06.842212 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_44f36c8b-51b8-4bce-afbb-1962eb317dec/mysql-bootstrap/0.log" Nov 26 02:29:07 crc kubenswrapper[4766]: I1126 02:29:07.006406 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_44f36c8b-51b8-4bce-afbb-1962eb317dec/galera/0.log" Nov 26 02:29:07 crc kubenswrapper[4766]: I1126 02:29:07.027533 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_44f36c8b-51b8-4bce-afbb-1962eb317dec/mysql-bootstrap/0.log" Nov 26 02:29:07 crc kubenswrapper[4766]: I1126 02:29:07.208172 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_54c97191-0bfc-4a65-8168-10833aab2814/mysql-bootstrap/0.log" Nov 26 02:29:07 crc kubenswrapper[4766]: I1126 02:29:07.410910 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_54c97191-0bfc-4a65-8168-10833aab2814/mysql-bootstrap/0.log" Nov 26 02:29:07 crc kubenswrapper[4766]: I1126 02:29:07.479315 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_54c97191-0bfc-4a65-8168-10833aab2814/galera/0.log" Nov 26 02:29:07 crc kubenswrapper[4766]: I1126 02:29:07.742919 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4snm6" Nov 26 02:29:07 crc kubenswrapper[4766]: I1126 02:29:07.744261 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4snm6" Nov 26 02:29:07 crc kubenswrapper[4766]: I1126 02:29:07.814423 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4snm6" Nov 26 02:29:07 crc kubenswrapper[4766]: I1126 02:29:07.848260 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_d8551b13-7602-4c7c-bfb5-0bcf7e64954d/openstackclient/0.log" Nov 26 02:29:07 crc kubenswrapper[4766]: I1126 02:29:07.911118 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-qnhgk_342b8104-66eb-4fe1-88ad-ef5afc617d65/openstack-network-exporter/0.log" Nov 26 02:29:08 crc kubenswrapper[4766]: I1126 02:29:08.107083 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6ghtc_71963f95-01c9-473a-be8a-7798f6038ffb/ovsdb-server-init/0.log" Nov 26 02:29:08 crc kubenswrapper[4766]: I1126 02:29:08.382581 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6ghtc_71963f95-01c9-473a-be8a-7798f6038ffb/ovsdb-server-init/0.log" Nov 26 02:29:08 crc kubenswrapper[4766]: I1126 02:29:08.413254 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6ghtc_71963f95-01c9-473a-be8a-7798f6038ffb/ovsdb-server/0.log" Nov 26 02:29:08 crc kubenswrapper[4766]: I1126 02:29:08.434285 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6ghtc_71963f95-01c9-473a-be8a-7798f6038ffb/ovs-vswitchd/0.log" Nov 26 02:29:08 crc kubenswrapper[4766]: I1126 02:29:08.625616 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-p42tr_22168318-cef8-4363-8b3e-3042e82037ea/ovn-controller/0.log" Nov 26 02:29:08 crc kubenswrapper[4766]: I1126 02:29:08.883405 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4snm6" Nov 26 02:29:08 crc kubenswrapper[4766]: I1126 02:29:08.893149 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-8xf9z_45de00ed-b796-4ac1-8e38-ee601eab6dc0/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:29:09 crc kubenswrapper[4766]: I1126 02:29:09.046823 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_64b18a9d-088b-4a4a-9c1a-4b63cbb1d642/openstack-network-exporter/0.log" Nov 26 02:29:09 crc kubenswrapper[4766]: I1126 02:29:09.088554 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_64b18a9d-088b-4a4a-9c1a-4b63cbb1d642/ovn-northd/0.log" Nov 26 02:29:09 crc kubenswrapper[4766]: I1126 02:29:09.261825 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1f172da7-f357-44b5-a4c0-5a2cc493f906/nova-metadata-metadata/0.log" Nov 26 02:29:09 crc kubenswrapper[4766]: I1126 02:29:09.269584 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d346209c-ec54-4719-9b56-7a079c603ebc/openstack-network-exporter/0.log" Nov 26 02:29:09 crc kubenswrapper[4766]: I1126 02:29:09.307315 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d346209c-ec54-4719-9b56-7a079c603ebc/ovsdbserver-nb/0.log" Nov 26 02:29:09 crc kubenswrapper[4766]: I1126 02:29:09.508370 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_954395e4-178d-46ef-a695-8895ebe444d8/openstack-network-exporter/0.log" Nov 26 02:29:09 crc kubenswrapper[4766]: I1126 02:29:09.546278 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_954395e4-178d-46ef-a695-8895ebe444d8/ovsdbserver-sb/0.log" Nov 26 02:29:09 crc kubenswrapper[4766]: I1126 02:29:09.793643 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0c404609-2299-4fe5-8c43-da5614df251e/init-config-reloader/0.log" Nov 26 02:29:09 crc kubenswrapper[4766]: I1126 02:29:09.949193 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5fd6484f46-cqsxg_be4073df-dee9-4a70-8911-b23a317a0a57/placement-log/0.log" Nov 26 02:29:09 crc kubenswrapper[4766]: I1126 02:29:09.972787 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5fd6484f46-cqsxg_be4073df-dee9-4a70-8911-b23a317a0a57/placement-api/0.log" Nov 26 02:29:10 crc kubenswrapper[4766]: I1126 02:29:10.045471 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0c404609-2299-4fe5-8c43-da5614df251e/init-config-reloader/0.log" Nov 26 02:29:10 crc kubenswrapper[4766]: I1126 02:29:10.046444 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0c404609-2299-4fe5-8c43-da5614df251e/config-reloader/0.log" Nov 26 02:29:10 crc kubenswrapper[4766]: I1126 02:29:10.163323 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0c404609-2299-4fe5-8c43-da5614df251e/thanos-sidecar/0.log" Nov 26 02:29:10 crc kubenswrapper[4766]: I1126 02:29:10.225397 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0c404609-2299-4fe5-8c43-da5614df251e/prometheus/0.log" Nov 26 02:29:10 crc kubenswrapper[4766]: I1126 02:29:10.284757 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_64400ff7-9c3a-42c3-b1fc-2f9696f1206c/setup-container/0.log" Nov 26 02:29:10 crc kubenswrapper[4766]: I1126 02:29:10.464848 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4snm6"] Nov 26 02:29:10 crc kubenswrapper[4766]: I1126 02:29:10.488962 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_64400ff7-9c3a-42c3-b1fc-2f9696f1206c/setup-container/0.log" Nov 26 02:29:10 crc kubenswrapper[4766]: I1126 02:29:10.572553 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_64400ff7-9c3a-42c3-b1fc-2f9696f1206c/rabbitmq/0.log" Nov 26 02:29:10 crc kubenswrapper[4766]: I1126 02:29:10.616744 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_797e86ef-8e0d-499e-859f-6f6732826d7e/setup-container/0.log" Nov 26 02:29:10 crc kubenswrapper[4766]: I1126 02:29:10.814029 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_797e86ef-8e0d-499e-859f-6f6732826d7e/setup-container/0.log" Nov 26 02:29:10 crc kubenswrapper[4766]: I1126 02:29:10.832174 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4snm6" podUID="a1064657-b06b-4370-928e-5972ff8b46ae" containerName="registry-server" containerID="cri-o://1b129982706444f72be0a69c788ebe8be308f1f177151a33e61fb8059a8dc993" gracePeriod=2 Nov 26 02:29:10 crc kubenswrapper[4766]: I1126 02:29:10.907433 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj_b59fb3c0-2626-47c2-b2f4-f13c13b011a5/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:29:10 crc kubenswrapper[4766]: I1126 02:29:10.935179 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_797e86ef-8e0d-499e-859f-6f6732826d7e/rabbitmq/0.log" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.179474 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl_7ce6b646-f313-478b-b425-bc3b56504c3f/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.255146 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-tzd58_14ea5f7a-f044-426a-919e-a05cdcfe900c/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.427312 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-qlskl_57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa/ssh-known-hosts-edpm-deployment/0.log" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.479036 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.479090 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.490106 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4snm6" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.620899 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6t44w\" (UniqueName: \"kubernetes.io/projected/a1064657-b06b-4370-928e-5972ff8b46ae-kube-api-access-6t44w\") pod \"a1064657-b06b-4370-928e-5972ff8b46ae\" (UID: \"a1064657-b06b-4370-928e-5972ff8b46ae\") " Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.620960 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1064657-b06b-4370-928e-5972ff8b46ae-catalog-content\") pod \"a1064657-b06b-4370-928e-5972ff8b46ae\" (UID: \"a1064657-b06b-4370-928e-5972ff8b46ae\") " Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.621047 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1064657-b06b-4370-928e-5972ff8b46ae-utilities\") pod \"a1064657-b06b-4370-928e-5972ff8b46ae\" (UID: \"a1064657-b06b-4370-928e-5972ff8b46ae\") " Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.622288 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1064657-b06b-4370-928e-5972ff8b46ae-utilities" (OuterVolumeSpecName: "utilities") pod "a1064657-b06b-4370-928e-5972ff8b46ae" (UID: "a1064657-b06b-4370-928e-5972ff8b46ae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.630569 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1064657-b06b-4370-928e-5972ff8b46ae-kube-api-access-6t44w" (OuterVolumeSpecName: "kube-api-access-6t44w") pod "a1064657-b06b-4370-928e-5972ff8b46ae" (UID: "a1064657-b06b-4370-928e-5972ff8b46ae"). InnerVolumeSpecName "kube-api-access-6t44w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.648536 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1064657-b06b-4370-928e-5972ff8b46ae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a1064657-b06b-4370-928e-5972ff8b46ae" (UID: "a1064657-b06b-4370-928e-5972ff8b46ae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.678132 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-74d678b987-q89lp_3d4e6997-4bbc-4cb2-a0f2-abf3b940090c/proxy-server/0.log" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.723321 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6t44w\" (UniqueName: \"kubernetes.io/projected/a1064657-b06b-4370-928e-5972ff8b46ae-kube-api-access-6t44w\") on node \"crc\" DevicePath \"\"" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.723359 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1064657-b06b-4370-928e-5972ff8b46ae-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.723374 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1064657-b06b-4370-928e-5972ff8b46ae-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.805257 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-tmfzn_6df26511-75fd-4bcb-ab3e-65cde68c9f46/swift-ring-rebalance/0.log" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.844989 4766 generic.go:334] "Generic (PLEG): container finished" podID="a1064657-b06b-4370-928e-5972ff8b46ae" containerID="1b129982706444f72be0a69c788ebe8be308f1f177151a33e61fb8059a8dc993" exitCode=0 Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.845233 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4snm6" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.845311 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4snm6" event={"ID":"a1064657-b06b-4370-928e-5972ff8b46ae","Type":"ContainerDied","Data":"1b129982706444f72be0a69c788ebe8be308f1f177151a33e61fb8059a8dc993"} Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.845387 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4snm6" event={"ID":"a1064657-b06b-4370-928e-5972ff8b46ae","Type":"ContainerDied","Data":"2dab757e83732522486b228b30eaff79cab6bbe49b1c2a9cc9374b94a88554f5"} Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.845465 4766 scope.go:117] "RemoveContainer" containerID="1b129982706444f72be0a69c788ebe8be308f1f177151a33e61fb8059a8dc993" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.848461 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-74d678b987-q89lp_3d4e6997-4bbc-4cb2-a0f2-abf3b940090c/proxy-httpd/0.log" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.868515 4766 scope.go:117] "RemoveContainer" containerID="bf36d21bce57c1db501d2c69a3916d0708757f005dcb0ce7edc76cb3c42840fa" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.893586 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4snm6"] Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.896048 4766 scope.go:117] "RemoveContainer" containerID="de8fa1c2126280db25301dfe7a944455ac09bdd9c065e48ff1f0386146a06a9d" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.900191 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4snm6"] Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.915259 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/account-auditor/0.log" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.943877 4766 scope.go:117] "RemoveContainer" containerID="1b129982706444f72be0a69c788ebe8be308f1f177151a33e61fb8059a8dc993" Nov 26 02:29:11 crc kubenswrapper[4766]: E1126 02:29:11.944451 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b129982706444f72be0a69c788ebe8be308f1f177151a33e61fb8059a8dc993\": container with ID starting with 1b129982706444f72be0a69c788ebe8be308f1f177151a33e61fb8059a8dc993 not found: ID does not exist" containerID="1b129982706444f72be0a69c788ebe8be308f1f177151a33e61fb8059a8dc993" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.944492 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b129982706444f72be0a69c788ebe8be308f1f177151a33e61fb8059a8dc993"} err="failed to get container status \"1b129982706444f72be0a69c788ebe8be308f1f177151a33e61fb8059a8dc993\": rpc error: code = NotFound desc = could not find container \"1b129982706444f72be0a69c788ebe8be308f1f177151a33e61fb8059a8dc993\": container with ID starting with 1b129982706444f72be0a69c788ebe8be308f1f177151a33e61fb8059a8dc993 not found: ID does not exist" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.944516 4766 scope.go:117] "RemoveContainer" containerID="bf36d21bce57c1db501d2c69a3916d0708757f005dcb0ce7edc76cb3c42840fa" Nov 26 02:29:11 crc kubenswrapper[4766]: E1126 02:29:11.944852 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf36d21bce57c1db501d2c69a3916d0708757f005dcb0ce7edc76cb3c42840fa\": container with ID starting with bf36d21bce57c1db501d2c69a3916d0708757f005dcb0ce7edc76cb3c42840fa not found: ID does not exist" containerID="bf36d21bce57c1db501d2c69a3916d0708757f005dcb0ce7edc76cb3c42840fa" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.944874 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf36d21bce57c1db501d2c69a3916d0708757f005dcb0ce7edc76cb3c42840fa"} err="failed to get container status \"bf36d21bce57c1db501d2c69a3916d0708757f005dcb0ce7edc76cb3c42840fa\": rpc error: code = NotFound desc = could not find container \"bf36d21bce57c1db501d2c69a3916d0708757f005dcb0ce7edc76cb3c42840fa\": container with ID starting with bf36d21bce57c1db501d2c69a3916d0708757f005dcb0ce7edc76cb3c42840fa not found: ID does not exist" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.944889 4766 scope.go:117] "RemoveContainer" containerID="de8fa1c2126280db25301dfe7a944455ac09bdd9c065e48ff1f0386146a06a9d" Nov 26 02:29:11 crc kubenswrapper[4766]: E1126 02:29:11.945169 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de8fa1c2126280db25301dfe7a944455ac09bdd9c065e48ff1f0386146a06a9d\": container with ID starting with de8fa1c2126280db25301dfe7a944455ac09bdd9c065e48ff1f0386146a06a9d not found: ID does not exist" containerID="de8fa1c2126280db25301dfe7a944455ac09bdd9c065e48ff1f0386146a06a9d" Nov 26 02:29:11 crc kubenswrapper[4766]: I1126 02:29:11.945206 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de8fa1c2126280db25301dfe7a944455ac09bdd9c065e48ff1f0386146a06a9d"} err="failed to get container status \"de8fa1c2126280db25301dfe7a944455ac09bdd9c065e48ff1f0386146a06a9d\": rpc error: code = NotFound desc = could not find container \"de8fa1c2126280db25301dfe7a944455ac09bdd9c065e48ff1f0386146a06a9d\": container with ID starting with de8fa1c2126280db25301dfe7a944455ac09bdd9c065e48ff1f0386146a06a9d not found: ID does not exist" Nov 26 02:29:12 crc kubenswrapper[4766]: I1126 02:29:12.085797 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/account-reaper/0.log" Nov 26 02:29:12 crc kubenswrapper[4766]: I1126 02:29:12.100585 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/account-replicator/0.log" Nov 26 02:29:12 crc kubenswrapper[4766]: I1126 02:29:12.131024 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/account-server/0.log" Nov 26 02:29:12 crc kubenswrapper[4766]: I1126 02:29:12.171967 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/container-auditor/0.log" Nov 26 02:29:12 crc kubenswrapper[4766]: I1126 02:29:12.308244 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/container-server/0.log" Nov 26 02:29:12 crc kubenswrapper[4766]: I1126 02:29:12.337272 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/container-replicator/0.log" Nov 26 02:29:12 crc kubenswrapper[4766]: I1126 02:29:12.371691 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/container-updater/0.log" Nov 26 02:29:12 crc kubenswrapper[4766]: I1126 02:29:12.379804 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/object-auditor/0.log" Nov 26 02:29:12 crc kubenswrapper[4766]: I1126 02:29:12.557164 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/object-expirer/0.log" Nov 26 02:29:12 crc kubenswrapper[4766]: I1126 02:29:12.579698 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/object-replicator/0.log" Nov 26 02:29:12 crc kubenswrapper[4766]: I1126 02:29:12.585122 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/object-updater/0.log" Nov 26 02:29:12 crc kubenswrapper[4766]: I1126 02:29:12.606415 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/object-server/0.log" Nov 26 02:29:12 crc kubenswrapper[4766]: I1126 02:29:12.728079 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/rsync/0.log" Nov 26 02:29:13 crc kubenswrapper[4766]: I1126 02:29:13.014541 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/swift-recon-cron/0.log" Nov 26 02:29:13 crc kubenswrapper[4766]: I1126 02:29:13.097904 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-g6znx_eef488e4-2c28-4247-a8f4-b8e9008d518e/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:29:13 crc kubenswrapper[4766]: I1126 02:29:13.301791 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8_0667c3b6-bdae-4a36-acce-947435c51d91/telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:29:13 crc kubenswrapper[4766]: I1126 02:29:13.499300 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_f4a47960-5a63-4a92-a4e1-89d48aaa6515/test-operator-logs-container/0.log" Nov 26 02:29:13 crc kubenswrapper[4766]: I1126 02:29:13.643908 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs_d2efe136-b7e0-475d-9204-f0e36e720e96/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:29:13 crc kubenswrapper[4766]: I1126 02:29:13.839163 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1064657-b06b-4370-928e-5972ff8b46ae" path="/var/lib/kubelet/pods/a1064657-b06b-4370-928e-5972ff8b46ae/volumes" Nov 26 02:29:14 crc kubenswrapper[4766]: I1126 02:29:14.187522 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_edfc5f7f-1ff2-4147-b68e-b9be01ff6a99/tempest-tests-tempest-tests-runner/0.log" Nov 26 02:29:23 crc kubenswrapper[4766]: I1126 02:29:23.422236 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_f3679589-d23a-48c7-8b4a-712935d229f1/memcached/0.log" Nov 26 02:29:41 crc kubenswrapper[4766]: I1126 02:29:41.480196 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:29:41 crc kubenswrapper[4766]: I1126 02:29:41.480588 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:29:44 crc kubenswrapper[4766]: I1126 02:29:44.229281 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b_28a6a73a-8156-4df1-a41e-d2fe03ce2b25/util/0.log" Nov 26 02:29:44 crc kubenswrapper[4766]: I1126 02:29:44.408215 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b_28a6a73a-8156-4df1-a41e-d2fe03ce2b25/util/0.log" Nov 26 02:29:44 crc kubenswrapper[4766]: I1126 02:29:44.494342 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b_28a6a73a-8156-4df1-a41e-d2fe03ce2b25/pull/0.log" Nov 26 02:29:44 crc kubenswrapper[4766]: I1126 02:29:44.503978 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b_28a6a73a-8156-4df1-a41e-d2fe03ce2b25/pull/0.log" Nov 26 02:29:44 crc kubenswrapper[4766]: I1126 02:29:44.703608 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b_28a6a73a-8156-4df1-a41e-d2fe03ce2b25/util/0.log" Nov 26 02:29:44 crc kubenswrapper[4766]: I1126 02:29:44.706359 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b_28a6a73a-8156-4df1-a41e-d2fe03ce2b25/pull/0.log" Nov 26 02:29:44 crc kubenswrapper[4766]: I1126 02:29:44.753782 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b_28a6a73a-8156-4df1-a41e-d2fe03ce2b25/extract/0.log" Nov 26 02:29:44 crc kubenswrapper[4766]: I1126 02:29:44.890239 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-gbdhz_cc87f631-ba4e-4943-b51d-bf1012cca754/kube-rbac-proxy/0.log" Nov 26 02:29:44 crc kubenswrapper[4766]: I1126 02:29:44.930745 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-gbdhz_cc87f631-ba4e-4943-b51d-bf1012cca754/manager/3.log" Nov 26 02:29:44 crc kubenswrapper[4766]: I1126 02:29:44.953021 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-gbdhz_cc87f631-ba4e-4943-b51d-bf1012cca754/manager/2.log" Nov 26 02:29:45 crc kubenswrapper[4766]: I1126 02:29:45.068060 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-jgz6c_31b1b73b-1d10-49ac-8549-b202c965b63e/kube-rbac-proxy/0.log" Nov 26 02:29:45 crc kubenswrapper[4766]: I1126 02:29:45.134153 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-jgz6c_31b1b73b-1d10-49ac-8549-b202c965b63e/manager/3.log" Nov 26 02:29:45 crc kubenswrapper[4766]: I1126 02:29:45.186335 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-jgz6c_31b1b73b-1d10-49ac-8549-b202c965b63e/manager/2.log" Nov 26 02:29:45 crc kubenswrapper[4766]: I1126 02:29:45.280687 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-dqtcj_0e6cfef3-86af-4cf0-ab06-b3eef687836b/kube-rbac-proxy/0.log" Nov 26 02:29:45 crc kubenswrapper[4766]: I1126 02:29:45.357425 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-dqtcj_0e6cfef3-86af-4cf0-ab06-b3eef687836b/manager/3.log" Nov 26 02:29:45 crc kubenswrapper[4766]: I1126 02:29:45.380084 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-dqtcj_0e6cfef3-86af-4cf0-ab06-b3eef687836b/manager/2.log" Nov 26 02:29:45 crc kubenswrapper[4766]: I1126 02:29:45.505055 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-k68q9_45353fc4-9955-460b-9276-9792dfdaf8f4/kube-rbac-proxy/0.log" Nov 26 02:29:45 crc kubenswrapper[4766]: I1126 02:29:45.574557 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-k68q9_45353fc4-9955-460b-9276-9792dfdaf8f4/manager/2.log" Nov 26 02:29:45 crc kubenswrapper[4766]: I1126 02:29:45.596468 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-k68q9_45353fc4-9955-460b-9276-9792dfdaf8f4/manager/1.log" Nov 26 02:29:45 crc kubenswrapper[4766]: I1126 02:29:45.694299 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-tkj9b_1ecbd1d2-cb7f-432f-bc75-903d0643ea29/kube-rbac-proxy/0.log" Nov 26 02:29:45 crc kubenswrapper[4766]: I1126 02:29:45.991069 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-tkj9b_1ecbd1d2-cb7f-432f-bc75-903d0643ea29/manager/2.log" Nov 26 02:29:46 crc kubenswrapper[4766]: I1126 02:29:46.032048 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-tkj9b_1ecbd1d2-cb7f-432f-bc75-903d0643ea29/manager/3.log" Nov 26 02:29:46 crc kubenswrapper[4766]: I1126 02:29:46.147071 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-h4mrn_5284ad70-a36c-4666-8202-84633d83f50b/kube-rbac-proxy/0.log" Nov 26 02:29:46 crc kubenswrapper[4766]: I1126 02:29:46.166210 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-h4mrn_5284ad70-a36c-4666-8202-84633d83f50b/manager/3.log" Nov 26 02:29:46 crc kubenswrapper[4766]: I1126 02:29:46.202468 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-h4mrn_5284ad70-a36c-4666-8202-84633d83f50b/manager/2.log" Nov 26 02:29:46 crc kubenswrapper[4766]: I1126 02:29:46.345835 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-x6h4v_c3cec209-fdef-44a6-90d0-2d06cfd7e5cc/kube-rbac-proxy/0.log" Nov 26 02:29:46 crc kubenswrapper[4766]: I1126 02:29:46.404026 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-x6h4v_c3cec209-fdef-44a6-90d0-2d06cfd7e5cc/manager/2.log" Nov 26 02:29:46 crc kubenswrapper[4766]: I1126 02:29:46.435615 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-x6h4v_c3cec209-fdef-44a6-90d0-2d06cfd7e5cc/manager/3.log" Nov 26 02:29:46 crc kubenswrapper[4766]: I1126 02:29:46.574042 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-qtnxq_08bd3753-3b70-44f8-92ff-98bedd88f662/kube-rbac-proxy/0.log" Nov 26 02:29:46 crc kubenswrapper[4766]: I1126 02:29:46.699381 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-qtnxq_08bd3753-3b70-44f8-92ff-98bedd88f662/manager/2.log" Nov 26 02:29:46 crc kubenswrapper[4766]: I1126 02:29:46.701227 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-qtnxq_08bd3753-3b70-44f8-92ff-98bedd88f662/manager/3.log" Nov 26 02:29:46 crc kubenswrapper[4766]: I1126 02:29:46.910930 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-fcr76_3d0d59b2-0486-40a2-ae60-b738a4f76dc4/kube-rbac-proxy/0.log" Nov 26 02:29:46 crc kubenswrapper[4766]: I1126 02:29:46.951504 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-fcr76_3d0d59b2-0486-40a2-ae60-b738a4f76dc4/manager/3.log" Nov 26 02:29:46 crc kubenswrapper[4766]: I1126 02:29:46.959862 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-fcr76_3d0d59b2-0486-40a2-ae60-b738a4f76dc4/manager/2.log" Nov 26 02:29:47 crc kubenswrapper[4766]: I1126 02:29:47.161428 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-rs4kn_a16c5563-750f-4e91-93bd-4044c656cb88/kube-rbac-proxy/0.log" Nov 26 02:29:47 crc kubenswrapper[4766]: I1126 02:29:47.202883 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-rs4kn_a16c5563-750f-4e91-93bd-4044c656cb88/manager/2.log" Nov 26 02:29:47 crc kubenswrapper[4766]: I1126 02:29:47.236069 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-rs4kn_a16c5563-750f-4e91-93bd-4044c656cb88/manager/1.log" Nov 26 02:29:47 crc kubenswrapper[4766]: I1126 02:29:47.401463 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-z865g_749a25e6-23eb-4fdd-8e15-ac2874cdee7f/kube-rbac-proxy/0.log" Nov 26 02:29:47 crc kubenswrapper[4766]: I1126 02:29:47.444144 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-z865g_749a25e6-23eb-4fdd-8e15-ac2874cdee7f/manager/3.log" Nov 26 02:29:47 crc kubenswrapper[4766]: I1126 02:29:47.533421 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-z865g_749a25e6-23eb-4fdd-8e15-ac2874cdee7f/manager/2.log" Nov 26 02:29:47 crc kubenswrapper[4766]: I1126 02:29:47.605370 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-wdzgw_2bb23a3b-1f4a-409f-9d9c-0afe73a33352/kube-rbac-proxy/0.log" Nov 26 02:29:47 crc kubenswrapper[4766]: I1126 02:29:47.701341 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-wdzgw_2bb23a3b-1f4a-409f-9d9c-0afe73a33352/manager/3.log" Nov 26 02:29:47 crc kubenswrapper[4766]: I1126 02:29:47.768784 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-wdzgw_2bb23a3b-1f4a-409f-9d9c-0afe73a33352/manager/2.log" Nov 26 02:29:47 crc kubenswrapper[4766]: I1126 02:29:47.815438 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-dm7xb_8969652f-14c9-4618-ad2d-1b3fd1ad3627/kube-rbac-proxy/0.log" Nov 26 02:29:48 crc kubenswrapper[4766]: I1126 02:29:48.012417 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-dm7xb_8969652f-14c9-4618-ad2d-1b3fd1ad3627/manager/3.log" Nov 26 02:29:48 crc kubenswrapper[4766]: I1126 02:29:48.093542 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-dm7xb_8969652f-14c9-4618-ad2d-1b3fd1ad3627/manager/2.log" Nov 26 02:29:48 crc kubenswrapper[4766]: I1126 02:29:48.141148 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-bfw5f_91d3e9c8-8bbc-43b9-947b-71c4bb474b21/kube-rbac-proxy/0.log" Nov 26 02:29:48 crc kubenswrapper[4766]: I1126 02:29:48.180578 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-bfw5f_91d3e9c8-8bbc-43b9-947b-71c4bb474b21/manager/3.log" Nov 26 02:29:48 crc kubenswrapper[4766]: I1126 02:29:48.214722 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-bfw5f_91d3e9c8-8bbc-43b9-947b-71c4bb474b21/manager/2.log" Nov 26 02:29:48 crc kubenswrapper[4766]: I1126 02:29:48.316520 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t_150d5aa9-549d-4b43-b5c3-e6c81e942146/kube-rbac-proxy/0.log" Nov 26 02:29:48 crc kubenswrapper[4766]: I1126 02:29:48.344399 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t_150d5aa9-549d-4b43-b5c3-e6c81e942146/manager/1.log" Nov 26 02:29:48 crc kubenswrapper[4766]: I1126 02:29:48.373870 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t_150d5aa9-549d-4b43-b5c3-e6c81e942146/manager/0.log" Nov 26 02:29:48 crc kubenswrapper[4766]: I1126 02:29:48.512941 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-86674bdb4b-vq98m_d00b9c60-0fb7-49c6-944a-fac598cc22a4/manager/2.log" Nov 26 02:29:48 crc kubenswrapper[4766]: I1126 02:29:48.706569 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7d5fdccdc4-5sh97_62c72069-cf63-4244-a55e-0028aaaed380/operator/1.log" Nov 26 02:29:48 crc kubenswrapper[4766]: I1126 02:29:48.913963 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-sp492_5244fd2a-7bec-4649-af25-c77d4ec96ed0/registry-server/0.log" Nov 26 02:29:48 crc kubenswrapper[4766]: I1126 02:29:48.920040 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7d5fdccdc4-5sh97_62c72069-cf63-4244-a55e-0028aaaed380/operator/0.log" Nov 26 02:29:49 crc kubenswrapper[4766]: I1126 02:29:49.134556 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-9zhql_a2e7c098-bf9d-4d99-84bf-adfb49ee6b12/kube-rbac-proxy/0.log" Nov 26 02:29:49 crc kubenswrapper[4766]: I1126 02:29:49.204006 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-9zhql_a2e7c098-bf9d-4d99-84bf-adfb49ee6b12/manager/2.log" Nov 26 02:29:49 crc kubenswrapper[4766]: I1126 02:29:49.268230 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-9zhql_a2e7c098-bf9d-4d99-84bf-adfb49ee6b12/manager/3.log" Nov 26 02:29:49 crc kubenswrapper[4766]: I1126 02:29:49.377516 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-xlxgj_97419f32-8768-495f-a652-8a1e8701546f/kube-rbac-proxy/0.log" Nov 26 02:29:49 crc kubenswrapper[4766]: I1126 02:29:49.462387 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-xlxgj_97419f32-8768-495f-a652-8a1e8701546f/manager/2.log" Nov 26 02:29:49 crc kubenswrapper[4766]: I1126 02:29:49.491884 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-86674bdb4b-vq98m_d00b9c60-0fb7-49c6-944a-fac598cc22a4/manager/3.log" Nov 26 02:29:49 crc kubenswrapper[4766]: I1126 02:29:49.524443 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-xlxgj_97419f32-8768-495f-a652-8a1e8701546f/manager/1.log" Nov 26 02:29:49 crc kubenswrapper[4766]: I1126 02:29:49.647440 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-mx9vw_91ee52b5-0e2e-4732-87bf-809e378d33d3/operator/3.log" Nov 26 02:29:49 crc kubenswrapper[4766]: I1126 02:29:49.694405 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-mx9vw_91ee52b5-0e2e-4732-87bf-809e378d33d3/operator/2.log" Nov 26 02:29:49 crc kubenswrapper[4766]: I1126 02:29:49.705461 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-2t9w2_00a32e6c-edcc-4625-ac35-26fed19f1638/kube-rbac-proxy/0.log" Nov 26 02:29:49 crc kubenswrapper[4766]: I1126 02:29:49.970376 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-2t9w2_00a32e6c-edcc-4625-ac35-26fed19f1638/manager/2.log" Nov 26 02:29:49 crc kubenswrapper[4766]: I1126 02:29:49.988066 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-2t9w2_00a32e6c-edcc-4625-ac35-26fed19f1638/manager/3.log" Nov 26 02:29:50 crc kubenswrapper[4766]: I1126 02:29:50.102125 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5ccf99cfdd-9cfft_04e53d09-de84-4682-b17b-61f9ea7bf152/kube-rbac-proxy/0.log" Nov 26 02:29:50 crc kubenswrapper[4766]: I1126 02:29:50.199461 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5ccf99cfdd-9cfft_04e53d09-de84-4682-b17b-61f9ea7bf152/manager/2.log" Nov 26 02:29:50 crc kubenswrapper[4766]: I1126 02:29:50.201491 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5ccf99cfdd-9cfft_04e53d09-de84-4682-b17b-61f9ea7bf152/manager/3.log" Nov 26 02:29:50 crc kubenswrapper[4766]: I1126 02:29:50.290691 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-glkl6_40780971-20ed-4b4c-8d19-e74f6582ef8c/kube-rbac-proxy/0.log" Nov 26 02:29:50 crc kubenswrapper[4766]: I1126 02:29:50.364923 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-glkl6_40780971-20ed-4b4c-8d19-e74f6582ef8c/manager/0.log" Nov 26 02:29:50 crc kubenswrapper[4766]: I1126 02:29:50.375593 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-glkl6_40780971-20ed-4b4c-8d19-e74f6582ef8c/manager/1.log" Nov 26 02:29:50 crc kubenswrapper[4766]: I1126 02:29:50.397156 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-2hpg8_751b8e60-46d4-4f5e-b62d-800f641ccd40/kube-rbac-proxy/0.log" Nov 26 02:29:50 crc kubenswrapper[4766]: I1126 02:29:50.467712 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-2hpg8_751b8e60-46d4-4f5e-b62d-800f641ccd40/manager/3.log" Nov 26 02:29:50 crc kubenswrapper[4766]: I1126 02:29:50.544204 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-2hpg8_751b8e60-46d4-4f5e-b62d-800f641ccd40/manager/2.log" Nov 26 02:30:00 crc kubenswrapper[4766]: I1126 02:30:00.183408 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402070-9bnw9"] Nov 26 02:30:00 crc kubenswrapper[4766]: E1126 02:30:00.184498 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1064657-b06b-4370-928e-5972ff8b46ae" containerName="extract-content" Nov 26 02:30:00 crc kubenswrapper[4766]: I1126 02:30:00.184510 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1064657-b06b-4370-928e-5972ff8b46ae" containerName="extract-content" Nov 26 02:30:00 crc kubenswrapper[4766]: E1126 02:30:00.184530 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1064657-b06b-4370-928e-5972ff8b46ae" containerName="registry-server" Nov 26 02:30:00 crc kubenswrapper[4766]: I1126 02:30:00.184536 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1064657-b06b-4370-928e-5972ff8b46ae" containerName="registry-server" Nov 26 02:30:00 crc kubenswrapper[4766]: E1126 02:30:00.184561 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1064657-b06b-4370-928e-5972ff8b46ae" containerName="extract-utilities" Nov 26 02:30:00 crc kubenswrapper[4766]: I1126 02:30:00.184569 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1064657-b06b-4370-928e-5972ff8b46ae" containerName="extract-utilities" Nov 26 02:30:00 crc kubenswrapper[4766]: I1126 02:30:00.184815 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1064657-b06b-4370-928e-5972ff8b46ae" containerName="registry-server" Nov 26 02:30:00 crc kubenswrapper[4766]: I1126 02:30:00.185761 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402070-9bnw9" Nov 26 02:30:00 crc kubenswrapper[4766]: I1126 02:30:00.225057 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 26 02:30:00 crc kubenswrapper[4766]: I1126 02:30:00.226590 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402070-9bnw9"] Nov 26 02:30:00 crc kubenswrapper[4766]: I1126 02:30:00.226730 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 26 02:30:00 crc kubenswrapper[4766]: I1126 02:30:00.320889 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nbk2\" (UniqueName: \"kubernetes.io/projected/c24b7232-063d-4dc2-8f8b-d669c9ddfc42-kube-api-access-7nbk2\") pod \"collect-profiles-29402070-9bnw9\" (UID: \"c24b7232-063d-4dc2-8f8b-d669c9ddfc42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402070-9bnw9" Nov 26 02:30:00 crc kubenswrapper[4766]: I1126 02:30:00.320987 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c24b7232-063d-4dc2-8f8b-d669c9ddfc42-secret-volume\") pod \"collect-profiles-29402070-9bnw9\" (UID: \"c24b7232-063d-4dc2-8f8b-d669c9ddfc42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402070-9bnw9" Nov 26 02:30:00 crc kubenswrapper[4766]: I1126 02:30:00.321017 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c24b7232-063d-4dc2-8f8b-d669c9ddfc42-config-volume\") pod \"collect-profiles-29402070-9bnw9\" (UID: \"c24b7232-063d-4dc2-8f8b-d669c9ddfc42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402070-9bnw9" Nov 26 02:30:00 crc kubenswrapper[4766]: I1126 02:30:00.423077 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nbk2\" (UniqueName: \"kubernetes.io/projected/c24b7232-063d-4dc2-8f8b-d669c9ddfc42-kube-api-access-7nbk2\") pod \"collect-profiles-29402070-9bnw9\" (UID: \"c24b7232-063d-4dc2-8f8b-d669c9ddfc42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402070-9bnw9" Nov 26 02:30:00 crc kubenswrapper[4766]: I1126 02:30:00.423181 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c24b7232-063d-4dc2-8f8b-d669c9ddfc42-secret-volume\") pod \"collect-profiles-29402070-9bnw9\" (UID: \"c24b7232-063d-4dc2-8f8b-d669c9ddfc42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402070-9bnw9" Nov 26 02:30:00 crc kubenswrapper[4766]: I1126 02:30:00.423219 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c24b7232-063d-4dc2-8f8b-d669c9ddfc42-config-volume\") pod \"collect-profiles-29402070-9bnw9\" (UID: \"c24b7232-063d-4dc2-8f8b-d669c9ddfc42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402070-9bnw9" Nov 26 02:30:00 crc kubenswrapper[4766]: I1126 02:30:00.425710 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c24b7232-063d-4dc2-8f8b-d669c9ddfc42-config-volume\") pod \"collect-profiles-29402070-9bnw9\" (UID: \"c24b7232-063d-4dc2-8f8b-d669c9ddfc42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402070-9bnw9" Nov 26 02:30:00 crc kubenswrapper[4766]: I1126 02:30:00.430599 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c24b7232-063d-4dc2-8f8b-d669c9ddfc42-secret-volume\") pod \"collect-profiles-29402070-9bnw9\" (UID: \"c24b7232-063d-4dc2-8f8b-d669c9ddfc42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402070-9bnw9" Nov 26 02:30:00 crc kubenswrapper[4766]: I1126 02:30:00.458589 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nbk2\" (UniqueName: \"kubernetes.io/projected/c24b7232-063d-4dc2-8f8b-d669c9ddfc42-kube-api-access-7nbk2\") pod \"collect-profiles-29402070-9bnw9\" (UID: \"c24b7232-063d-4dc2-8f8b-d669c9ddfc42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402070-9bnw9" Nov 26 02:30:00 crc kubenswrapper[4766]: I1126 02:30:00.542303 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402070-9bnw9" Nov 26 02:30:01 crc kubenswrapper[4766]: I1126 02:30:01.043559 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402070-9bnw9"] Nov 26 02:30:01 crc kubenswrapper[4766]: I1126 02:30:01.369648 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402070-9bnw9" event={"ID":"c24b7232-063d-4dc2-8f8b-d669c9ddfc42","Type":"ContainerStarted","Data":"dc5b935a9b7554b85118b1de2debedaf6b8704437517dda93b5b8ce0a3ef5d50"} Nov 26 02:30:01 crc kubenswrapper[4766]: I1126 02:30:01.369986 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402070-9bnw9" event={"ID":"c24b7232-063d-4dc2-8f8b-d669c9ddfc42","Type":"ContainerStarted","Data":"f45e44f23f1ca65c4baac8084177eea6102fc3ce4d8f29195e754174ec46dab0"} Nov 26 02:30:01 crc kubenswrapper[4766]: I1126 02:30:01.395897 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29402070-9bnw9" podStartSLOduration=1.395875751 podStartE2EDuration="1.395875751s" podCreationTimestamp="2025-11-26 02:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 02:30:01.385439195 +0000 UTC m=+7582.234209635" watchObservedRunningTime="2025-11-26 02:30:01.395875751 +0000 UTC m=+7582.244646181" Nov 26 02:30:02 crc kubenswrapper[4766]: I1126 02:30:02.385570 4766 generic.go:334] "Generic (PLEG): container finished" podID="c24b7232-063d-4dc2-8f8b-d669c9ddfc42" containerID="dc5b935a9b7554b85118b1de2debedaf6b8704437517dda93b5b8ce0a3ef5d50" exitCode=0 Nov 26 02:30:02 crc kubenswrapper[4766]: I1126 02:30:02.385713 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402070-9bnw9" event={"ID":"c24b7232-063d-4dc2-8f8b-d669c9ddfc42","Type":"ContainerDied","Data":"dc5b935a9b7554b85118b1de2debedaf6b8704437517dda93b5b8ce0a3ef5d50"} Nov 26 02:30:03 crc kubenswrapper[4766]: I1126 02:30:03.872812 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402070-9bnw9" Nov 26 02:30:04 crc kubenswrapper[4766]: I1126 02:30:04.014715 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c24b7232-063d-4dc2-8f8b-d669c9ddfc42-config-volume\") pod \"c24b7232-063d-4dc2-8f8b-d669c9ddfc42\" (UID: \"c24b7232-063d-4dc2-8f8b-d669c9ddfc42\") " Nov 26 02:30:04 crc kubenswrapper[4766]: I1126 02:30:04.014918 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nbk2\" (UniqueName: \"kubernetes.io/projected/c24b7232-063d-4dc2-8f8b-d669c9ddfc42-kube-api-access-7nbk2\") pod \"c24b7232-063d-4dc2-8f8b-d669c9ddfc42\" (UID: \"c24b7232-063d-4dc2-8f8b-d669c9ddfc42\") " Nov 26 02:30:04 crc kubenswrapper[4766]: I1126 02:30:04.015074 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c24b7232-063d-4dc2-8f8b-d669c9ddfc42-secret-volume\") pod \"c24b7232-063d-4dc2-8f8b-d669c9ddfc42\" (UID: \"c24b7232-063d-4dc2-8f8b-d669c9ddfc42\") " Nov 26 02:30:04 crc kubenswrapper[4766]: I1126 02:30:04.015335 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c24b7232-063d-4dc2-8f8b-d669c9ddfc42-config-volume" (OuterVolumeSpecName: "config-volume") pod "c24b7232-063d-4dc2-8f8b-d669c9ddfc42" (UID: "c24b7232-063d-4dc2-8f8b-d669c9ddfc42"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 02:30:04 crc kubenswrapper[4766]: I1126 02:30:04.016074 4766 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c24b7232-063d-4dc2-8f8b-d669c9ddfc42-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 02:30:04 crc kubenswrapper[4766]: I1126 02:30:04.021591 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c24b7232-063d-4dc2-8f8b-d669c9ddfc42-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c24b7232-063d-4dc2-8f8b-d669c9ddfc42" (UID: "c24b7232-063d-4dc2-8f8b-d669c9ddfc42"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 02:30:04 crc kubenswrapper[4766]: I1126 02:30:04.025070 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c24b7232-063d-4dc2-8f8b-d669c9ddfc42-kube-api-access-7nbk2" (OuterVolumeSpecName: "kube-api-access-7nbk2") pod "c24b7232-063d-4dc2-8f8b-d669c9ddfc42" (UID: "c24b7232-063d-4dc2-8f8b-d669c9ddfc42"). InnerVolumeSpecName "kube-api-access-7nbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:30:04 crc kubenswrapper[4766]: I1126 02:30:04.119575 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nbk2\" (UniqueName: \"kubernetes.io/projected/c24b7232-063d-4dc2-8f8b-d669c9ddfc42-kube-api-access-7nbk2\") on node \"crc\" DevicePath \"\"" Nov 26 02:30:04 crc kubenswrapper[4766]: I1126 02:30:04.119619 4766 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c24b7232-063d-4dc2-8f8b-d669c9ddfc42-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 26 02:30:04 crc kubenswrapper[4766]: I1126 02:30:04.417269 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402070-9bnw9" event={"ID":"c24b7232-063d-4dc2-8f8b-d669c9ddfc42","Type":"ContainerDied","Data":"f45e44f23f1ca65c4baac8084177eea6102fc3ce4d8f29195e754174ec46dab0"} Nov 26 02:30:04 crc kubenswrapper[4766]: I1126 02:30:04.417312 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402070-9bnw9" Nov 26 02:30:04 crc kubenswrapper[4766]: I1126 02:30:04.417314 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f45e44f23f1ca65c4baac8084177eea6102fc3ce4d8f29195e754174ec46dab0" Nov 26 02:30:04 crc kubenswrapper[4766]: I1126 02:30:04.480853 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402025-vsmwp"] Nov 26 02:30:04 crc kubenswrapper[4766]: I1126 02:30:04.491103 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402025-vsmwp"] Nov 26 02:30:05 crc kubenswrapper[4766]: I1126 02:30:05.846496 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb60fd86-c072-40b9-a2e3-047f3b83c5bd" path="/var/lib/kubelet/pods/bb60fd86-c072-40b9-a2e3-047f3b83c5bd/volumes" Nov 26 02:30:06 crc kubenswrapper[4766]: I1126 02:30:06.549380 4766 scope.go:117] "RemoveContainer" containerID="eef8ee593ead07a9e091d4276577af4eca5229e4b3eab352eb6c8e78f8cbd6ce" Nov 26 02:30:10 crc kubenswrapper[4766]: I1126 02:30:10.923981 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-pm5j5_752819e1-d0ad-4820-979a-bf6ebfc59b4d/control-plane-machine-set-operator/0.log" Nov 26 02:30:11 crc kubenswrapper[4766]: I1126 02:30:11.128996 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-x2vlb_c8260737-caa6-42f1-8854-ad2b97a1ee42/kube-rbac-proxy/0.log" Nov 26 02:30:11 crc kubenswrapper[4766]: I1126 02:30:11.129046 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-x2vlb_c8260737-caa6-42f1-8854-ad2b97a1ee42/machine-api-operator/0.log" Nov 26 02:30:11 crc kubenswrapper[4766]: I1126 02:30:11.480205 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:30:11 crc kubenswrapper[4766]: I1126 02:30:11.480274 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:30:11 crc kubenswrapper[4766]: I1126 02:30:11.480329 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 02:30:11 crc kubenswrapper[4766]: I1126 02:30:11.481571 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e12146d79b16ca01dd3ddaff20fdb9b94b5dd64df0a00c9f78979d3aecbe9802"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 02:30:11 crc kubenswrapper[4766]: I1126 02:30:11.481685 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://e12146d79b16ca01dd3ddaff20fdb9b94b5dd64df0a00c9f78979d3aecbe9802" gracePeriod=600 Nov 26 02:30:12 crc kubenswrapper[4766]: I1126 02:30:12.534070 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="e12146d79b16ca01dd3ddaff20fdb9b94b5dd64df0a00c9f78979d3aecbe9802" exitCode=0 Nov 26 02:30:12 crc kubenswrapper[4766]: I1126 02:30:12.534166 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"e12146d79b16ca01dd3ddaff20fdb9b94b5dd64df0a00c9f78979d3aecbe9802"} Nov 26 02:30:12 crc kubenswrapper[4766]: I1126 02:30:12.534589 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a"} Nov 26 02:30:12 crc kubenswrapper[4766]: I1126 02:30:12.534612 4766 scope.go:117] "RemoveContainer" containerID="f97a1385b9b58286347e0fa8236c89d0c51cd411c71f60a53eb5f85521eda562" Nov 26 02:30:25 crc kubenswrapper[4766]: I1126 02:30:25.604899 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-nd9m2_954f9d95-f1ab-4e36-9b7a-d16fbe282e53/cert-manager-controller/1.log" Nov 26 02:30:25 crc kubenswrapper[4766]: I1126 02:30:25.667908 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-nd9m2_954f9d95-f1ab-4e36-9b7a-d16fbe282e53/cert-manager-controller/0.log" Nov 26 02:30:25 crc kubenswrapper[4766]: I1126 02:30:25.809550 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-qjmjr_e6ee9725-56d1-426a-afeb-c00b1abb9e2a/cert-manager-cainjector/0.log" Nov 26 02:30:25 crc kubenswrapper[4766]: I1126 02:30:25.869299 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-btczr_403aadd6-8bd1-4446-b0e7-f342f31e54c9/cert-manager-webhook/0.log" Nov 26 02:30:40 crc kubenswrapper[4766]: I1126 02:30:40.337812 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-nzhrs_d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130/nmstate-console-plugin/0.log" Nov 26 02:30:40 crc kubenswrapper[4766]: I1126 02:30:40.553156 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-q4sb9_58a75814-134a-4a52-aae0-5c21109f658f/nmstate-handler/0.log" Nov 26 02:30:40 crc kubenswrapper[4766]: I1126 02:30:40.579709 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-f58l6_8f6f1e2d-6ebd-4f98-bc08-8cba715f33ae/nmstate-metrics/0.log" Nov 26 02:30:40 crc kubenswrapper[4766]: I1126 02:30:40.600002 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-f58l6_8f6f1e2d-6ebd-4f98-bc08-8cba715f33ae/kube-rbac-proxy/0.log" Nov 26 02:30:40 crc kubenswrapper[4766]: I1126 02:30:40.731014 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-ztjbb_61f41116-4bc2-47c5-b805-d0a0990c6024/nmstate-operator/0.log" Nov 26 02:30:40 crc kubenswrapper[4766]: I1126 02:30:40.804308 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-mn64d_d874f3cf-9f30-431e-81e1-6334a0cf4351/nmstate-webhook/0.log" Nov 26 02:30:54 crc kubenswrapper[4766]: I1126 02:30:54.212271 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-7bbf4564c5-m77vx_b90f7164-a5e9-452f-8733-f549f15e17ec/kube-rbac-proxy/0.log" Nov 26 02:30:54 crc kubenswrapper[4766]: I1126 02:30:54.295228 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-7bbf4564c5-m77vx_b90f7164-a5e9-452f-8733-f549f15e17ec/manager/1.log" Nov 26 02:30:54 crc kubenswrapper[4766]: I1126 02:30:54.402828 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-7bbf4564c5-m77vx_b90f7164-a5e9-452f-8733-f549f15e17ec/manager/0.log" Nov 26 02:31:09 crc kubenswrapper[4766]: I1126 02:31:09.152997 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-ff9846bd-mq2w5_2850bfaf-5303-44e6-8fc6-616612cb68e1/cluster-logging-operator/0.log" Nov 26 02:31:09 crc kubenswrapper[4766]: I1126 02:31:09.373075 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-zzrrp_bed1e93e-177c-4e72-99b7-e35f6d8f6cf9/collector/0.log" Nov 26 02:31:09 crc kubenswrapper[4766]: I1126 02:31:09.534917 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e/loki-compactor/0.log" Nov 26 02:31:09 crc kubenswrapper[4766]: I1126 02:31:09.623513 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-76cc67bf56-7khhv_bfb11063-6ad6-463b-87e6-eaf9333e2d4a/loki-distributor/0.log" Nov 26 02:31:09 crc kubenswrapper[4766]: I1126 02:31:09.813394 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-5d9974794d-bbcbk_261b6830-cc71-4704-ab0d-fd3a9708a786/opa/0.log" Nov 26 02:31:09 crc kubenswrapper[4766]: I1126 02:31:09.836718 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-5d9974794d-bbcbk_261b6830-cc71-4704-ab0d-fd3a9708a786/gateway/0.log" Nov 26 02:31:09 crc kubenswrapper[4766]: I1126 02:31:09.890356 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-5d9974794d-hgkj4_bc3b1ab0-39c8-42a1-b0d1-c45baa43df87/gateway/0.log" Nov 26 02:31:10 crc kubenswrapper[4766]: I1126 02:31:10.025438 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_87bee09b-03e7-4b6c-a28f-7e12aab1eb45/loki-index-gateway/0.log" Nov 26 02:31:10 crc kubenswrapper[4766]: I1126 02:31:10.040831 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-5d9974794d-hgkj4_bc3b1ab0-39c8-42a1-b0d1-c45baa43df87/opa/0.log" Nov 26 02:31:10 crc kubenswrapper[4766]: I1126 02:31:10.294932 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-5895d59bb8-tv6v8_036dedff-92b9-4ad9-89b4-15dfb60e73c2/loki-querier/0.log" Nov 26 02:31:10 crc kubenswrapper[4766]: I1126 02:31:10.305843 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_aae58301-788d-430a-b020-a5f530d70c5a/loki-ingester/0.log" Nov 26 02:31:10 crc kubenswrapper[4766]: I1126 02:31:10.529197 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-84558f7c9f-tltn8_7fb68151-644a-4d8b-9274-18bb7073cc41/loki-query-frontend/0.log" Nov 26 02:31:26 crc kubenswrapper[4766]: I1126 02:31:26.020715 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-9dzn2_0c9c337a-681d-4a68-8e51-5d690a326fe7/kube-rbac-proxy/0.log" Nov 26 02:31:26 crc kubenswrapper[4766]: I1126 02:31:26.176975 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-9dzn2_0c9c337a-681d-4a68-8e51-5d690a326fe7/controller/0.log" Nov 26 02:31:26 crc kubenswrapper[4766]: I1126 02:31:26.272301 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-frr-files/0.log" Nov 26 02:31:26 crc kubenswrapper[4766]: I1126 02:31:26.412243 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-reloader/0.log" Nov 26 02:31:26 crc kubenswrapper[4766]: I1126 02:31:26.416321 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-frr-files/0.log" Nov 26 02:31:26 crc kubenswrapper[4766]: I1126 02:31:26.456298 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-metrics/0.log" Nov 26 02:31:26 crc kubenswrapper[4766]: I1126 02:31:26.556947 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-reloader/0.log" Nov 26 02:31:26 crc kubenswrapper[4766]: I1126 02:31:26.719055 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-metrics/0.log" Nov 26 02:31:26 crc kubenswrapper[4766]: I1126 02:31:26.755030 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-metrics/0.log" Nov 26 02:31:26 crc kubenswrapper[4766]: I1126 02:31:26.764022 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-reloader/0.log" Nov 26 02:31:26 crc kubenswrapper[4766]: I1126 02:31:26.775635 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-frr-files/0.log" Nov 26 02:31:26 crc kubenswrapper[4766]: I1126 02:31:26.950044 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-frr-files/0.log" Nov 26 02:31:26 crc kubenswrapper[4766]: I1126 02:31:26.958536 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-metrics/0.log" Nov 26 02:31:26 crc kubenswrapper[4766]: I1126 02:31:26.979025 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-reloader/0.log" Nov 26 02:31:26 crc kubenswrapper[4766]: I1126 02:31:26.984597 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/controller/0.log" Nov 26 02:31:27 crc kubenswrapper[4766]: I1126 02:31:27.128167 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/frr-metrics/0.log" Nov 26 02:31:27 crc kubenswrapper[4766]: I1126 02:31:27.176419 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/kube-rbac-proxy/0.log" Nov 26 02:31:27 crc kubenswrapper[4766]: I1126 02:31:27.214438 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/kube-rbac-proxy-frr/0.log" Nov 26 02:31:27 crc kubenswrapper[4766]: I1126 02:31:27.466222 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/reloader/0.log" Nov 26 02:31:27 crc kubenswrapper[4766]: I1126 02:31:27.535482 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-k6vct_88faf6a0-1824-43f6-b711-09cb78c54a5d/frr-k8s-webhook-server/0.log" Nov 26 02:31:27 crc kubenswrapper[4766]: I1126 02:31:27.770180 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-764c8cf56f-vrwpt_fd3a6275-9f47-4799-bc15-96a51b93f31b/manager/3.log" Nov 26 02:31:27 crc kubenswrapper[4766]: I1126 02:31:27.832354 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-764c8cf56f-vrwpt_fd3a6275-9f47-4799-bc15-96a51b93f31b/manager/2.log" Nov 26 02:31:28 crc kubenswrapper[4766]: I1126 02:31:28.271384 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6f5d8f677c-qjhpt_ce7c66d0-5857-46ce-a725-6531c37891cb/webhook-server/0.log" Nov 26 02:31:28 crc kubenswrapper[4766]: I1126 02:31:28.400207 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-d5tdc_76a9ad7d-db95-4177-839f-db3e7cc1ac81/kube-rbac-proxy/0.log" Nov 26 02:31:29 crc kubenswrapper[4766]: I1126 02:31:29.251868 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/frr/0.log" Nov 26 02:31:29 crc kubenswrapper[4766]: I1126 02:31:29.261145 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-d5tdc_76a9ad7d-db95-4177-839f-db3e7cc1ac81/speaker/0.log" Nov 26 02:31:35 crc kubenswrapper[4766]: I1126 02:31:35.095379 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9gddf"] Nov 26 02:31:35 crc kubenswrapper[4766]: E1126 02:31:35.097908 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c24b7232-063d-4dc2-8f8b-d669c9ddfc42" containerName="collect-profiles" Nov 26 02:31:35 crc kubenswrapper[4766]: I1126 02:31:35.098035 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="c24b7232-063d-4dc2-8f8b-d669c9ddfc42" containerName="collect-profiles" Nov 26 02:31:35 crc kubenswrapper[4766]: I1126 02:31:35.098452 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="c24b7232-063d-4dc2-8f8b-d669c9ddfc42" containerName="collect-profiles" Nov 26 02:31:35 crc kubenswrapper[4766]: I1126 02:31:35.100748 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9gddf" Nov 26 02:31:35 crc kubenswrapper[4766]: I1126 02:31:35.112068 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9gddf"] Nov 26 02:31:35 crc kubenswrapper[4766]: I1126 02:31:35.153387 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a218fd1-d674-44e8-97dd-fae53b3ae179-catalog-content\") pod \"redhat-operators-9gddf\" (UID: \"2a218fd1-d674-44e8-97dd-fae53b3ae179\") " pod="openshift-marketplace/redhat-operators-9gddf" Nov 26 02:31:35 crc kubenswrapper[4766]: I1126 02:31:35.153533 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cgjj\" (UniqueName: \"kubernetes.io/projected/2a218fd1-d674-44e8-97dd-fae53b3ae179-kube-api-access-4cgjj\") pod \"redhat-operators-9gddf\" (UID: \"2a218fd1-d674-44e8-97dd-fae53b3ae179\") " pod="openshift-marketplace/redhat-operators-9gddf" Nov 26 02:31:35 crc kubenswrapper[4766]: I1126 02:31:35.153554 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a218fd1-d674-44e8-97dd-fae53b3ae179-utilities\") pod \"redhat-operators-9gddf\" (UID: \"2a218fd1-d674-44e8-97dd-fae53b3ae179\") " pod="openshift-marketplace/redhat-operators-9gddf" Nov 26 02:31:35 crc kubenswrapper[4766]: I1126 02:31:35.256275 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a218fd1-d674-44e8-97dd-fae53b3ae179-catalog-content\") pod \"redhat-operators-9gddf\" (UID: \"2a218fd1-d674-44e8-97dd-fae53b3ae179\") " pod="openshift-marketplace/redhat-operators-9gddf" Nov 26 02:31:35 crc kubenswrapper[4766]: I1126 02:31:35.256613 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cgjj\" (UniqueName: \"kubernetes.io/projected/2a218fd1-d674-44e8-97dd-fae53b3ae179-kube-api-access-4cgjj\") pod \"redhat-operators-9gddf\" (UID: \"2a218fd1-d674-44e8-97dd-fae53b3ae179\") " pod="openshift-marketplace/redhat-operators-9gddf" Nov 26 02:31:35 crc kubenswrapper[4766]: I1126 02:31:35.256715 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a218fd1-d674-44e8-97dd-fae53b3ae179-utilities\") pod \"redhat-operators-9gddf\" (UID: \"2a218fd1-d674-44e8-97dd-fae53b3ae179\") " pod="openshift-marketplace/redhat-operators-9gddf" Nov 26 02:31:35 crc kubenswrapper[4766]: I1126 02:31:35.256715 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a218fd1-d674-44e8-97dd-fae53b3ae179-catalog-content\") pod \"redhat-operators-9gddf\" (UID: \"2a218fd1-d674-44e8-97dd-fae53b3ae179\") " pod="openshift-marketplace/redhat-operators-9gddf" Nov 26 02:31:35 crc kubenswrapper[4766]: I1126 02:31:35.256934 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a218fd1-d674-44e8-97dd-fae53b3ae179-utilities\") pod \"redhat-operators-9gddf\" (UID: \"2a218fd1-d674-44e8-97dd-fae53b3ae179\") " pod="openshift-marketplace/redhat-operators-9gddf" Nov 26 02:31:35 crc kubenswrapper[4766]: I1126 02:31:35.287816 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cgjj\" (UniqueName: \"kubernetes.io/projected/2a218fd1-d674-44e8-97dd-fae53b3ae179-kube-api-access-4cgjj\") pod \"redhat-operators-9gddf\" (UID: \"2a218fd1-d674-44e8-97dd-fae53b3ae179\") " pod="openshift-marketplace/redhat-operators-9gddf" Nov 26 02:31:35 crc kubenswrapper[4766]: I1126 02:31:35.441056 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9gddf" Nov 26 02:31:36 crc kubenswrapper[4766]: I1126 02:31:36.030682 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9gddf"] Nov 26 02:31:36 crc kubenswrapper[4766]: I1126 02:31:36.462835 4766 generic.go:334] "Generic (PLEG): container finished" podID="2a218fd1-d674-44e8-97dd-fae53b3ae179" containerID="374867ec51df9b8d6e9a945f072cac226f8fdb62b6191207f5004ddcf69327eb" exitCode=0 Nov 26 02:31:36 crc kubenswrapper[4766]: I1126 02:31:36.462952 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9gddf" event={"ID":"2a218fd1-d674-44e8-97dd-fae53b3ae179","Type":"ContainerDied","Data":"374867ec51df9b8d6e9a945f072cac226f8fdb62b6191207f5004ddcf69327eb"} Nov 26 02:31:36 crc kubenswrapper[4766]: I1126 02:31:36.464887 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9gddf" event={"ID":"2a218fd1-d674-44e8-97dd-fae53b3ae179","Type":"ContainerStarted","Data":"4a1401dda042657c3ffdaa3ee74d3a22feb83858be9851d835f60ec6ffb4c034"} Nov 26 02:31:36 crc kubenswrapper[4766]: I1126 02:31:36.465509 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 02:31:37 crc kubenswrapper[4766]: I1126 02:31:37.475985 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9gddf" event={"ID":"2a218fd1-d674-44e8-97dd-fae53b3ae179","Type":"ContainerStarted","Data":"c6e5511c31f7246fa7efbb516a68215e1a398304639a033f76c59431f675db41"} Nov 26 02:31:41 crc kubenswrapper[4766]: I1126 02:31:41.530803 4766 generic.go:334] "Generic (PLEG): container finished" podID="2a218fd1-d674-44e8-97dd-fae53b3ae179" containerID="c6e5511c31f7246fa7efbb516a68215e1a398304639a033f76c59431f675db41" exitCode=0 Nov 26 02:31:41 crc kubenswrapper[4766]: I1126 02:31:41.530864 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9gddf" event={"ID":"2a218fd1-d674-44e8-97dd-fae53b3ae179","Type":"ContainerDied","Data":"c6e5511c31f7246fa7efbb516a68215e1a398304639a033f76c59431f675db41"} Nov 26 02:31:42 crc kubenswrapper[4766]: I1126 02:31:42.556423 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9gddf" event={"ID":"2a218fd1-d674-44e8-97dd-fae53b3ae179","Type":"ContainerStarted","Data":"102669f53cca04ea057cacf1845cf71cd5f172994ec8892c00f0918af24ae80b"} Nov 26 02:31:42 crc kubenswrapper[4766]: I1126 02:31:42.577905 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9gddf" podStartSLOduration=2.114233237 podStartE2EDuration="7.57788962s" podCreationTimestamp="2025-11-26 02:31:35 +0000 UTC" firstStartedPulling="2025-11-26 02:31:36.464515721 +0000 UTC m=+7677.313286151" lastFinishedPulling="2025-11-26 02:31:41.928172084 +0000 UTC m=+7682.776942534" observedRunningTime="2025-11-26 02:31:42.577331737 +0000 UTC m=+7683.426102167" watchObservedRunningTime="2025-11-26 02:31:42.57788962 +0000 UTC m=+7683.426660040" Nov 26 02:31:43 crc kubenswrapper[4766]: I1126 02:31:43.459224 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58_64a27a6e-0ec0-42e2-8606-370bafc71c04/util/0.log" Nov 26 02:31:43 crc kubenswrapper[4766]: I1126 02:31:43.672051 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58_64a27a6e-0ec0-42e2-8606-370bafc71c04/util/0.log" Nov 26 02:31:43 crc kubenswrapper[4766]: I1126 02:31:43.676226 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58_64a27a6e-0ec0-42e2-8606-370bafc71c04/pull/0.log" Nov 26 02:31:43 crc kubenswrapper[4766]: I1126 02:31:43.711093 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58_64a27a6e-0ec0-42e2-8606-370bafc71c04/pull/0.log" Nov 26 02:31:43 crc kubenswrapper[4766]: I1126 02:31:43.917963 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58_64a27a6e-0ec0-42e2-8606-370bafc71c04/extract/0.log" Nov 26 02:31:43 crc kubenswrapper[4766]: I1126 02:31:43.934862 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58_64a27a6e-0ec0-42e2-8606-370bafc71c04/util/0.log" Nov 26 02:31:43 crc kubenswrapper[4766]: I1126 02:31:43.970413 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58_64a27a6e-0ec0-42e2-8606-370bafc71c04/pull/0.log" Nov 26 02:31:44 crc kubenswrapper[4766]: I1126 02:31:44.097410 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm_9e6e8bab-7232-4619-a55d-95125dfaf6d0/util/0.log" Nov 26 02:31:44 crc kubenswrapper[4766]: I1126 02:31:44.323809 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm_9e6e8bab-7232-4619-a55d-95125dfaf6d0/pull/0.log" Nov 26 02:31:44 crc kubenswrapper[4766]: I1126 02:31:44.331426 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm_9e6e8bab-7232-4619-a55d-95125dfaf6d0/util/0.log" Nov 26 02:31:44 crc kubenswrapper[4766]: I1126 02:31:44.404478 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm_9e6e8bab-7232-4619-a55d-95125dfaf6d0/pull/0.log" Nov 26 02:31:44 crc kubenswrapper[4766]: I1126 02:31:44.651823 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm_9e6e8bab-7232-4619-a55d-95125dfaf6d0/util/0.log" Nov 26 02:31:44 crc kubenswrapper[4766]: I1126 02:31:44.712714 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm_9e6e8bab-7232-4619-a55d-95125dfaf6d0/pull/0.log" Nov 26 02:31:44 crc kubenswrapper[4766]: I1126 02:31:44.801891 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm_9e6e8bab-7232-4619-a55d-95125dfaf6d0/extract/0.log" Nov 26 02:31:44 crc kubenswrapper[4766]: I1126 02:31:44.931841 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq_4d00d8e8-b83e-4585-9678-fb64e663fec3/util/0.log" Nov 26 02:31:45 crc kubenswrapper[4766]: I1126 02:31:45.129293 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq_4d00d8e8-b83e-4585-9678-fb64e663fec3/pull/0.log" Nov 26 02:31:45 crc kubenswrapper[4766]: I1126 02:31:45.129466 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq_4d00d8e8-b83e-4585-9678-fb64e663fec3/pull/0.log" Nov 26 02:31:45 crc kubenswrapper[4766]: I1126 02:31:45.139927 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq_4d00d8e8-b83e-4585-9678-fb64e663fec3/util/0.log" Nov 26 02:31:45 crc kubenswrapper[4766]: I1126 02:31:45.337407 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq_4d00d8e8-b83e-4585-9678-fb64e663fec3/pull/0.log" Nov 26 02:31:45 crc kubenswrapper[4766]: I1126 02:31:45.351006 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq_4d00d8e8-b83e-4585-9678-fb64e663fec3/util/0.log" Nov 26 02:31:45 crc kubenswrapper[4766]: I1126 02:31:45.352074 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq_4d00d8e8-b83e-4585-9678-fb64e663fec3/extract/0.log" Nov 26 02:31:45 crc kubenswrapper[4766]: I1126 02:31:45.442032 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9gddf" Nov 26 02:31:45 crc kubenswrapper[4766]: I1126 02:31:45.442094 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9gddf" Nov 26 02:31:45 crc kubenswrapper[4766]: I1126 02:31:45.536867 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9_36c1a0d9-7873-4ce0-b926-17e535470807/util/0.log" Nov 26 02:31:45 crc kubenswrapper[4766]: I1126 02:31:45.726148 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9_36c1a0d9-7873-4ce0-b926-17e535470807/util/0.log" Nov 26 02:31:45 crc kubenswrapper[4766]: I1126 02:31:45.749068 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9_36c1a0d9-7873-4ce0-b926-17e535470807/pull/0.log" Nov 26 02:31:45 crc kubenswrapper[4766]: I1126 02:31:45.749360 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9_36c1a0d9-7873-4ce0-b926-17e535470807/pull/0.log" Nov 26 02:31:45 crc kubenswrapper[4766]: I1126 02:31:45.997926 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9_36c1a0d9-7873-4ce0-b926-17e535470807/util/0.log" Nov 26 02:31:46 crc kubenswrapper[4766]: I1126 02:31:46.026140 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9_36c1a0d9-7873-4ce0-b926-17e535470807/pull/0.log" Nov 26 02:31:46 crc kubenswrapper[4766]: I1126 02:31:46.041566 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9_36c1a0d9-7873-4ce0-b926-17e535470807/extract/0.log" Nov 26 02:31:46 crc kubenswrapper[4766]: I1126 02:31:46.170254 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvvwl_74038ed5-c727-40ae-bc0f-d779aa47613d/extract-utilities/0.log" Nov 26 02:31:46 crc kubenswrapper[4766]: I1126 02:31:46.384989 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvvwl_74038ed5-c727-40ae-bc0f-d779aa47613d/extract-content/0.log" Nov 26 02:31:46 crc kubenswrapper[4766]: I1126 02:31:46.397420 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvvwl_74038ed5-c727-40ae-bc0f-d779aa47613d/extract-utilities/0.log" Nov 26 02:31:46 crc kubenswrapper[4766]: I1126 02:31:46.402958 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvvwl_74038ed5-c727-40ae-bc0f-d779aa47613d/extract-content/0.log" Nov 26 02:31:46 crc kubenswrapper[4766]: I1126 02:31:46.506615 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9gddf" podUID="2a218fd1-d674-44e8-97dd-fae53b3ae179" containerName="registry-server" probeResult="failure" output=< Nov 26 02:31:46 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 02:31:46 crc kubenswrapper[4766]: > Nov 26 02:31:46 crc kubenswrapper[4766]: I1126 02:31:46.585441 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvvwl_74038ed5-c727-40ae-bc0f-d779aa47613d/extract-content/0.log" Nov 26 02:31:46 crc kubenswrapper[4766]: I1126 02:31:46.606489 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvvwl_74038ed5-c727-40ae-bc0f-d779aa47613d/extract-utilities/0.log" Nov 26 02:31:46 crc kubenswrapper[4766]: I1126 02:31:46.806613 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6vj5j_e1f0994c-ef19-451f-bc61-fe7b66c3399b/extract-utilities/0.log" Nov 26 02:31:46 crc kubenswrapper[4766]: I1126 02:31:46.984203 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6vj5j_e1f0994c-ef19-451f-bc61-fe7b66c3399b/extract-utilities/0.log" Nov 26 02:31:47 crc kubenswrapper[4766]: I1126 02:31:47.018583 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6vj5j_e1f0994c-ef19-451f-bc61-fe7b66c3399b/extract-content/0.log" Nov 26 02:31:47 crc kubenswrapper[4766]: I1126 02:31:47.165962 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6vj5j_e1f0994c-ef19-451f-bc61-fe7b66c3399b/extract-content/0.log" Nov 26 02:31:47 crc kubenswrapper[4766]: I1126 02:31:47.174730 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvvwl_74038ed5-c727-40ae-bc0f-d779aa47613d/registry-server/0.log" Nov 26 02:31:47 crc kubenswrapper[4766]: I1126 02:31:47.223714 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6vj5j_e1f0994c-ef19-451f-bc61-fe7b66c3399b/extract-utilities/0.log" Nov 26 02:31:47 crc kubenswrapper[4766]: I1126 02:31:47.276104 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6vj5j_e1f0994c-ef19-451f-bc61-fe7b66c3399b/extract-content/0.log" Nov 26 02:31:47 crc kubenswrapper[4766]: I1126 02:31:47.374595 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk_631bfaa9-4706-44f7-9417-500418a18c73/util/0.log" Nov 26 02:31:47 crc kubenswrapper[4766]: I1126 02:31:47.617462 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk_631bfaa9-4706-44f7-9417-500418a18c73/pull/0.log" Nov 26 02:31:47 crc kubenswrapper[4766]: I1126 02:31:47.654539 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk_631bfaa9-4706-44f7-9417-500418a18c73/util/0.log" Nov 26 02:31:47 crc kubenswrapper[4766]: I1126 02:31:47.728264 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk_631bfaa9-4706-44f7-9417-500418a18c73/pull/0.log" Nov 26 02:31:47 crc kubenswrapper[4766]: I1126 02:31:47.745196 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6vj5j_e1f0994c-ef19-451f-bc61-fe7b66c3399b/registry-server/0.log" Nov 26 02:31:47 crc kubenswrapper[4766]: I1126 02:31:47.869866 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk_631bfaa9-4706-44f7-9417-500418a18c73/util/0.log" Nov 26 02:31:47 crc kubenswrapper[4766]: I1126 02:31:47.870782 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk_631bfaa9-4706-44f7-9417-500418a18c73/pull/0.log" Nov 26 02:31:47 crc kubenswrapper[4766]: I1126 02:31:47.886012 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk_631bfaa9-4706-44f7-9417-500418a18c73/extract/0.log" Nov 26 02:31:47 crc kubenswrapper[4766]: I1126 02:31:47.953218 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-2xsk5_cad040b4-5d74-4cd0-b726-93d9c6cb73e2/marketplace-operator/0.log" Nov 26 02:31:48 crc kubenswrapper[4766]: I1126 02:31:48.038619 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vhw86_f18c8ba9-685a-4cd8-92e9-797074b81278/extract-utilities/0.log" Nov 26 02:31:48 crc kubenswrapper[4766]: I1126 02:31:48.180255 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vhw86_f18c8ba9-685a-4cd8-92e9-797074b81278/extract-utilities/0.log" Nov 26 02:31:48 crc kubenswrapper[4766]: I1126 02:31:48.221842 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vhw86_f18c8ba9-685a-4cd8-92e9-797074b81278/extract-content/0.log" Nov 26 02:31:48 crc kubenswrapper[4766]: I1126 02:31:48.221849 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vhw86_f18c8ba9-685a-4cd8-92e9-797074b81278/extract-content/0.log" Nov 26 02:31:48 crc kubenswrapper[4766]: I1126 02:31:48.380456 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vhw86_f18c8ba9-685a-4cd8-92e9-797074b81278/extract-content/0.log" Nov 26 02:31:48 crc kubenswrapper[4766]: I1126 02:31:48.402909 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vhw86_f18c8ba9-685a-4cd8-92e9-797074b81278/extract-utilities/0.log" Nov 26 02:31:48 crc kubenswrapper[4766]: I1126 02:31:48.509089 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9gddf_2a218fd1-d674-44e8-97dd-fae53b3ae179/extract-utilities/0.log" Nov 26 02:31:48 crc kubenswrapper[4766]: I1126 02:31:48.713590 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vhw86_f18c8ba9-685a-4cd8-92e9-797074b81278/registry-server/0.log" Nov 26 02:31:48 crc kubenswrapper[4766]: I1126 02:31:48.761203 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9gddf_2a218fd1-d674-44e8-97dd-fae53b3ae179/extract-utilities/0.log" Nov 26 02:31:48 crc kubenswrapper[4766]: I1126 02:31:48.761306 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9gddf_2a218fd1-d674-44e8-97dd-fae53b3ae179/extract-content/0.log" Nov 26 02:31:48 crc kubenswrapper[4766]: I1126 02:31:48.812030 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9gddf_2a218fd1-d674-44e8-97dd-fae53b3ae179/extract-content/0.log" Nov 26 02:31:48 crc kubenswrapper[4766]: I1126 02:31:48.954060 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9gddf_2a218fd1-d674-44e8-97dd-fae53b3ae179/registry-server/0.log" Nov 26 02:31:48 crc kubenswrapper[4766]: I1126 02:31:48.986084 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9gddf_2a218fd1-d674-44e8-97dd-fae53b3ae179/extract-content/0.log" Nov 26 02:31:49 crc kubenswrapper[4766]: I1126 02:31:49.005128 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9gddf_2a218fd1-d674-44e8-97dd-fae53b3ae179/extract-utilities/0.log" Nov 26 02:31:49 crc kubenswrapper[4766]: I1126 02:31:49.042535 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pjtsc_34c7586e-58a4-4e97-b559-cb805d91e9d8/extract-utilities/0.log" Nov 26 02:31:49 crc kubenswrapper[4766]: I1126 02:31:49.188272 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pjtsc_34c7586e-58a4-4e97-b559-cb805d91e9d8/extract-utilities/0.log" Nov 26 02:31:49 crc kubenswrapper[4766]: I1126 02:31:49.200554 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pjtsc_34c7586e-58a4-4e97-b559-cb805d91e9d8/extract-content/0.log" Nov 26 02:31:49 crc kubenswrapper[4766]: I1126 02:31:49.231153 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pjtsc_34c7586e-58a4-4e97-b559-cb805d91e9d8/extract-content/0.log" Nov 26 02:31:49 crc kubenswrapper[4766]: I1126 02:31:49.366014 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pjtsc_34c7586e-58a4-4e97-b559-cb805d91e9d8/extract-content/0.log" Nov 26 02:31:49 crc kubenswrapper[4766]: I1126 02:31:49.394375 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pjtsc_34c7586e-58a4-4e97-b559-cb805d91e9d8/extract-utilities/0.log" Nov 26 02:31:50 crc kubenswrapper[4766]: I1126 02:31:50.483389 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pjtsc_34c7586e-58a4-4e97-b559-cb805d91e9d8/registry-server/0.log" Nov 26 02:31:56 crc kubenswrapper[4766]: I1126 02:31:56.511313 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9gddf" podUID="2a218fd1-d674-44e8-97dd-fae53b3ae179" containerName="registry-server" probeResult="failure" output=< Nov 26 02:31:56 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 02:31:56 crc kubenswrapper[4766]: > Nov 26 02:32:03 crc kubenswrapper[4766]: I1126 02:32:03.020821 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-bt977_bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70/prometheus-operator/0.log" Nov 26 02:32:03 crc kubenswrapper[4766]: I1126 02:32:03.145686 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs_1125f245-de34-4508-a4b0-2a468e73df36/prometheus-operator-admission-webhook/0.log" Nov 26 02:32:03 crc kubenswrapper[4766]: I1126 02:32:03.173124 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx_490105d8-e5f5-4fe9-89b9-d67ee8e07b87/prometheus-operator-admission-webhook/0.log" Nov 26 02:32:03 crc kubenswrapper[4766]: I1126 02:32:03.308032 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-b8qgs_a6dba801-0cee-4c4b-856c-db3be51c0ed9/operator/0.log" Nov 26 02:32:03 crc kubenswrapper[4766]: I1126 02:32:03.379016 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-ui-dashboards-7d5fb4cbfb-cmzfk_468c5f0c-30d9-4493-a87c-2f4482a3b4d3/observability-ui-dashboards/0.log" Nov 26 02:32:03 crc kubenswrapper[4766]: I1126 02:32:03.514384 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-twhqz_0be2b756-c941-49e3-b8f0-0977eaaf59d0/perses-operator/0.log" Nov 26 02:32:06 crc kubenswrapper[4766]: I1126 02:32:06.501448 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9gddf" podUID="2a218fd1-d674-44e8-97dd-fae53b3ae179" containerName="registry-server" probeResult="failure" output=< Nov 26 02:32:06 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 02:32:06 crc kubenswrapper[4766]: > Nov 26 02:32:11 crc kubenswrapper[4766]: I1126 02:32:11.479643 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:32:11 crc kubenswrapper[4766]: I1126 02:32:11.480497 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:32:15 crc kubenswrapper[4766]: I1126 02:32:15.509192 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9gddf" Nov 26 02:32:15 crc kubenswrapper[4766]: I1126 02:32:15.588189 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9gddf" Nov 26 02:32:15 crc kubenswrapper[4766]: I1126 02:32:15.760902 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9gddf"] Nov 26 02:32:16 crc kubenswrapper[4766]: I1126 02:32:16.964626 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9gddf" podUID="2a218fd1-d674-44e8-97dd-fae53b3ae179" containerName="registry-server" containerID="cri-o://102669f53cca04ea057cacf1845cf71cd5f172994ec8892c00f0918af24ae80b" gracePeriod=2 Nov 26 02:32:17 crc kubenswrapper[4766]: I1126 02:32:17.501872 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9gddf" Nov 26 02:32:17 crc kubenswrapper[4766]: I1126 02:32:17.599607 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-7bbf4564c5-m77vx_b90f7164-a5e9-452f-8733-f549f15e17ec/manager/0.log" Nov 26 02:32:17 crc kubenswrapper[4766]: I1126 02:32:17.634619 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-7bbf4564c5-m77vx_b90f7164-a5e9-452f-8733-f549f15e17ec/manager/1.log" Nov 26 02:32:17 crc kubenswrapper[4766]: I1126 02:32:17.638038 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-7bbf4564c5-m77vx_b90f7164-a5e9-452f-8733-f549f15e17ec/kube-rbac-proxy/0.log" Nov 26 02:32:17 crc kubenswrapper[4766]: I1126 02:32:17.644820 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a218fd1-d674-44e8-97dd-fae53b3ae179-utilities\") pod \"2a218fd1-d674-44e8-97dd-fae53b3ae179\" (UID: \"2a218fd1-d674-44e8-97dd-fae53b3ae179\") " Nov 26 02:32:17 crc kubenswrapper[4766]: I1126 02:32:17.646068 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a218fd1-d674-44e8-97dd-fae53b3ae179-utilities" (OuterVolumeSpecName: "utilities") pod "2a218fd1-d674-44e8-97dd-fae53b3ae179" (UID: "2a218fd1-d674-44e8-97dd-fae53b3ae179"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:32:17 crc kubenswrapper[4766]: I1126 02:32:17.646141 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a218fd1-d674-44e8-97dd-fae53b3ae179-catalog-content\") pod \"2a218fd1-d674-44e8-97dd-fae53b3ae179\" (UID: \"2a218fd1-d674-44e8-97dd-fae53b3ae179\") " Nov 26 02:32:17 crc kubenswrapper[4766]: I1126 02:32:17.646196 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cgjj\" (UniqueName: \"kubernetes.io/projected/2a218fd1-d674-44e8-97dd-fae53b3ae179-kube-api-access-4cgjj\") pod \"2a218fd1-d674-44e8-97dd-fae53b3ae179\" (UID: \"2a218fd1-d674-44e8-97dd-fae53b3ae179\") " Nov 26 02:32:17 crc kubenswrapper[4766]: I1126 02:32:17.647897 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a218fd1-d674-44e8-97dd-fae53b3ae179-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 02:32:17 crc kubenswrapper[4766]: I1126 02:32:17.653026 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a218fd1-d674-44e8-97dd-fae53b3ae179-kube-api-access-4cgjj" (OuterVolumeSpecName: "kube-api-access-4cgjj") pod "2a218fd1-d674-44e8-97dd-fae53b3ae179" (UID: "2a218fd1-d674-44e8-97dd-fae53b3ae179"). InnerVolumeSpecName "kube-api-access-4cgjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:32:17 crc kubenswrapper[4766]: I1126 02:32:17.738326 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a218fd1-d674-44e8-97dd-fae53b3ae179-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2a218fd1-d674-44e8-97dd-fae53b3ae179" (UID: "2a218fd1-d674-44e8-97dd-fae53b3ae179"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:32:17 crc kubenswrapper[4766]: I1126 02:32:17.751281 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a218fd1-d674-44e8-97dd-fae53b3ae179-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 02:32:17 crc kubenswrapper[4766]: I1126 02:32:17.751314 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cgjj\" (UniqueName: \"kubernetes.io/projected/2a218fd1-d674-44e8-97dd-fae53b3ae179-kube-api-access-4cgjj\") on node \"crc\" DevicePath \"\"" Nov 26 02:32:17 crc kubenswrapper[4766]: I1126 02:32:17.973859 4766 generic.go:334] "Generic (PLEG): container finished" podID="2a218fd1-d674-44e8-97dd-fae53b3ae179" containerID="102669f53cca04ea057cacf1845cf71cd5f172994ec8892c00f0918af24ae80b" exitCode=0 Nov 26 02:32:17 crc kubenswrapper[4766]: I1126 02:32:17.973904 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9gddf" event={"ID":"2a218fd1-d674-44e8-97dd-fae53b3ae179","Type":"ContainerDied","Data":"102669f53cca04ea057cacf1845cf71cd5f172994ec8892c00f0918af24ae80b"} Nov 26 02:32:17 crc kubenswrapper[4766]: I1126 02:32:17.974235 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9gddf" event={"ID":"2a218fd1-d674-44e8-97dd-fae53b3ae179","Type":"ContainerDied","Data":"4a1401dda042657c3ffdaa3ee74d3a22feb83858be9851d835f60ec6ffb4c034"} Nov 26 02:32:17 crc kubenswrapper[4766]: I1126 02:32:17.974257 4766 scope.go:117] "RemoveContainer" containerID="102669f53cca04ea057cacf1845cf71cd5f172994ec8892c00f0918af24ae80b" Nov 26 02:32:17 crc kubenswrapper[4766]: I1126 02:32:17.973964 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9gddf" Nov 26 02:32:18 crc kubenswrapper[4766]: I1126 02:32:18.004231 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9gddf"] Nov 26 02:32:18 crc kubenswrapper[4766]: I1126 02:32:18.006645 4766 scope.go:117] "RemoveContainer" containerID="c6e5511c31f7246fa7efbb516a68215e1a398304639a033f76c59431f675db41" Nov 26 02:32:18 crc kubenswrapper[4766]: I1126 02:32:18.015500 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9gddf"] Nov 26 02:32:18 crc kubenswrapper[4766]: I1126 02:32:18.037257 4766 scope.go:117] "RemoveContainer" containerID="374867ec51df9b8d6e9a945f072cac226f8fdb62b6191207f5004ddcf69327eb" Nov 26 02:32:18 crc kubenswrapper[4766]: I1126 02:32:18.083504 4766 scope.go:117] "RemoveContainer" containerID="102669f53cca04ea057cacf1845cf71cd5f172994ec8892c00f0918af24ae80b" Nov 26 02:32:18 crc kubenswrapper[4766]: E1126 02:32:18.085971 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"102669f53cca04ea057cacf1845cf71cd5f172994ec8892c00f0918af24ae80b\": container with ID starting with 102669f53cca04ea057cacf1845cf71cd5f172994ec8892c00f0918af24ae80b not found: ID does not exist" containerID="102669f53cca04ea057cacf1845cf71cd5f172994ec8892c00f0918af24ae80b" Nov 26 02:32:18 crc kubenswrapper[4766]: I1126 02:32:18.086022 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"102669f53cca04ea057cacf1845cf71cd5f172994ec8892c00f0918af24ae80b"} err="failed to get container status \"102669f53cca04ea057cacf1845cf71cd5f172994ec8892c00f0918af24ae80b\": rpc error: code = NotFound desc = could not find container \"102669f53cca04ea057cacf1845cf71cd5f172994ec8892c00f0918af24ae80b\": container with ID starting with 102669f53cca04ea057cacf1845cf71cd5f172994ec8892c00f0918af24ae80b not found: ID does not exist" Nov 26 02:32:18 crc kubenswrapper[4766]: I1126 02:32:18.086054 4766 scope.go:117] "RemoveContainer" containerID="c6e5511c31f7246fa7efbb516a68215e1a398304639a033f76c59431f675db41" Nov 26 02:32:18 crc kubenswrapper[4766]: E1126 02:32:18.086545 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6e5511c31f7246fa7efbb516a68215e1a398304639a033f76c59431f675db41\": container with ID starting with c6e5511c31f7246fa7efbb516a68215e1a398304639a033f76c59431f675db41 not found: ID does not exist" containerID="c6e5511c31f7246fa7efbb516a68215e1a398304639a033f76c59431f675db41" Nov 26 02:32:18 crc kubenswrapper[4766]: I1126 02:32:18.086612 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6e5511c31f7246fa7efbb516a68215e1a398304639a033f76c59431f675db41"} err="failed to get container status \"c6e5511c31f7246fa7efbb516a68215e1a398304639a033f76c59431f675db41\": rpc error: code = NotFound desc = could not find container \"c6e5511c31f7246fa7efbb516a68215e1a398304639a033f76c59431f675db41\": container with ID starting with c6e5511c31f7246fa7efbb516a68215e1a398304639a033f76c59431f675db41 not found: ID does not exist" Nov 26 02:32:18 crc kubenswrapper[4766]: I1126 02:32:18.086676 4766 scope.go:117] "RemoveContainer" containerID="374867ec51df9b8d6e9a945f072cac226f8fdb62b6191207f5004ddcf69327eb" Nov 26 02:32:18 crc kubenswrapper[4766]: E1126 02:32:18.087155 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"374867ec51df9b8d6e9a945f072cac226f8fdb62b6191207f5004ddcf69327eb\": container with ID starting with 374867ec51df9b8d6e9a945f072cac226f8fdb62b6191207f5004ddcf69327eb not found: ID does not exist" containerID="374867ec51df9b8d6e9a945f072cac226f8fdb62b6191207f5004ddcf69327eb" Nov 26 02:32:18 crc kubenswrapper[4766]: I1126 02:32:18.087182 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"374867ec51df9b8d6e9a945f072cac226f8fdb62b6191207f5004ddcf69327eb"} err="failed to get container status \"374867ec51df9b8d6e9a945f072cac226f8fdb62b6191207f5004ddcf69327eb\": rpc error: code = NotFound desc = could not find container \"374867ec51df9b8d6e9a945f072cac226f8fdb62b6191207f5004ddcf69327eb\": container with ID starting with 374867ec51df9b8d6e9a945f072cac226f8fdb62b6191207f5004ddcf69327eb not found: ID does not exist" Nov 26 02:32:19 crc kubenswrapper[4766]: I1126 02:32:19.849853 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a218fd1-d674-44e8-97dd-fae53b3ae179" path="/var/lib/kubelet/pods/2a218fd1-d674-44e8-97dd-fae53b3ae179/volumes" Nov 26 02:32:36 crc kubenswrapper[4766]: E1126 02:32:36.635714 4766 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.32:56490->38.102.83.32:36993: read tcp 38.102.83.32:56490->38.102.83.32:36993: read: connection reset by peer Nov 26 02:32:41 crc kubenswrapper[4766]: I1126 02:32:41.479113 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:32:41 crc kubenswrapper[4766]: I1126 02:32:41.479702 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:33:06 crc kubenswrapper[4766]: I1126 02:33:06.769854 4766 scope.go:117] "RemoveContainer" containerID="50ec60584e0e9b18b2c90f62dc5ed056feb979841a2b8cbf242dc1da347c1130" Nov 26 02:33:11 crc kubenswrapper[4766]: I1126 02:33:11.480273 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:33:11 crc kubenswrapper[4766]: I1126 02:33:11.480985 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:33:11 crc kubenswrapper[4766]: I1126 02:33:11.481061 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 02:33:11 crc kubenswrapper[4766]: I1126 02:33:11.482243 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 02:33:11 crc kubenswrapper[4766]: I1126 02:33:11.482348 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" gracePeriod=600 Nov 26 02:33:11 crc kubenswrapper[4766]: E1126 02:33:11.609761 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:33:11 crc kubenswrapper[4766]: I1126 02:33:11.649266 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" exitCode=0 Nov 26 02:33:11 crc kubenswrapper[4766]: I1126 02:33:11.649334 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a"} Nov 26 02:33:11 crc kubenswrapper[4766]: I1126 02:33:11.649384 4766 scope.go:117] "RemoveContainer" containerID="e12146d79b16ca01dd3ddaff20fdb9b94b5dd64df0a00c9f78979d3aecbe9802" Nov 26 02:33:11 crc kubenswrapper[4766]: I1126 02:33:11.650117 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:33:11 crc kubenswrapper[4766]: E1126 02:33:11.650441 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:33:24 crc kubenswrapper[4766]: I1126 02:33:24.826733 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:33:24 crc kubenswrapper[4766]: E1126 02:33:24.827799 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:33:38 crc kubenswrapper[4766]: I1126 02:33:38.826522 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:33:38 crc kubenswrapper[4766]: E1126 02:33:38.827286 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:33:50 crc kubenswrapper[4766]: I1126 02:33:50.827329 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:33:50 crc kubenswrapper[4766]: E1126 02:33:50.828418 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:34:01 crc kubenswrapper[4766]: I1126 02:34:01.827587 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:34:01 crc kubenswrapper[4766]: E1126 02:34:01.828750 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:34:06 crc kubenswrapper[4766]: I1126 02:34:06.906370 4766 scope.go:117] "RemoveContainer" containerID="62a84cd8dcfadcb8288b1512069ccb0620c98789b17e318ef701fa58a25fd935" Nov 26 02:34:13 crc kubenswrapper[4766]: I1126 02:34:13.827041 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:34:13 crc kubenswrapper[4766]: E1126 02:34:13.828788 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:34:14 crc kubenswrapper[4766]: I1126 02:34:14.875309 4766 generic.go:334] "Generic (PLEG): container finished" podID="1e340437-6f94-4ea6-82c5-bed5ac1b9c71" containerID="ae717b377a1717d9d05706682945c3e64ebb7ca8aecf7b9a4ee91bbb6e288daf" exitCode=0 Nov 26 02:34:14 crc kubenswrapper[4766]: I1126 02:34:14.875444 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv2fj/must-gather-sfxwh" event={"ID":"1e340437-6f94-4ea6-82c5-bed5ac1b9c71","Type":"ContainerDied","Data":"ae717b377a1717d9d05706682945c3e64ebb7ca8aecf7b9a4ee91bbb6e288daf"} Nov 26 02:34:14 crc kubenswrapper[4766]: I1126 02:34:14.876411 4766 scope.go:117] "RemoveContainer" containerID="ae717b377a1717d9d05706682945c3e64ebb7ca8aecf7b9a4ee91bbb6e288daf" Nov 26 02:34:15 crc kubenswrapper[4766]: I1126 02:34:15.065278 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-tv2fj_must-gather-sfxwh_1e340437-6f94-4ea6-82c5-bed5ac1b9c71/gather/0.log" Nov 26 02:34:23 crc kubenswrapper[4766]: I1126 02:34:23.350465 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tv2fj/must-gather-sfxwh"] Nov 26 02:34:23 crc kubenswrapper[4766]: I1126 02:34:23.351149 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-tv2fj/must-gather-sfxwh" podUID="1e340437-6f94-4ea6-82c5-bed5ac1b9c71" containerName="copy" containerID="cri-o://7836e1c1d70e7eff8c083085cf73a68a29333eb8b481ac700911467164edc9ee" gracePeriod=2 Nov 26 02:34:23 crc kubenswrapper[4766]: I1126 02:34:23.365402 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tv2fj/must-gather-sfxwh"] Nov 26 02:34:23 crc kubenswrapper[4766]: I1126 02:34:23.917269 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-tv2fj_must-gather-sfxwh_1e340437-6f94-4ea6-82c5-bed5ac1b9c71/copy/0.log" Nov 26 02:34:23 crc kubenswrapper[4766]: I1126 02:34:23.917975 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv2fj/must-gather-sfxwh" Nov 26 02:34:23 crc kubenswrapper[4766]: I1126 02:34:23.982976 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbg78\" (UniqueName: \"kubernetes.io/projected/1e340437-6f94-4ea6-82c5-bed5ac1b9c71-kube-api-access-gbg78\") pod \"1e340437-6f94-4ea6-82c5-bed5ac1b9c71\" (UID: \"1e340437-6f94-4ea6-82c5-bed5ac1b9c71\") " Nov 26 02:34:23 crc kubenswrapper[4766]: I1126 02:34:23.983217 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1e340437-6f94-4ea6-82c5-bed5ac1b9c71-must-gather-output\") pod \"1e340437-6f94-4ea6-82c5-bed5ac1b9c71\" (UID: \"1e340437-6f94-4ea6-82c5-bed5ac1b9c71\") " Nov 26 02:34:23 crc kubenswrapper[4766]: I1126 02:34:23.990918 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e340437-6f94-4ea6-82c5-bed5ac1b9c71-kube-api-access-gbg78" (OuterVolumeSpecName: "kube-api-access-gbg78") pod "1e340437-6f94-4ea6-82c5-bed5ac1b9c71" (UID: "1e340437-6f94-4ea6-82c5-bed5ac1b9c71"). InnerVolumeSpecName "kube-api-access-gbg78". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:34:24 crc kubenswrapper[4766]: I1126 02:34:24.002027 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-tv2fj_must-gather-sfxwh_1e340437-6f94-4ea6-82c5-bed5ac1b9c71/copy/0.log" Nov 26 02:34:24 crc kubenswrapper[4766]: I1126 02:34:24.002489 4766 generic.go:334] "Generic (PLEG): container finished" podID="1e340437-6f94-4ea6-82c5-bed5ac1b9c71" containerID="7836e1c1d70e7eff8c083085cf73a68a29333eb8b481ac700911467164edc9ee" exitCode=143 Nov 26 02:34:24 crc kubenswrapper[4766]: I1126 02:34:24.002533 4766 scope.go:117] "RemoveContainer" containerID="7836e1c1d70e7eff8c083085cf73a68a29333eb8b481ac700911467164edc9ee" Nov 26 02:34:24 crc kubenswrapper[4766]: I1126 02:34:24.002695 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv2fj/must-gather-sfxwh" Nov 26 02:34:24 crc kubenswrapper[4766]: I1126 02:34:24.065756 4766 scope.go:117] "RemoveContainer" containerID="ae717b377a1717d9d05706682945c3e64ebb7ca8aecf7b9a4ee91bbb6e288daf" Nov 26 02:34:24 crc kubenswrapper[4766]: I1126 02:34:24.086818 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbg78\" (UniqueName: \"kubernetes.io/projected/1e340437-6f94-4ea6-82c5-bed5ac1b9c71-kube-api-access-gbg78\") on node \"crc\" DevicePath \"\"" Nov 26 02:34:24 crc kubenswrapper[4766]: I1126 02:34:24.135476 4766 scope.go:117] "RemoveContainer" containerID="7836e1c1d70e7eff8c083085cf73a68a29333eb8b481ac700911467164edc9ee" Nov 26 02:34:24 crc kubenswrapper[4766]: E1126 02:34:24.136071 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7836e1c1d70e7eff8c083085cf73a68a29333eb8b481ac700911467164edc9ee\": container with ID starting with 7836e1c1d70e7eff8c083085cf73a68a29333eb8b481ac700911467164edc9ee not found: ID does not exist" containerID="7836e1c1d70e7eff8c083085cf73a68a29333eb8b481ac700911467164edc9ee" Nov 26 02:34:24 crc kubenswrapper[4766]: I1126 02:34:24.136128 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7836e1c1d70e7eff8c083085cf73a68a29333eb8b481ac700911467164edc9ee"} err="failed to get container status \"7836e1c1d70e7eff8c083085cf73a68a29333eb8b481ac700911467164edc9ee\": rpc error: code = NotFound desc = could not find container \"7836e1c1d70e7eff8c083085cf73a68a29333eb8b481ac700911467164edc9ee\": container with ID starting with 7836e1c1d70e7eff8c083085cf73a68a29333eb8b481ac700911467164edc9ee not found: ID does not exist" Nov 26 02:34:24 crc kubenswrapper[4766]: I1126 02:34:24.136162 4766 scope.go:117] "RemoveContainer" containerID="ae717b377a1717d9d05706682945c3e64ebb7ca8aecf7b9a4ee91bbb6e288daf" Nov 26 02:34:24 crc kubenswrapper[4766]: E1126 02:34:24.136867 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae717b377a1717d9d05706682945c3e64ebb7ca8aecf7b9a4ee91bbb6e288daf\": container with ID starting with ae717b377a1717d9d05706682945c3e64ebb7ca8aecf7b9a4ee91bbb6e288daf not found: ID does not exist" containerID="ae717b377a1717d9d05706682945c3e64ebb7ca8aecf7b9a4ee91bbb6e288daf" Nov 26 02:34:24 crc kubenswrapper[4766]: I1126 02:34:24.136909 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae717b377a1717d9d05706682945c3e64ebb7ca8aecf7b9a4ee91bbb6e288daf"} err="failed to get container status \"ae717b377a1717d9d05706682945c3e64ebb7ca8aecf7b9a4ee91bbb6e288daf\": rpc error: code = NotFound desc = could not find container \"ae717b377a1717d9d05706682945c3e64ebb7ca8aecf7b9a4ee91bbb6e288daf\": container with ID starting with ae717b377a1717d9d05706682945c3e64ebb7ca8aecf7b9a4ee91bbb6e288daf not found: ID does not exist" Nov 26 02:34:24 crc kubenswrapper[4766]: I1126 02:34:24.174256 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e340437-6f94-4ea6-82c5-bed5ac1b9c71-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "1e340437-6f94-4ea6-82c5-bed5ac1b9c71" (UID: "1e340437-6f94-4ea6-82c5-bed5ac1b9c71"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:34:24 crc kubenswrapper[4766]: I1126 02:34:24.189334 4766 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1e340437-6f94-4ea6-82c5-bed5ac1b9c71-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 26 02:34:24 crc kubenswrapper[4766]: I1126 02:34:24.826765 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:34:24 crc kubenswrapper[4766]: E1126 02:34:24.827361 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:34:25 crc kubenswrapper[4766]: I1126 02:34:25.837290 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e340437-6f94-4ea6-82c5-bed5ac1b9c71" path="/var/lib/kubelet/pods/1e340437-6f94-4ea6-82c5-bed5ac1b9c71/volumes" Nov 26 02:34:39 crc kubenswrapper[4766]: I1126 02:34:39.836808 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:34:39 crc kubenswrapper[4766]: E1126 02:34:39.837610 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:34:42 crc kubenswrapper[4766]: I1126 02:34:42.163216 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hbjbv"] Nov 26 02:34:42 crc kubenswrapper[4766]: E1126 02:34:42.164352 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e340437-6f94-4ea6-82c5-bed5ac1b9c71" containerName="copy" Nov 26 02:34:42 crc kubenswrapper[4766]: I1126 02:34:42.164370 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e340437-6f94-4ea6-82c5-bed5ac1b9c71" containerName="copy" Nov 26 02:34:42 crc kubenswrapper[4766]: E1126 02:34:42.164391 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e340437-6f94-4ea6-82c5-bed5ac1b9c71" containerName="gather" Nov 26 02:34:42 crc kubenswrapper[4766]: I1126 02:34:42.164399 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e340437-6f94-4ea6-82c5-bed5ac1b9c71" containerName="gather" Nov 26 02:34:42 crc kubenswrapper[4766]: E1126 02:34:42.164417 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a218fd1-d674-44e8-97dd-fae53b3ae179" containerName="extract-content" Nov 26 02:34:42 crc kubenswrapper[4766]: I1126 02:34:42.164426 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a218fd1-d674-44e8-97dd-fae53b3ae179" containerName="extract-content" Nov 26 02:34:42 crc kubenswrapper[4766]: E1126 02:34:42.164434 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a218fd1-d674-44e8-97dd-fae53b3ae179" containerName="registry-server" Nov 26 02:34:42 crc kubenswrapper[4766]: I1126 02:34:42.164441 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a218fd1-d674-44e8-97dd-fae53b3ae179" containerName="registry-server" Nov 26 02:34:42 crc kubenswrapper[4766]: E1126 02:34:42.164482 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a218fd1-d674-44e8-97dd-fae53b3ae179" containerName="extract-utilities" Nov 26 02:34:42 crc kubenswrapper[4766]: I1126 02:34:42.164490 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a218fd1-d674-44e8-97dd-fae53b3ae179" containerName="extract-utilities" Nov 26 02:34:42 crc kubenswrapper[4766]: I1126 02:34:42.164767 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e340437-6f94-4ea6-82c5-bed5ac1b9c71" containerName="gather" Nov 26 02:34:42 crc kubenswrapper[4766]: I1126 02:34:42.164787 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a218fd1-d674-44e8-97dd-fae53b3ae179" containerName="registry-server" Nov 26 02:34:42 crc kubenswrapper[4766]: I1126 02:34:42.164800 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e340437-6f94-4ea6-82c5-bed5ac1b9c71" containerName="copy" Nov 26 02:34:42 crc kubenswrapper[4766]: I1126 02:34:42.167146 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hbjbv" Nov 26 02:34:42 crc kubenswrapper[4766]: I1126 02:34:42.184981 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hbjbv"] Nov 26 02:34:42 crc kubenswrapper[4766]: I1126 02:34:42.269064 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0eff4d6c-f2b8-493b-a830-3ebbc5854eea-utilities\") pod \"certified-operators-hbjbv\" (UID: \"0eff4d6c-f2b8-493b-a830-3ebbc5854eea\") " pod="openshift-marketplace/certified-operators-hbjbv" Nov 26 02:34:42 crc kubenswrapper[4766]: I1126 02:34:42.269356 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0eff4d6c-f2b8-493b-a830-3ebbc5854eea-catalog-content\") pod \"certified-operators-hbjbv\" (UID: \"0eff4d6c-f2b8-493b-a830-3ebbc5854eea\") " pod="openshift-marketplace/certified-operators-hbjbv" Nov 26 02:34:42 crc kubenswrapper[4766]: I1126 02:34:42.269435 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzss5\" (UniqueName: \"kubernetes.io/projected/0eff4d6c-f2b8-493b-a830-3ebbc5854eea-kube-api-access-pzss5\") pod \"certified-operators-hbjbv\" (UID: \"0eff4d6c-f2b8-493b-a830-3ebbc5854eea\") " pod="openshift-marketplace/certified-operators-hbjbv" Nov 26 02:34:42 crc kubenswrapper[4766]: I1126 02:34:42.371541 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0eff4d6c-f2b8-493b-a830-3ebbc5854eea-catalog-content\") pod \"certified-operators-hbjbv\" (UID: \"0eff4d6c-f2b8-493b-a830-3ebbc5854eea\") " pod="openshift-marketplace/certified-operators-hbjbv" Nov 26 02:34:42 crc kubenswrapper[4766]: I1126 02:34:42.371660 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzss5\" (UniqueName: \"kubernetes.io/projected/0eff4d6c-f2b8-493b-a830-3ebbc5854eea-kube-api-access-pzss5\") pod \"certified-operators-hbjbv\" (UID: \"0eff4d6c-f2b8-493b-a830-3ebbc5854eea\") " pod="openshift-marketplace/certified-operators-hbjbv" Nov 26 02:34:42 crc kubenswrapper[4766]: I1126 02:34:42.371727 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0eff4d6c-f2b8-493b-a830-3ebbc5854eea-utilities\") pod \"certified-operators-hbjbv\" (UID: \"0eff4d6c-f2b8-493b-a830-3ebbc5854eea\") " pod="openshift-marketplace/certified-operators-hbjbv" Nov 26 02:34:42 crc kubenswrapper[4766]: I1126 02:34:42.372249 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0eff4d6c-f2b8-493b-a830-3ebbc5854eea-utilities\") pod \"certified-operators-hbjbv\" (UID: \"0eff4d6c-f2b8-493b-a830-3ebbc5854eea\") " pod="openshift-marketplace/certified-operators-hbjbv" Nov 26 02:34:42 crc kubenswrapper[4766]: I1126 02:34:42.372544 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0eff4d6c-f2b8-493b-a830-3ebbc5854eea-catalog-content\") pod \"certified-operators-hbjbv\" (UID: \"0eff4d6c-f2b8-493b-a830-3ebbc5854eea\") " pod="openshift-marketplace/certified-operators-hbjbv" Nov 26 02:34:42 crc kubenswrapper[4766]: I1126 02:34:42.397187 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzss5\" (UniqueName: \"kubernetes.io/projected/0eff4d6c-f2b8-493b-a830-3ebbc5854eea-kube-api-access-pzss5\") pod \"certified-operators-hbjbv\" (UID: \"0eff4d6c-f2b8-493b-a830-3ebbc5854eea\") " pod="openshift-marketplace/certified-operators-hbjbv" Nov 26 02:34:42 crc kubenswrapper[4766]: I1126 02:34:42.503875 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hbjbv" Nov 26 02:34:43 crc kubenswrapper[4766]: I1126 02:34:43.073508 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hbjbv"] Nov 26 02:34:43 crc kubenswrapper[4766]: I1126 02:34:43.233062 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbjbv" event={"ID":"0eff4d6c-f2b8-493b-a830-3ebbc5854eea","Type":"ContainerStarted","Data":"85e8fec5f8ab3874010c241f8bd972265a8c7ba525ad1eda2a1ff848f5e72292"} Nov 26 02:34:44 crc kubenswrapper[4766]: I1126 02:34:44.251523 4766 generic.go:334] "Generic (PLEG): container finished" podID="0eff4d6c-f2b8-493b-a830-3ebbc5854eea" containerID="4b683313ddaf5a330416b8bdf43dee1e2ac479b33f0bc46dc6898ca2a4028b4b" exitCode=0 Nov 26 02:34:44 crc kubenswrapper[4766]: I1126 02:34:44.251614 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbjbv" event={"ID":"0eff4d6c-f2b8-493b-a830-3ebbc5854eea","Type":"ContainerDied","Data":"4b683313ddaf5a330416b8bdf43dee1e2ac479b33f0bc46dc6898ca2a4028b4b"} Nov 26 02:34:45 crc kubenswrapper[4766]: I1126 02:34:45.293207 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbjbv" event={"ID":"0eff4d6c-f2b8-493b-a830-3ebbc5854eea","Type":"ContainerStarted","Data":"6d9f6e9757d90f17eba19feba3f6801c06dca432db823fa085eda6913b44d04a"} Nov 26 02:34:46 crc kubenswrapper[4766]: I1126 02:34:46.308049 4766 generic.go:334] "Generic (PLEG): container finished" podID="0eff4d6c-f2b8-493b-a830-3ebbc5854eea" containerID="6d9f6e9757d90f17eba19feba3f6801c06dca432db823fa085eda6913b44d04a" exitCode=0 Nov 26 02:34:46 crc kubenswrapper[4766]: I1126 02:34:46.308313 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbjbv" event={"ID":"0eff4d6c-f2b8-493b-a830-3ebbc5854eea","Type":"ContainerDied","Data":"6d9f6e9757d90f17eba19feba3f6801c06dca432db823fa085eda6913b44d04a"} Nov 26 02:34:47 crc kubenswrapper[4766]: I1126 02:34:47.322746 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbjbv" event={"ID":"0eff4d6c-f2b8-493b-a830-3ebbc5854eea","Type":"ContainerStarted","Data":"065d0d3a755ab9e468d894c736971bc9103af510ea09537839d2185983023261"} Nov 26 02:34:47 crc kubenswrapper[4766]: I1126 02:34:47.349290 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hbjbv" podStartSLOduration=2.831569152 podStartE2EDuration="5.349269789s" podCreationTimestamp="2025-11-26 02:34:42 +0000 UTC" firstStartedPulling="2025-11-26 02:34:44.254909683 +0000 UTC m=+7865.103680153" lastFinishedPulling="2025-11-26 02:34:46.77261036 +0000 UTC m=+7867.621380790" observedRunningTime="2025-11-26 02:34:47.345889076 +0000 UTC m=+7868.194659516" watchObservedRunningTime="2025-11-26 02:34:47.349269789 +0000 UTC m=+7868.198040219" Nov 26 02:34:52 crc kubenswrapper[4766]: I1126 02:34:52.504075 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hbjbv" Nov 26 02:34:52 crc kubenswrapper[4766]: I1126 02:34:52.504554 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hbjbv" Nov 26 02:34:52 crc kubenswrapper[4766]: I1126 02:34:52.580645 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hbjbv" Nov 26 02:34:53 crc kubenswrapper[4766]: I1126 02:34:53.452896 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hbjbv" Nov 26 02:34:53 crc kubenswrapper[4766]: I1126 02:34:53.547984 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hbjbv"] Nov 26 02:34:53 crc kubenswrapper[4766]: I1126 02:34:53.826799 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:34:53 crc kubenswrapper[4766]: E1126 02:34:53.827167 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:34:55 crc kubenswrapper[4766]: I1126 02:34:55.405323 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hbjbv" podUID="0eff4d6c-f2b8-493b-a830-3ebbc5854eea" containerName="registry-server" containerID="cri-o://065d0d3a755ab9e468d894c736971bc9103af510ea09537839d2185983023261" gracePeriod=2 Nov 26 02:34:56 crc kubenswrapper[4766]: I1126 02:34:56.420439 4766 generic.go:334] "Generic (PLEG): container finished" podID="0eff4d6c-f2b8-493b-a830-3ebbc5854eea" containerID="065d0d3a755ab9e468d894c736971bc9103af510ea09537839d2185983023261" exitCode=0 Nov 26 02:34:56 crc kubenswrapper[4766]: I1126 02:34:56.420797 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbjbv" event={"ID":"0eff4d6c-f2b8-493b-a830-3ebbc5854eea","Type":"ContainerDied","Data":"065d0d3a755ab9e468d894c736971bc9103af510ea09537839d2185983023261"} Nov 26 02:34:56 crc kubenswrapper[4766]: I1126 02:34:56.421122 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbjbv" event={"ID":"0eff4d6c-f2b8-493b-a830-3ebbc5854eea","Type":"ContainerDied","Data":"85e8fec5f8ab3874010c241f8bd972265a8c7ba525ad1eda2a1ff848f5e72292"} Nov 26 02:34:56 crc kubenswrapper[4766]: I1126 02:34:56.421138 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85e8fec5f8ab3874010c241f8bd972265a8c7ba525ad1eda2a1ff848f5e72292" Nov 26 02:34:56 crc kubenswrapper[4766]: I1126 02:34:56.536742 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hbjbv" Nov 26 02:34:56 crc kubenswrapper[4766]: I1126 02:34:56.679271 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0eff4d6c-f2b8-493b-a830-3ebbc5854eea-utilities\") pod \"0eff4d6c-f2b8-493b-a830-3ebbc5854eea\" (UID: \"0eff4d6c-f2b8-493b-a830-3ebbc5854eea\") " Nov 26 02:34:56 crc kubenswrapper[4766]: I1126 02:34:56.679540 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0eff4d6c-f2b8-493b-a830-3ebbc5854eea-catalog-content\") pod \"0eff4d6c-f2b8-493b-a830-3ebbc5854eea\" (UID: \"0eff4d6c-f2b8-493b-a830-3ebbc5854eea\") " Nov 26 02:34:56 crc kubenswrapper[4766]: I1126 02:34:56.679674 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzss5\" (UniqueName: \"kubernetes.io/projected/0eff4d6c-f2b8-493b-a830-3ebbc5854eea-kube-api-access-pzss5\") pod \"0eff4d6c-f2b8-493b-a830-3ebbc5854eea\" (UID: \"0eff4d6c-f2b8-493b-a830-3ebbc5854eea\") " Nov 26 02:34:56 crc kubenswrapper[4766]: I1126 02:34:56.680771 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0eff4d6c-f2b8-493b-a830-3ebbc5854eea-utilities" (OuterVolumeSpecName: "utilities") pod "0eff4d6c-f2b8-493b-a830-3ebbc5854eea" (UID: "0eff4d6c-f2b8-493b-a830-3ebbc5854eea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:34:56 crc kubenswrapper[4766]: I1126 02:34:56.685447 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0eff4d6c-f2b8-493b-a830-3ebbc5854eea-kube-api-access-pzss5" (OuterVolumeSpecName: "kube-api-access-pzss5") pod "0eff4d6c-f2b8-493b-a830-3ebbc5854eea" (UID: "0eff4d6c-f2b8-493b-a830-3ebbc5854eea"). InnerVolumeSpecName "kube-api-access-pzss5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:34:56 crc kubenswrapper[4766]: I1126 02:34:56.733985 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0eff4d6c-f2b8-493b-a830-3ebbc5854eea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0eff4d6c-f2b8-493b-a830-3ebbc5854eea" (UID: "0eff4d6c-f2b8-493b-a830-3ebbc5854eea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:34:56 crc kubenswrapper[4766]: I1126 02:34:56.781675 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0eff4d6c-f2b8-493b-a830-3ebbc5854eea-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 02:34:56 crc kubenswrapper[4766]: I1126 02:34:56.781710 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzss5\" (UniqueName: \"kubernetes.io/projected/0eff4d6c-f2b8-493b-a830-3ebbc5854eea-kube-api-access-pzss5\") on node \"crc\" DevicePath \"\"" Nov 26 02:34:56 crc kubenswrapper[4766]: I1126 02:34:56.781723 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0eff4d6c-f2b8-493b-a830-3ebbc5854eea-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 02:34:57 crc kubenswrapper[4766]: I1126 02:34:57.433560 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hbjbv" Nov 26 02:34:57 crc kubenswrapper[4766]: I1126 02:34:57.489631 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hbjbv"] Nov 26 02:34:57 crc kubenswrapper[4766]: I1126 02:34:57.507537 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hbjbv"] Nov 26 02:34:57 crc kubenswrapper[4766]: I1126 02:34:57.860074 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0eff4d6c-f2b8-493b-a830-3ebbc5854eea" path="/var/lib/kubelet/pods/0eff4d6c-f2b8-493b-a830-3ebbc5854eea/volumes" Nov 26 02:35:08 crc kubenswrapper[4766]: I1126 02:35:08.827167 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:35:08 crc kubenswrapper[4766]: E1126 02:35:08.828201 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:35:20 crc kubenswrapper[4766]: I1126 02:35:20.828780 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:35:20 crc kubenswrapper[4766]: E1126 02:35:20.829983 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:35:34 crc kubenswrapper[4766]: I1126 02:35:34.826731 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:35:34 crc kubenswrapper[4766]: E1126 02:35:34.827677 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:35:44 crc kubenswrapper[4766]: I1126 02:35:44.070209 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-q8z4k"] Nov 26 02:35:44 crc kubenswrapper[4766]: E1126 02:35:44.071032 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0eff4d6c-f2b8-493b-a830-3ebbc5854eea" containerName="registry-server" Nov 26 02:35:44 crc kubenswrapper[4766]: I1126 02:35:44.071045 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0eff4d6c-f2b8-493b-a830-3ebbc5854eea" containerName="registry-server" Nov 26 02:35:44 crc kubenswrapper[4766]: E1126 02:35:44.071062 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0eff4d6c-f2b8-493b-a830-3ebbc5854eea" containerName="extract-content" Nov 26 02:35:44 crc kubenswrapper[4766]: I1126 02:35:44.071068 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0eff4d6c-f2b8-493b-a830-3ebbc5854eea" containerName="extract-content" Nov 26 02:35:44 crc kubenswrapper[4766]: E1126 02:35:44.071099 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0eff4d6c-f2b8-493b-a830-3ebbc5854eea" containerName="extract-utilities" Nov 26 02:35:44 crc kubenswrapper[4766]: I1126 02:35:44.071107 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="0eff4d6c-f2b8-493b-a830-3ebbc5854eea" containerName="extract-utilities" Nov 26 02:35:44 crc kubenswrapper[4766]: I1126 02:35:44.071323 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="0eff4d6c-f2b8-493b-a830-3ebbc5854eea" containerName="registry-server" Nov 26 02:35:44 crc kubenswrapper[4766]: I1126 02:35:44.073000 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q8z4k" Nov 26 02:35:44 crc kubenswrapper[4766]: I1126 02:35:44.093394 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q8z4k"] Nov 26 02:35:44 crc kubenswrapper[4766]: I1126 02:35:44.169123 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30b17774-f46f-4d8e-8211-56148141054a-utilities\") pod \"community-operators-q8z4k\" (UID: \"30b17774-f46f-4d8e-8211-56148141054a\") " pod="openshift-marketplace/community-operators-q8z4k" Nov 26 02:35:44 crc kubenswrapper[4766]: I1126 02:35:44.169208 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30b17774-f46f-4d8e-8211-56148141054a-catalog-content\") pod \"community-operators-q8z4k\" (UID: \"30b17774-f46f-4d8e-8211-56148141054a\") " pod="openshift-marketplace/community-operators-q8z4k" Nov 26 02:35:44 crc kubenswrapper[4766]: I1126 02:35:44.169320 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqj57\" (UniqueName: \"kubernetes.io/projected/30b17774-f46f-4d8e-8211-56148141054a-kube-api-access-hqj57\") pod \"community-operators-q8z4k\" (UID: \"30b17774-f46f-4d8e-8211-56148141054a\") " pod="openshift-marketplace/community-operators-q8z4k" Nov 26 02:35:44 crc kubenswrapper[4766]: I1126 02:35:44.270734 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqj57\" (UniqueName: \"kubernetes.io/projected/30b17774-f46f-4d8e-8211-56148141054a-kube-api-access-hqj57\") pod \"community-operators-q8z4k\" (UID: \"30b17774-f46f-4d8e-8211-56148141054a\") " pod="openshift-marketplace/community-operators-q8z4k" Nov 26 02:35:44 crc kubenswrapper[4766]: I1126 02:35:44.270932 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30b17774-f46f-4d8e-8211-56148141054a-utilities\") pod \"community-operators-q8z4k\" (UID: \"30b17774-f46f-4d8e-8211-56148141054a\") " pod="openshift-marketplace/community-operators-q8z4k" Nov 26 02:35:44 crc kubenswrapper[4766]: I1126 02:35:44.270971 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30b17774-f46f-4d8e-8211-56148141054a-catalog-content\") pod \"community-operators-q8z4k\" (UID: \"30b17774-f46f-4d8e-8211-56148141054a\") " pod="openshift-marketplace/community-operators-q8z4k" Nov 26 02:35:44 crc kubenswrapper[4766]: I1126 02:35:44.271450 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30b17774-f46f-4d8e-8211-56148141054a-catalog-content\") pod \"community-operators-q8z4k\" (UID: \"30b17774-f46f-4d8e-8211-56148141054a\") " pod="openshift-marketplace/community-operators-q8z4k" Nov 26 02:35:44 crc kubenswrapper[4766]: I1126 02:35:44.271525 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30b17774-f46f-4d8e-8211-56148141054a-utilities\") pod \"community-operators-q8z4k\" (UID: \"30b17774-f46f-4d8e-8211-56148141054a\") " pod="openshift-marketplace/community-operators-q8z4k" Nov 26 02:35:44 crc kubenswrapper[4766]: I1126 02:35:44.292498 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqj57\" (UniqueName: \"kubernetes.io/projected/30b17774-f46f-4d8e-8211-56148141054a-kube-api-access-hqj57\") pod \"community-operators-q8z4k\" (UID: \"30b17774-f46f-4d8e-8211-56148141054a\") " pod="openshift-marketplace/community-operators-q8z4k" Nov 26 02:35:44 crc kubenswrapper[4766]: I1126 02:35:44.394311 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q8z4k" Nov 26 02:35:44 crc kubenswrapper[4766]: I1126 02:35:44.947315 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q8z4k"] Nov 26 02:35:45 crc kubenswrapper[4766]: I1126 02:35:45.049955 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q8z4k" event={"ID":"30b17774-f46f-4d8e-8211-56148141054a","Type":"ContainerStarted","Data":"15e535994b1fc6643d49d0b7caa7f467e254b93d068321677b8fe4a44958b456"} Nov 26 02:35:46 crc kubenswrapper[4766]: I1126 02:35:46.062629 4766 generic.go:334] "Generic (PLEG): container finished" podID="30b17774-f46f-4d8e-8211-56148141054a" containerID="067426c10d5407eb8a6c3f90f35c338363d00155cdba76863563aef83ee1e69a" exitCode=0 Nov 26 02:35:46 crc kubenswrapper[4766]: I1126 02:35:46.062687 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q8z4k" event={"ID":"30b17774-f46f-4d8e-8211-56148141054a","Type":"ContainerDied","Data":"067426c10d5407eb8a6c3f90f35c338363d00155cdba76863563aef83ee1e69a"} Nov 26 02:35:46 crc kubenswrapper[4766]: I1126 02:35:46.827143 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:35:46 crc kubenswrapper[4766]: E1126 02:35:46.827794 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:35:48 crc kubenswrapper[4766]: I1126 02:35:48.113378 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q8z4k" event={"ID":"30b17774-f46f-4d8e-8211-56148141054a","Type":"ContainerStarted","Data":"c1dd85c5014bb1724a8c4cff8b05f239a9a9a871e9ca9ec0a9d58c8b759f8ef1"} Nov 26 02:35:49 crc kubenswrapper[4766]: I1126 02:35:49.132284 4766 generic.go:334] "Generic (PLEG): container finished" podID="30b17774-f46f-4d8e-8211-56148141054a" containerID="c1dd85c5014bb1724a8c4cff8b05f239a9a9a871e9ca9ec0a9d58c8b759f8ef1" exitCode=0 Nov 26 02:35:49 crc kubenswrapper[4766]: I1126 02:35:49.132625 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q8z4k" event={"ID":"30b17774-f46f-4d8e-8211-56148141054a","Type":"ContainerDied","Data":"c1dd85c5014bb1724a8c4cff8b05f239a9a9a871e9ca9ec0a9d58c8b759f8ef1"} Nov 26 02:35:50 crc kubenswrapper[4766]: I1126 02:35:50.145733 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q8z4k" event={"ID":"30b17774-f46f-4d8e-8211-56148141054a","Type":"ContainerStarted","Data":"fa288cfbc07dd9cb3603c28b0fc77913a9d5aa136ccc8fe23527caf4f86d070d"} Nov 26 02:35:50 crc kubenswrapper[4766]: I1126 02:35:50.172368 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-q8z4k" podStartSLOduration=2.6431884610000003 podStartE2EDuration="6.172344295s" podCreationTimestamp="2025-11-26 02:35:44 +0000 UTC" firstStartedPulling="2025-11-26 02:35:46.070166551 +0000 UTC m=+7926.918936981" lastFinishedPulling="2025-11-26 02:35:49.599322395 +0000 UTC m=+7930.448092815" observedRunningTime="2025-11-26 02:35:50.161309295 +0000 UTC m=+7931.010079735" watchObservedRunningTime="2025-11-26 02:35:50.172344295 +0000 UTC m=+7931.021114735" Nov 26 02:35:54 crc kubenswrapper[4766]: I1126 02:35:54.395313 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-q8z4k" Nov 26 02:35:54 crc kubenswrapper[4766]: I1126 02:35:54.396637 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-q8z4k" Nov 26 02:35:54 crc kubenswrapper[4766]: I1126 02:35:54.475868 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-q8z4k" Nov 26 02:35:55 crc kubenswrapper[4766]: I1126 02:35:55.291452 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-q8z4k" Nov 26 02:35:55 crc kubenswrapper[4766]: I1126 02:35:55.364291 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q8z4k"] Nov 26 02:35:57 crc kubenswrapper[4766]: I1126 02:35:57.239348 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-q8z4k" podUID="30b17774-f46f-4d8e-8211-56148141054a" containerName="registry-server" containerID="cri-o://fa288cfbc07dd9cb3603c28b0fc77913a9d5aa136ccc8fe23527caf4f86d070d" gracePeriod=2 Nov 26 02:35:57 crc kubenswrapper[4766]: I1126 02:35:57.829571 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q8z4k" Nov 26 02:35:57 crc kubenswrapper[4766]: I1126 02:35:57.946215 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30b17774-f46f-4d8e-8211-56148141054a-catalog-content\") pod \"30b17774-f46f-4d8e-8211-56148141054a\" (UID: \"30b17774-f46f-4d8e-8211-56148141054a\") " Nov 26 02:35:57 crc kubenswrapper[4766]: I1126 02:35:57.946327 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqj57\" (UniqueName: \"kubernetes.io/projected/30b17774-f46f-4d8e-8211-56148141054a-kube-api-access-hqj57\") pod \"30b17774-f46f-4d8e-8211-56148141054a\" (UID: \"30b17774-f46f-4d8e-8211-56148141054a\") " Nov 26 02:35:57 crc kubenswrapper[4766]: I1126 02:35:57.946437 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30b17774-f46f-4d8e-8211-56148141054a-utilities\") pod \"30b17774-f46f-4d8e-8211-56148141054a\" (UID: \"30b17774-f46f-4d8e-8211-56148141054a\") " Nov 26 02:35:57 crc kubenswrapper[4766]: I1126 02:35:57.947256 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30b17774-f46f-4d8e-8211-56148141054a-utilities" (OuterVolumeSpecName: "utilities") pod "30b17774-f46f-4d8e-8211-56148141054a" (UID: "30b17774-f46f-4d8e-8211-56148141054a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:35:57 crc kubenswrapper[4766]: I1126 02:35:57.948075 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30b17774-f46f-4d8e-8211-56148141054a-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 02:35:57 crc kubenswrapper[4766]: I1126 02:35:57.953058 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30b17774-f46f-4d8e-8211-56148141054a-kube-api-access-hqj57" (OuterVolumeSpecName: "kube-api-access-hqj57") pod "30b17774-f46f-4d8e-8211-56148141054a" (UID: "30b17774-f46f-4d8e-8211-56148141054a"). InnerVolumeSpecName "kube-api-access-hqj57". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:35:58 crc kubenswrapper[4766]: I1126 02:35:58.030512 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30b17774-f46f-4d8e-8211-56148141054a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30b17774-f46f-4d8e-8211-56148141054a" (UID: "30b17774-f46f-4d8e-8211-56148141054a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:35:58 crc kubenswrapper[4766]: I1126 02:35:58.053358 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30b17774-f46f-4d8e-8211-56148141054a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 02:35:58 crc kubenswrapper[4766]: I1126 02:35:58.053401 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqj57\" (UniqueName: \"kubernetes.io/projected/30b17774-f46f-4d8e-8211-56148141054a-kube-api-access-hqj57\") on node \"crc\" DevicePath \"\"" Nov 26 02:35:58 crc kubenswrapper[4766]: I1126 02:35:58.257862 4766 generic.go:334] "Generic (PLEG): container finished" podID="30b17774-f46f-4d8e-8211-56148141054a" containerID="fa288cfbc07dd9cb3603c28b0fc77913a9d5aa136ccc8fe23527caf4f86d070d" exitCode=0 Nov 26 02:35:58 crc kubenswrapper[4766]: I1126 02:35:58.257914 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q8z4k" event={"ID":"30b17774-f46f-4d8e-8211-56148141054a","Type":"ContainerDied","Data":"fa288cfbc07dd9cb3603c28b0fc77913a9d5aa136ccc8fe23527caf4f86d070d"} Nov 26 02:35:58 crc kubenswrapper[4766]: I1126 02:35:58.257969 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q8z4k" event={"ID":"30b17774-f46f-4d8e-8211-56148141054a","Type":"ContainerDied","Data":"15e535994b1fc6643d49d0b7caa7f467e254b93d068321677b8fe4a44958b456"} Nov 26 02:35:58 crc kubenswrapper[4766]: I1126 02:35:58.257989 4766 scope.go:117] "RemoveContainer" containerID="fa288cfbc07dd9cb3603c28b0fc77913a9d5aa136ccc8fe23527caf4f86d070d" Nov 26 02:35:58 crc kubenswrapper[4766]: I1126 02:35:58.258006 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q8z4k" Nov 26 02:35:58 crc kubenswrapper[4766]: I1126 02:35:58.283408 4766 scope.go:117] "RemoveContainer" containerID="c1dd85c5014bb1724a8c4cff8b05f239a9a9a871e9ca9ec0a9d58c8b759f8ef1" Nov 26 02:35:58 crc kubenswrapper[4766]: I1126 02:35:58.311444 4766 scope.go:117] "RemoveContainer" containerID="067426c10d5407eb8a6c3f90f35c338363d00155cdba76863563aef83ee1e69a" Nov 26 02:35:58 crc kubenswrapper[4766]: I1126 02:35:58.321066 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q8z4k"] Nov 26 02:35:58 crc kubenswrapper[4766]: I1126 02:35:58.333922 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-q8z4k"] Nov 26 02:35:58 crc kubenswrapper[4766]: I1126 02:35:58.372171 4766 scope.go:117] "RemoveContainer" containerID="fa288cfbc07dd9cb3603c28b0fc77913a9d5aa136ccc8fe23527caf4f86d070d" Nov 26 02:35:58 crc kubenswrapper[4766]: E1126 02:35:58.372763 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa288cfbc07dd9cb3603c28b0fc77913a9d5aa136ccc8fe23527caf4f86d070d\": container with ID starting with fa288cfbc07dd9cb3603c28b0fc77913a9d5aa136ccc8fe23527caf4f86d070d not found: ID does not exist" containerID="fa288cfbc07dd9cb3603c28b0fc77913a9d5aa136ccc8fe23527caf4f86d070d" Nov 26 02:35:58 crc kubenswrapper[4766]: I1126 02:35:58.372798 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa288cfbc07dd9cb3603c28b0fc77913a9d5aa136ccc8fe23527caf4f86d070d"} err="failed to get container status \"fa288cfbc07dd9cb3603c28b0fc77913a9d5aa136ccc8fe23527caf4f86d070d\": rpc error: code = NotFound desc = could not find container \"fa288cfbc07dd9cb3603c28b0fc77913a9d5aa136ccc8fe23527caf4f86d070d\": container with ID starting with fa288cfbc07dd9cb3603c28b0fc77913a9d5aa136ccc8fe23527caf4f86d070d not found: ID does not exist" Nov 26 02:35:58 crc kubenswrapper[4766]: I1126 02:35:58.372816 4766 scope.go:117] "RemoveContainer" containerID="c1dd85c5014bb1724a8c4cff8b05f239a9a9a871e9ca9ec0a9d58c8b759f8ef1" Nov 26 02:35:58 crc kubenswrapper[4766]: E1126 02:35:58.373133 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1dd85c5014bb1724a8c4cff8b05f239a9a9a871e9ca9ec0a9d58c8b759f8ef1\": container with ID starting with c1dd85c5014bb1724a8c4cff8b05f239a9a9a871e9ca9ec0a9d58c8b759f8ef1 not found: ID does not exist" containerID="c1dd85c5014bb1724a8c4cff8b05f239a9a9a871e9ca9ec0a9d58c8b759f8ef1" Nov 26 02:35:58 crc kubenswrapper[4766]: I1126 02:35:58.373176 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1dd85c5014bb1724a8c4cff8b05f239a9a9a871e9ca9ec0a9d58c8b759f8ef1"} err="failed to get container status \"c1dd85c5014bb1724a8c4cff8b05f239a9a9a871e9ca9ec0a9d58c8b759f8ef1\": rpc error: code = NotFound desc = could not find container \"c1dd85c5014bb1724a8c4cff8b05f239a9a9a871e9ca9ec0a9d58c8b759f8ef1\": container with ID starting with c1dd85c5014bb1724a8c4cff8b05f239a9a9a871e9ca9ec0a9d58c8b759f8ef1 not found: ID does not exist" Nov 26 02:35:58 crc kubenswrapper[4766]: I1126 02:35:58.373211 4766 scope.go:117] "RemoveContainer" containerID="067426c10d5407eb8a6c3f90f35c338363d00155cdba76863563aef83ee1e69a" Nov 26 02:35:58 crc kubenswrapper[4766]: E1126 02:35:58.373503 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"067426c10d5407eb8a6c3f90f35c338363d00155cdba76863563aef83ee1e69a\": container with ID starting with 067426c10d5407eb8a6c3f90f35c338363d00155cdba76863563aef83ee1e69a not found: ID does not exist" containerID="067426c10d5407eb8a6c3f90f35c338363d00155cdba76863563aef83ee1e69a" Nov 26 02:35:58 crc kubenswrapper[4766]: I1126 02:35:58.373532 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"067426c10d5407eb8a6c3f90f35c338363d00155cdba76863563aef83ee1e69a"} err="failed to get container status \"067426c10d5407eb8a6c3f90f35c338363d00155cdba76863563aef83ee1e69a\": rpc error: code = NotFound desc = could not find container \"067426c10d5407eb8a6c3f90f35c338363d00155cdba76863563aef83ee1e69a\": container with ID starting with 067426c10d5407eb8a6c3f90f35c338363d00155cdba76863563aef83ee1e69a not found: ID does not exist" Nov 26 02:35:59 crc kubenswrapper[4766]: I1126 02:35:59.854913 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30b17774-f46f-4d8e-8211-56148141054a" path="/var/lib/kubelet/pods/30b17774-f46f-4d8e-8211-56148141054a/volumes" Nov 26 02:36:00 crc kubenswrapper[4766]: I1126 02:36:00.827861 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:36:00 crc kubenswrapper[4766]: E1126 02:36:00.828760 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:36:12 crc kubenswrapper[4766]: I1126 02:36:12.826696 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:36:12 crc kubenswrapper[4766]: E1126 02:36:12.827506 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:36:27 crc kubenswrapper[4766]: I1126 02:36:27.830556 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:36:27 crc kubenswrapper[4766]: E1126 02:36:27.831604 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:36:38 crc kubenswrapper[4766]: I1126 02:36:38.826699 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:36:38 crc kubenswrapper[4766]: E1126 02:36:38.827800 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:36:53 crc kubenswrapper[4766]: I1126 02:36:53.827365 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:36:53 crc kubenswrapper[4766]: E1126 02:36:53.831591 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:37:05 crc kubenswrapper[4766]: I1126 02:37:05.828395 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:37:05 crc kubenswrapper[4766]: E1126 02:37:05.829367 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:37:20 crc kubenswrapper[4766]: I1126 02:37:20.827370 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:37:20 crc kubenswrapper[4766]: E1126 02:37:20.828152 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:37:32 crc kubenswrapper[4766]: I1126 02:37:32.827791 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:37:32 crc kubenswrapper[4766]: E1126 02:37:32.828805 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:37:33 crc kubenswrapper[4766]: I1126 02:37:33.264105 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ffdxw/must-gather-x859m"] Nov 26 02:37:33 crc kubenswrapper[4766]: E1126 02:37:33.264991 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30b17774-f46f-4d8e-8211-56148141054a" containerName="extract-content" Nov 26 02:37:33 crc kubenswrapper[4766]: I1126 02:37:33.265016 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="30b17774-f46f-4d8e-8211-56148141054a" containerName="extract-content" Nov 26 02:37:33 crc kubenswrapper[4766]: E1126 02:37:33.265049 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30b17774-f46f-4d8e-8211-56148141054a" containerName="extract-utilities" Nov 26 02:37:33 crc kubenswrapper[4766]: I1126 02:37:33.265060 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="30b17774-f46f-4d8e-8211-56148141054a" containerName="extract-utilities" Nov 26 02:37:33 crc kubenswrapper[4766]: E1126 02:37:33.265086 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30b17774-f46f-4d8e-8211-56148141054a" containerName="registry-server" Nov 26 02:37:33 crc kubenswrapper[4766]: I1126 02:37:33.265094 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="30b17774-f46f-4d8e-8211-56148141054a" containerName="registry-server" Nov 26 02:37:33 crc kubenswrapper[4766]: I1126 02:37:33.265375 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="30b17774-f46f-4d8e-8211-56148141054a" containerName="registry-server" Nov 26 02:37:33 crc kubenswrapper[4766]: I1126 02:37:33.268840 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffdxw/must-gather-x859m" Nov 26 02:37:33 crc kubenswrapper[4766]: I1126 02:37:33.270879 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ffdxw"/"openshift-service-ca.crt" Nov 26 02:37:33 crc kubenswrapper[4766]: I1126 02:37:33.272963 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ffdxw"/"kube-root-ca.crt" Nov 26 02:37:33 crc kubenswrapper[4766]: I1126 02:37:33.301795 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ffdxw/must-gather-x859m"] Nov 26 02:37:33 crc kubenswrapper[4766]: I1126 02:37:33.339136 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57ckq\" (UniqueName: \"kubernetes.io/projected/e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb-kube-api-access-57ckq\") pod \"must-gather-x859m\" (UID: \"e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb\") " pod="openshift-must-gather-ffdxw/must-gather-x859m" Nov 26 02:37:33 crc kubenswrapper[4766]: I1126 02:37:33.339505 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb-must-gather-output\") pod \"must-gather-x859m\" (UID: \"e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb\") " pod="openshift-must-gather-ffdxw/must-gather-x859m" Nov 26 02:37:33 crc kubenswrapper[4766]: I1126 02:37:33.441911 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb-must-gather-output\") pod \"must-gather-x859m\" (UID: \"e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb\") " pod="openshift-must-gather-ffdxw/must-gather-x859m" Nov 26 02:37:33 crc kubenswrapper[4766]: I1126 02:37:33.442051 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57ckq\" (UniqueName: \"kubernetes.io/projected/e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb-kube-api-access-57ckq\") pod \"must-gather-x859m\" (UID: \"e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb\") " pod="openshift-must-gather-ffdxw/must-gather-x859m" Nov 26 02:37:33 crc kubenswrapper[4766]: I1126 02:37:33.442321 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb-must-gather-output\") pod \"must-gather-x859m\" (UID: \"e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb\") " pod="openshift-must-gather-ffdxw/must-gather-x859m" Nov 26 02:37:33 crc kubenswrapper[4766]: I1126 02:37:33.463195 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57ckq\" (UniqueName: \"kubernetes.io/projected/e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb-kube-api-access-57ckq\") pod \"must-gather-x859m\" (UID: \"e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb\") " pod="openshift-must-gather-ffdxw/must-gather-x859m" Nov 26 02:37:33 crc kubenswrapper[4766]: I1126 02:37:33.588520 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffdxw/must-gather-x859m" Nov 26 02:37:34 crc kubenswrapper[4766]: I1126 02:37:34.183910 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ffdxw/must-gather-x859m"] Nov 26 02:37:34 crc kubenswrapper[4766]: I1126 02:37:34.822067 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffdxw/must-gather-x859m" event={"ID":"e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb","Type":"ContainerStarted","Data":"24e3e8da6468bd4fd80a5ae54b258044de46c878b291d6c281dd72a64b80b501"} Nov 26 02:37:34 crc kubenswrapper[4766]: I1126 02:37:34.822559 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffdxw/must-gather-x859m" event={"ID":"e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb","Type":"ContainerStarted","Data":"97e84cf8325e3d6640d7c8169043bd2cd2642d5b6cb59280ba24cf83bb89b4d5"} Nov 26 02:37:35 crc kubenswrapper[4766]: I1126 02:37:35.839323 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffdxw/must-gather-x859m" event={"ID":"e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb","Type":"ContainerStarted","Data":"dfe6aea51adb38a75d913e3b421827a39f2f9cd780d8595389a17b3f322ad9eb"} Nov 26 02:37:35 crc kubenswrapper[4766]: I1126 02:37:35.867112 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ffdxw/must-gather-x859m" podStartSLOduration=2.867093966 podStartE2EDuration="2.867093966s" podCreationTimestamp="2025-11-26 02:37:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 02:37:35.860364511 +0000 UTC m=+8036.709134951" watchObservedRunningTime="2025-11-26 02:37:35.867093966 +0000 UTC m=+8036.715864396" Nov 26 02:37:39 crc kubenswrapper[4766]: I1126 02:37:39.220405 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ffdxw/crc-debug-6x7q2"] Nov 26 02:37:39 crc kubenswrapper[4766]: I1126 02:37:39.222833 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffdxw/crc-debug-6x7q2" Nov 26 02:37:39 crc kubenswrapper[4766]: I1126 02:37:39.225294 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ffdxw"/"default-dockercfg-9m76s" Nov 26 02:37:39 crc kubenswrapper[4766]: I1126 02:37:39.228458 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjq8f\" (UniqueName: \"kubernetes.io/projected/a1c98313-0336-468c-b44f-1af2c7d6c7e6-kube-api-access-pjq8f\") pod \"crc-debug-6x7q2\" (UID: \"a1c98313-0336-468c-b44f-1af2c7d6c7e6\") " pod="openshift-must-gather-ffdxw/crc-debug-6x7q2" Nov 26 02:37:39 crc kubenswrapper[4766]: I1126 02:37:39.228963 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a1c98313-0336-468c-b44f-1af2c7d6c7e6-host\") pod \"crc-debug-6x7q2\" (UID: \"a1c98313-0336-468c-b44f-1af2c7d6c7e6\") " pod="openshift-must-gather-ffdxw/crc-debug-6x7q2" Nov 26 02:37:39 crc kubenswrapper[4766]: I1126 02:37:39.331894 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a1c98313-0336-468c-b44f-1af2c7d6c7e6-host\") pod \"crc-debug-6x7q2\" (UID: \"a1c98313-0336-468c-b44f-1af2c7d6c7e6\") " pod="openshift-must-gather-ffdxw/crc-debug-6x7q2" Nov 26 02:37:39 crc kubenswrapper[4766]: I1126 02:37:39.332102 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjq8f\" (UniqueName: \"kubernetes.io/projected/a1c98313-0336-468c-b44f-1af2c7d6c7e6-kube-api-access-pjq8f\") pod \"crc-debug-6x7q2\" (UID: \"a1c98313-0336-468c-b44f-1af2c7d6c7e6\") " pod="openshift-must-gather-ffdxw/crc-debug-6x7q2" Nov 26 02:37:39 crc kubenswrapper[4766]: I1126 02:37:39.332257 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a1c98313-0336-468c-b44f-1af2c7d6c7e6-host\") pod \"crc-debug-6x7q2\" (UID: \"a1c98313-0336-468c-b44f-1af2c7d6c7e6\") " pod="openshift-must-gather-ffdxw/crc-debug-6x7q2" Nov 26 02:37:39 crc kubenswrapper[4766]: I1126 02:37:39.361859 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjq8f\" (UniqueName: \"kubernetes.io/projected/a1c98313-0336-468c-b44f-1af2c7d6c7e6-kube-api-access-pjq8f\") pod \"crc-debug-6x7q2\" (UID: \"a1c98313-0336-468c-b44f-1af2c7d6c7e6\") " pod="openshift-must-gather-ffdxw/crc-debug-6x7q2" Nov 26 02:37:39 crc kubenswrapper[4766]: I1126 02:37:39.545330 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffdxw/crc-debug-6x7q2" Nov 26 02:37:39 crc kubenswrapper[4766]: W1126 02:37:39.584252 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1c98313_0336_468c_b44f_1af2c7d6c7e6.slice/crio-3751be814ae2b01de92530af1f06129b70f819c1c7b53553e3447f5ad7946387 WatchSource:0}: Error finding container 3751be814ae2b01de92530af1f06129b70f819c1c7b53553e3447f5ad7946387: Status 404 returned error can't find the container with id 3751be814ae2b01de92530af1f06129b70f819c1c7b53553e3447f5ad7946387 Nov 26 02:37:39 crc kubenswrapper[4766]: I1126 02:37:39.893882 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffdxw/crc-debug-6x7q2" event={"ID":"a1c98313-0336-468c-b44f-1af2c7d6c7e6","Type":"ContainerStarted","Data":"c8b18123802c7d0220efeb297a4b53f7040e8e4defba128dcc05dbf67c7aac68"} Nov 26 02:37:39 crc kubenswrapper[4766]: I1126 02:37:39.894462 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffdxw/crc-debug-6x7q2" event={"ID":"a1c98313-0336-468c-b44f-1af2c7d6c7e6","Type":"ContainerStarted","Data":"3751be814ae2b01de92530af1f06129b70f819c1c7b53553e3447f5ad7946387"} Nov 26 02:37:39 crc kubenswrapper[4766]: I1126 02:37:39.907250 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ffdxw/crc-debug-6x7q2" podStartSLOduration=0.90723541 podStartE2EDuration="907.23541ms" podCreationTimestamp="2025-11-26 02:37:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 02:37:39.906264407 +0000 UTC m=+8040.755034837" watchObservedRunningTime="2025-11-26 02:37:39.90723541 +0000 UTC m=+8040.756005830" Nov 26 02:37:47 crc kubenswrapper[4766]: I1126 02:37:47.830426 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:37:47 crc kubenswrapper[4766]: E1126 02:37:47.831166 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:37:59 crc kubenswrapper[4766]: I1126 02:37:59.834160 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:37:59 crc kubenswrapper[4766]: E1126 02:37:59.835177 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:38:11 crc kubenswrapper[4766]: I1126 02:38:11.828906 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:38:12 crc kubenswrapper[4766]: I1126 02:38:12.214707 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"93d6a4a6ec497ff7bc4cee7766539719d24858cdc0bc941ecde4de3cf6ff02cd"} Nov 26 02:38:22 crc kubenswrapper[4766]: I1126 02:38:22.337148 4766 generic.go:334] "Generic (PLEG): container finished" podID="a1c98313-0336-468c-b44f-1af2c7d6c7e6" containerID="c8b18123802c7d0220efeb297a4b53f7040e8e4defba128dcc05dbf67c7aac68" exitCode=0 Nov 26 02:38:22 crc kubenswrapper[4766]: I1126 02:38:22.337245 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffdxw/crc-debug-6x7q2" event={"ID":"a1c98313-0336-468c-b44f-1af2c7d6c7e6","Type":"ContainerDied","Data":"c8b18123802c7d0220efeb297a4b53f7040e8e4defba128dcc05dbf67c7aac68"} Nov 26 02:38:23 crc kubenswrapper[4766]: I1126 02:38:23.468751 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffdxw/crc-debug-6x7q2" Nov 26 02:38:23 crc kubenswrapper[4766]: I1126 02:38:23.512548 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ffdxw/crc-debug-6x7q2"] Nov 26 02:38:23 crc kubenswrapper[4766]: I1126 02:38:23.522487 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ffdxw/crc-debug-6x7q2"] Nov 26 02:38:23 crc kubenswrapper[4766]: I1126 02:38:23.561697 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a1c98313-0336-468c-b44f-1af2c7d6c7e6-host\") pod \"a1c98313-0336-468c-b44f-1af2c7d6c7e6\" (UID: \"a1c98313-0336-468c-b44f-1af2c7d6c7e6\") " Nov 26 02:38:23 crc kubenswrapper[4766]: I1126 02:38:23.561802 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1c98313-0336-468c-b44f-1af2c7d6c7e6-host" (OuterVolumeSpecName: "host") pod "a1c98313-0336-468c-b44f-1af2c7d6c7e6" (UID: "a1c98313-0336-468c-b44f-1af2c7d6c7e6"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 02:38:23 crc kubenswrapper[4766]: I1126 02:38:23.561901 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjq8f\" (UniqueName: \"kubernetes.io/projected/a1c98313-0336-468c-b44f-1af2c7d6c7e6-kube-api-access-pjq8f\") pod \"a1c98313-0336-468c-b44f-1af2c7d6c7e6\" (UID: \"a1c98313-0336-468c-b44f-1af2c7d6c7e6\") " Nov 26 02:38:23 crc kubenswrapper[4766]: I1126 02:38:23.562414 4766 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a1c98313-0336-468c-b44f-1af2c7d6c7e6-host\") on node \"crc\" DevicePath \"\"" Nov 26 02:38:23 crc kubenswrapper[4766]: I1126 02:38:23.569018 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1c98313-0336-468c-b44f-1af2c7d6c7e6-kube-api-access-pjq8f" (OuterVolumeSpecName: "kube-api-access-pjq8f") pod "a1c98313-0336-468c-b44f-1af2c7d6c7e6" (UID: "a1c98313-0336-468c-b44f-1af2c7d6c7e6"). InnerVolumeSpecName "kube-api-access-pjq8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:38:23 crc kubenswrapper[4766]: I1126 02:38:23.677312 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjq8f\" (UniqueName: \"kubernetes.io/projected/a1c98313-0336-468c-b44f-1af2c7d6c7e6-kube-api-access-pjq8f\") on node \"crc\" DevicePath \"\"" Nov 26 02:38:23 crc kubenswrapper[4766]: I1126 02:38:23.843365 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1c98313-0336-468c-b44f-1af2c7d6c7e6" path="/var/lib/kubelet/pods/a1c98313-0336-468c-b44f-1af2c7d6c7e6/volumes" Nov 26 02:38:24 crc kubenswrapper[4766]: I1126 02:38:24.359813 4766 scope.go:117] "RemoveContainer" containerID="c8b18123802c7d0220efeb297a4b53f7040e8e4defba128dcc05dbf67c7aac68" Nov 26 02:38:24 crc kubenswrapper[4766]: I1126 02:38:24.359875 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffdxw/crc-debug-6x7q2" Nov 26 02:38:24 crc kubenswrapper[4766]: I1126 02:38:24.714746 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ffdxw/crc-debug-96kpn"] Nov 26 02:38:24 crc kubenswrapper[4766]: E1126 02:38:24.715214 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c98313-0336-468c-b44f-1af2c7d6c7e6" containerName="container-00" Nov 26 02:38:24 crc kubenswrapper[4766]: I1126 02:38:24.715227 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c98313-0336-468c-b44f-1af2c7d6c7e6" containerName="container-00" Nov 26 02:38:24 crc kubenswrapper[4766]: I1126 02:38:24.715475 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1c98313-0336-468c-b44f-1af2c7d6c7e6" containerName="container-00" Nov 26 02:38:24 crc kubenswrapper[4766]: I1126 02:38:24.716227 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffdxw/crc-debug-96kpn" Nov 26 02:38:24 crc kubenswrapper[4766]: I1126 02:38:24.718170 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ffdxw"/"default-dockercfg-9m76s" Nov 26 02:38:24 crc kubenswrapper[4766]: I1126 02:38:24.801597 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d25d05fd-c8b1-4a8d-affe-305d2027b432-host\") pod \"crc-debug-96kpn\" (UID: \"d25d05fd-c8b1-4a8d-affe-305d2027b432\") " pod="openshift-must-gather-ffdxw/crc-debug-96kpn" Nov 26 02:38:24 crc kubenswrapper[4766]: I1126 02:38:24.801744 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbqkk\" (UniqueName: \"kubernetes.io/projected/d25d05fd-c8b1-4a8d-affe-305d2027b432-kube-api-access-sbqkk\") pod \"crc-debug-96kpn\" (UID: \"d25d05fd-c8b1-4a8d-affe-305d2027b432\") " pod="openshift-must-gather-ffdxw/crc-debug-96kpn" Nov 26 02:38:24 crc kubenswrapper[4766]: I1126 02:38:24.904251 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d25d05fd-c8b1-4a8d-affe-305d2027b432-host\") pod \"crc-debug-96kpn\" (UID: \"d25d05fd-c8b1-4a8d-affe-305d2027b432\") " pod="openshift-must-gather-ffdxw/crc-debug-96kpn" Nov 26 02:38:24 crc kubenswrapper[4766]: I1126 02:38:24.904328 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbqkk\" (UniqueName: \"kubernetes.io/projected/d25d05fd-c8b1-4a8d-affe-305d2027b432-kube-api-access-sbqkk\") pod \"crc-debug-96kpn\" (UID: \"d25d05fd-c8b1-4a8d-affe-305d2027b432\") " pod="openshift-must-gather-ffdxw/crc-debug-96kpn" Nov 26 02:38:24 crc kubenswrapper[4766]: I1126 02:38:24.906816 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d25d05fd-c8b1-4a8d-affe-305d2027b432-host\") pod \"crc-debug-96kpn\" (UID: \"d25d05fd-c8b1-4a8d-affe-305d2027b432\") " pod="openshift-must-gather-ffdxw/crc-debug-96kpn" Nov 26 02:38:24 crc kubenswrapper[4766]: I1126 02:38:24.926345 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbqkk\" (UniqueName: \"kubernetes.io/projected/d25d05fd-c8b1-4a8d-affe-305d2027b432-kube-api-access-sbqkk\") pod \"crc-debug-96kpn\" (UID: \"d25d05fd-c8b1-4a8d-affe-305d2027b432\") " pod="openshift-must-gather-ffdxw/crc-debug-96kpn" Nov 26 02:38:25 crc kubenswrapper[4766]: I1126 02:38:25.043132 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffdxw/crc-debug-96kpn" Nov 26 02:38:25 crc kubenswrapper[4766]: W1126 02:38:25.089275 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd25d05fd_c8b1_4a8d_affe_305d2027b432.slice/crio-391927e0b54b06b29f0c33d34b9ff81bda3d570702796cc9b36105de57c7d4d0 WatchSource:0}: Error finding container 391927e0b54b06b29f0c33d34b9ff81bda3d570702796cc9b36105de57c7d4d0: Status 404 returned error can't find the container with id 391927e0b54b06b29f0c33d34b9ff81bda3d570702796cc9b36105de57c7d4d0 Nov 26 02:38:25 crc kubenswrapper[4766]: I1126 02:38:25.373281 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffdxw/crc-debug-96kpn" event={"ID":"d25d05fd-c8b1-4a8d-affe-305d2027b432","Type":"ContainerStarted","Data":"391927e0b54b06b29f0c33d34b9ff81bda3d570702796cc9b36105de57c7d4d0"} Nov 26 02:38:26 crc kubenswrapper[4766]: I1126 02:38:26.384180 4766 generic.go:334] "Generic (PLEG): container finished" podID="d25d05fd-c8b1-4a8d-affe-305d2027b432" containerID="c3fdaa8a53fca84445f7eb62a215407d07a46d3ead6bb9e962216858cfa6e30f" exitCode=0 Nov 26 02:38:26 crc kubenswrapper[4766]: I1126 02:38:26.384237 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffdxw/crc-debug-96kpn" event={"ID":"d25d05fd-c8b1-4a8d-affe-305d2027b432","Type":"ContainerDied","Data":"c3fdaa8a53fca84445f7eb62a215407d07a46d3ead6bb9e962216858cfa6e30f"} Nov 26 02:38:27 crc kubenswrapper[4766]: I1126 02:38:27.551622 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffdxw/crc-debug-96kpn" Nov 26 02:38:27 crc kubenswrapper[4766]: I1126 02:38:27.658303 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbqkk\" (UniqueName: \"kubernetes.io/projected/d25d05fd-c8b1-4a8d-affe-305d2027b432-kube-api-access-sbqkk\") pod \"d25d05fd-c8b1-4a8d-affe-305d2027b432\" (UID: \"d25d05fd-c8b1-4a8d-affe-305d2027b432\") " Nov 26 02:38:27 crc kubenswrapper[4766]: I1126 02:38:27.658591 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d25d05fd-c8b1-4a8d-affe-305d2027b432-host\") pod \"d25d05fd-c8b1-4a8d-affe-305d2027b432\" (UID: \"d25d05fd-c8b1-4a8d-affe-305d2027b432\") " Nov 26 02:38:27 crc kubenswrapper[4766]: I1126 02:38:27.659237 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d25d05fd-c8b1-4a8d-affe-305d2027b432-host" (OuterVolumeSpecName: "host") pod "d25d05fd-c8b1-4a8d-affe-305d2027b432" (UID: "d25d05fd-c8b1-4a8d-affe-305d2027b432"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 02:38:27 crc kubenswrapper[4766]: I1126 02:38:27.689572 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d25d05fd-c8b1-4a8d-affe-305d2027b432-kube-api-access-sbqkk" (OuterVolumeSpecName: "kube-api-access-sbqkk") pod "d25d05fd-c8b1-4a8d-affe-305d2027b432" (UID: "d25d05fd-c8b1-4a8d-affe-305d2027b432"). InnerVolumeSpecName "kube-api-access-sbqkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:38:27 crc kubenswrapper[4766]: I1126 02:38:27.762719 4766 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d25d05fd-c8b1-4a8d-affe-305d2027b432-host\") on node \"crc\" DevicePath \"\"" Nov 26 02:38:27 crc kubenswrapper[4766]: I1126 02:38:27.762760 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbqkk\" (UniqueName: \"kubernetes.io/projected/d25d05fd-c8b1-4a8d-affe-305d2027b432-kube-api-access-sbqkk\") on node \"crc\" DevicePath \"\"" Nov 26 02:38:28 crc kubenswrapper[4766]: I1126 02:38:28.408547 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffdxw/crc-debug-96kpn" event={"ID":"d25d05fd-c8b1-4a8d-affe-305d2027b432","Type":"ContainerDied","Data":"391927e0b54b06b29f0c33d34b9ff81bda3d570702796cc9b36105de57c7d4d0"} Nov 26 02:38:28 crc kubenswrapper[4766]: I1126 02:38:28.408983 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="391927e0b54b06b29f0c33d34b9ff81bda3d570702796cc9b36105de57c7d4d0" Nov 26 02:38:28 crc kubenswrapper[4766]: I1126 02:38:28.408590 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffdxw/crc-debug-96kpn" Nov 26 02:38:28 crc kubenswrapper[4766]: I1126 02:38:28.686634 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ffdxw/crc-debug-96kpn"] Nov 26 02:38:28 crc kubenswrapper[4766]: I1126 02:38:28.697205 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ffdxw/crc-debug-96kpn"] Nov 26 02:38:29 crc kubenswrapper[4766]: I1126 02:38:29.849703 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d25d05fd-c8b1-4a8d-affe-305d2027b432" path="/var/lib/kubelet/pods/d25d05fd-c8b1-4a8d-affe-305d2027b432/volumes" Nov 26 02:38:29 crc kubenswrapper[4766]: I1126 02:38:29.956312 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ffdxw/crc-debug-7lvrt"] Nov 26 02:38:29 crc kubenswrapper[4766]: E1126 02:38:29.956851 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d25d05fd-c8b1-4a8d-affe-305d2027b432" containerName="container-00" Nov 26 02:38:29 crc kubenswrapper[4766]: I1126 02:38:29.956875 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="d25d05fd-c8b1-4a8d-affe-305d2027b432" containerName="container-00" Nov 26 02:38:29 crc kubenswrapper[4766]: I1126 02:38:29.957218 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="d25d05fd-c8b1-4a8d-affe-305d2027b432" containerName="container-00" Nov 26 02:38:29 crc kubenswrapper[4766]: I1126 02:38:29.958135 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffdxw/crc-debug-7lvrt" Nov 26 02:38:29 crc kubenswrapper[4766]: I1126 02:38:29.960502 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ffdxw"/"default-dockercfg-9m76s" Nov 26 02:38:30 crc kubenswrapper[4766]: I1126 02:38:30.118196 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6ktq\" (UniqueName: \"kubernetes.io/projected/a02b096a-e7bb-4e75-959c-764b0e9d5844-kube-api-access-w6ktq\") pod \"crc-debug-7lvrt\" (UID: \"a02b096a-e7bb-4e75-959c-764b0e9d5844\") " pod="openshift-must-gather-ffdxw/crc-debug-7lvrt" Nov 26 02:38:30 crc kubenswrapper[4766]: I1126 02:38:30.118261 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a02b096a-e7bb-4e75-959c-764b0e9d5844-host\") pod \"crc-debug-7lvrt\" (UID: \"a02b096a-e7bb-4e75-959c-764b0e9d5844\") " pod="openshift-must-gather-ffdxw/crc-debug-7lvrt" Nov 26 02:38:30 crc kubenswrapper[4766]: I1126 02:38:30.221002 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6ktq\" (UniqueName: \"kubernetes.io/projected/a02b096a-e7bb-4e75-959c-764b0e9d5844-kube-api-access-w6ktq\") pod \"crc-debug-7lvrt\" (UID: \"a02b096a-e7bb-4e75-959c-764b0e9d5844\") " pod="openshift-must-gather-ffdxw/crc-debug-7lvrt" Nov 26 02:38:30 crc kubenswrapper[4766]: I1126 02:38:30.221050 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a02b096a-e7bb-4e75-959c-764b0e9d5844-host\") pod \"crc-debug-7lvrt\" (UID: \"a02b096a-e7bb-4e75-959c-764b0e9d5844\") " pod="openshift-must-gather-ffdxw/crc-debug-7lvrt" Nov 26 02:38:30 crc kubenswrapper[4766]: I1126 02:38:30.221267 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a02b096a-e7bb-4e75-959c-764b0e9d5844-host\") pod \"crc-debug-7lvrt\" (UID: \"a02b096a-e7bb-4e75-959c-764b0e9d5844\") " pod="openshift-must-gather-ffdxw/crc-debug-7lvrt" Nov 26 02:38:30 crc kubenswrapper[4766]: I1126 02:38:30.253324 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6ktq\" (UniqueName: \"kubernetes.io/projected/a02b096a-e7bb-4e75-959c-764b0e9d5844-kube-api-access-w6ktq\") pod \"crc-debug-7lvrt\" (UID: \"a02b096a-e7bb-4e75-959c-764b0e9d5844\") " pod="openshift-must-gather-ffdxw/crc-debug-7lvrt" Nov 26 02:38:30 crc kubenswrapper[4766]: I1126 02:38:30.281380 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffdxw/crc-debug-7lvrt" Nov 26 02:38:30 crc kubenswrapper[4766]: W1126 02:38:30.322317 4766 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda02b096a_e7bb_4e75_959c_764b0e9d5844.slice/crio-56f924d65310755115c3846f2d53f6217a2c2d0e4e2136ebf35aa4e456d4d329 WatchSource:0}: Error finding container 56f924d65310755115c3846f2d53f6217a2c2d0e4e2136ebf35aa4e456d4d329: Status 404 returned error can't find the container with id 56f924d65310755115c3846f2d53f6217a2c2d0e4e2136ebf35aa4e456d4d329 Nov 26 02:38:30 crc kubenswrapper[4766]: I1126 02:38:30.432275 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffdxw/crc-debug-7lvrt" event={"ID":"a02b096a-e7bb-4e75-959c-764b0e9d5844","Type":"ContainerStarted","Data":"56f924d65310755115c3846f2d53f6217a2c2d0e4e2136ebf35aa4e456d4d329"} Nov 26 02:38:31 crc kubenswrapper[4766]: I1126 02:38:31.446030 4766 generic.go:334] "Generic (PLEG): container finished" podID="a02b096a-e7bb-4e75-959c-764b0e9d5844" containerID="8755df9d6d32400c296c1ffc1cedf23aab2dc6cfb0f4df5feb768026989abbd0" exitCode=0 Nov 26 02:38:31 crc kubenswrapper[4766]: I1126 02:38:31.446138 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffdxw/crc-debug-7lvrt" event={"ID":"a02b096a-e7bb-4e75-959c-764b0e9d5844","Type":"ContainerDied","Data":"8755df9d6d32400c296c1ffc1cedf23aab2dc6cfb0f4df5feb768026989abbd0"} Nov 26 02:38:31 crc kubenswrapper[4766]: I1126 02:38:31.519114 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ffdxw/crc-debug-7lvrt"] Nov 26 02:38:31 crc kubenswrapper[4766]: I1126 02:38:31.536747 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ffdxw/crc-debug-7lvrt"] Nov 26 02:38:32 crc kubenswrapper[4766]: I1126 02:38:32.593988 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffdxw/crc-debug-7lvrt" Nov 26 02:38:32 crc kubenswrapper[4766]: I1126 02:38:32.640909 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a02b096a-e7bb-4e75-959c-764b0e9d5844-host\") pod \"a02b096a-e7bb-4e75-959c-764b0e9d5844\" (UID: \"a02b096a-e7bb-4e75-959c-764b0e9d5844\") " Nov 26 02:38:32 crc kubenswrapper[4766]: I1126 02:38:32.641012 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a02b096a-e7bb-4e75-959c-764b0e9d5844-host" (OuterVolumeSpecName: "host") pod "a02b096a-e7bb-4e75-959c-764b0e9d5844" (UID: "a02b096a-e7bb-4e75-959c-764b0e9d5844"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 02:38:32 crc kubenswrapper[4766]: I1126 02:38:32.641172 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6ktq\" (UniqueName: \"kubernetes.io/projected/a02b096a-e7bb-4e75-959c-764b0e9d5844-kube-api-access-w6ktq\") pod \"a02b096a-e7bb-4e75-959c-764b0e9d5844\" (UID: \"a02b096a-e7bb-4e75-959c-764b0e9d5844\") " Nov 26 02:38:32 crc kubenswrapper[4766]: I1126 02:38:32.642176 4766 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a02b096a-e7bb-4e75-959c-764b0e9d5844-host\") on node \"crc\" DevicePath \"\"" Nov 26 02:38:32 crc kubenswrapper[4766]: I1126 02:38:32.647769 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a02b096a-e7bb-4e75-959c-764b0e9d5844-kube-api-access-w6ktq" (OuterVolumeSpecName: "kube-api-access-w6ktq") pod "a02b096a-e7bb-4e75-959c-764b0e9d5844" (UID: "a02b096a-e7bb-4e75-959c-764b0e9d5844"). InnerVolumeSpecName "kube-api-access-w6ktq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:38:32 crc kubenswrapper[4766]: I1126 02:38:32.744093 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6ktq\" (UniqueName: \"kubernetes.io/projected/a02b096a-e7bb-4e75-959c-764b0e9d5844-kube-api-access-w6ktq\") on node \"crc\" DevicePath \"\"" Nov 26 02:38:33 crc kubenswrapper[4766]: I1126 02:38:33.469667 4766 scope.go:117] "RemoveContainer" containerID="8755df9d6d32400c296c1ffc1cedf23aab2dc6cfb0f4df5feb768026989abbd0" Nov 26 02:38:33 crc kubenswrapper[4766]: I1126 02:38:33.469729 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffdxw/crc-debug-7lvrt" Nov 26 02:38:33 crc kubenswrapper[4766]: I1126 02:38:33.838017 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a02b096a-e7bb-4e75-959c-764b0e9d5844" path="/var/lib/kubelet/pods/a02b096a-e7bb-4e75-959c-764b0e9d5844/volumes" Nov 26 02:39:15 crc kubenswrapper[4766]: I1126 02:39:15.975391 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k7qcj"] Nov 26 02:39:15 crc kubenswrapper[4766]: E1126 02:39:15.976792 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a02b096a-e7bb-4e75-959c-764b0e9d5844" containerName="container-00" Nov 26 02:39:15 crc kubenswrapper[4766]: I1126 02:39:15.976813 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="a02b096a-e7bb-4e75-959c-764b0e9d5844" containerName="container-00" Nov 26 02:39:15 crc kubenswrapper[4766]: I1126 02:39:15.977187 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="a02b096a-e7bb-4e75-959c-764b0e9d5844" containerName="container-00" Nov 26 02:39:15 crc kubenswrapper[4766]: I1126 02:39:15.982021 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k7qcj" Nov 26 02:39:16 crc kubenswrapper[4766]: I1126 02:39:16.000584 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7qcj"] Nov 26 02:39:16 crc kubenswrapper[4766]: I1126 02:39:16.087223 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a-utilities\") pod \"redhat-marketplace-k7qcj\" (UID: \"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a\") " pod="openshift-marketplace/redhat-marketplace-k7qcj" Nov 26 02:39:16 crc kubenswrapper[4766]: I1126 02:39:16.087620 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs5sx\" (UniqueName: \"kubernetes.io/projected/6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a-kube-api-access-rs5sx\") pod \"redhat-marketplace-k7qcj\" (UID: \"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a\") " pod="openshift-marketplace/redhat-marketplace-k7qcj" Nov 26 02:39:16 crc kubenswrapper[4766]: I1126 02:39:16.088185 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a-catalog-content\") pod \"redhat-marketplace-k7qcj\" (UID: \"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a\") " pod="openshift-marketplace/redhat-marketplace-k7qcj" Nov 26 02:39:16 crc kubenswrapper[4766]: I1126 02:39:16.190963 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a-catalog-content\") pod \"redhat-marketplace-k7qcj\" (UID: \"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a\") " pod="openshift-marketplace/redhat-marketplace-k7qcj" Nov 26 02:39:16 crc kubenswrapper[4766]: I1126 02:39:16.191130 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a-utilities\") pod \"redhat-marketplace-k7qcj\" (UID: \"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a\") " pod="openshift-marketplace/redhat-marketplace-k7qcj" Nov 26 02:39:16 crc kubenswrapper[4766]: I1126 02:39:16.191197 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs5sx\" (UniqueName: \"kubernetes.io/projected/6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a-kube-api-access-rs5sx\") pod \"redhat-marketplace-k7qcj\" (UID: \"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a\") " pod="openshift-marketplace/redhat-marketplace-k7qcj" Nov 26 02:39:16 crc kubenswrapper[4766]: I1126 02:39:16.191499 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a-catalog-content\") pod \"redhat-marketplace-k7qcj\" (UID: \"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a\") " pod="openshift-marketplace/redhat-marketplace-k7qcj" Nov 26 02:39:16 crc kubenswrapper[4766]: I1126 02:39:16.191889 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a-utilities\") pod \"redhat-marketplace-k7qcj\" (UID: \"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a\") " pod="openshift-marketplace/redhat-marketplace-k7qcj" Nov 26 02:39:16 crc kubenswrapper[4766]: I1126 02:39:16.220880 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs5sx\" (UniqueName: \"kubernetes.io/projected/6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a-kube-api-access-rs5sx\") pod \"redhat-marketplace-k7qcj\" (UID: \"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a\") " pod="openshift-marketplace/redhat-marketplace-k7qcj" Nov 26 02:39:16 crc kubenswrapper[4766]: I1126 02:39:16.324794 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k7qcj" Nov 26 02:39:16 crc kubenswrapper[4766]: I1126 02:39:16.881007 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7qcj"] Nov 26 02:39:17 crc kubenswrapper[4766]: I1126 02:39:17.053800 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7qcj" event={"ID":"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a","Type":"ContainerStarted","Data":"426b1b062719da82686a7c7c18634487b68a69429625a54d058d797957fa4267"} Nov 26 02:39:18 crc kubenswrapper[4766]: I1126 02:39:18.064921 4766 generic.go:334] "Generic (PLEG): container finished" podID="6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a" containerID="e706490e40d0739c2e74c127fb7b60f004b24cc326c38a76a1a626a0b39206ae" exitCode=0 Nov 26 02:39:18 crc kubenswrapper[4766]: I1126 02:39:18.064983 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7qcj" event={"ID":"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a","Type":"ContainerDied","Data":"e706490e40d0739c2e74c127fb7b60f004b24cc326c38a76a1a626a0b39206ae"} Nov 26 02:39:18 crc kubenswrapper[4766]: I1126 02:39:18.067927 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 02:39:19 crc kubenswrapper[4766]: I1126 02:39:19.084431 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7qcj" event={"ID":"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a","Type":"ContainerStarted","Data":"26fd8c74148262fd3db7f1b53060d3f19b64676e63d3a0def1e18c69b41f5a8e"} Nov 26 02:39:20 crc kubenswrapper[4766]: I1126 02:39:20.095921 4766 generic.go:334] "Generic (PLEG): container finished" podID="6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a" containerID="26fd8c74148262fd3db7f1b53060d3f19b64676e63d3a0def1e18c69b41f5a8e" exitCode=0 Nov 26 02:39:20 crc kubenswrapper[4766]: I1126 02:39:20.096196 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7qcj" event={"ID":"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a","Type":"ContainerDied","Data":"26fd8c74148262fd3db7f1b53060d3f19b64676e63d3a0def1e18c69b41f5a8e"} Nov 26 02:39:21 crc kubenswrapper[4766]: I1126 02:39:21.108913 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7qcj" event={"ID":"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a","Type":"ContainerStarted","Data":"abf79ff2a792624cad3be8255052c5dcef53f2e9a477face31d8a9449fd6b1da"} Nov 26 02:39:21 crc kubenswrapper[4766]: I1126 02:39:21.132018 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k7qcj" podStartSLOduration=3.654371351 podStartE2EDuration="6.131999437s" podCreationTimestamp="2025-11-26 02:39:15 +0000 UTC" firstStartedPulling="2025-11-26 02:39:18.067676016 +0000 UTC m=+8138.916446446" lastFinishedPulling="2025-11-26 02:39:20.545304102 +0000 UTC m=+8141.394074532" observedRunningTime="2025-11-26 02:39:21.123717034 +0000 UTC m=+8141.972487464" watchObservedRunningTime="2025-11-26 02:39:21.131999437 +0000 UTC m=+8141.980769867" Nov 26 02:39:26 crc kubenswrapper[4766]: I1126 02:39:26.326004 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k7qcj" Nov 26 02:39:26 crc kubenswrapper[4766]: I1126 02:39:26.326419 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k7qcj" Nov 26 02:39:26 crc kubenswrapper[4766]: I1126 02:39:26.391931 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k7qcj" Nov 26 02:39:27 crc kubenswrapper[4766]: I1126 02:39:27.212032 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k7qcj" Nov 26 02:39:27 crc kubenswrapper[4766]: I1126 02:39:27.262320 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7qcj"] Nov 26 02:39:29 crc kubenswrapper[4766]: I1126 02:39:29.200571 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-k7qcj" podUID="6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a" containerName="registry-server" containerID="cri-o://abf79ff2a792624cad3be8255052c5dcef53f2e9a477face31d8a9449fd6b1da" gracePeriod=2 Nov 26 02:39:29 crc kubenswrapper[4766]: I1126 02:39:29.751888 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k7qcj" Nov 26 02:39:29 crc kubenswrapper[4766]: I1126 02:39:29.889831 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a-catalog-content\") pod \"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a\" (UID: \"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a\") " Nov 26 02:39:29 crc kubenswrapper[4766]: I1126 02:39:29.889878 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs5sx\" (UniqueName: \"kubernetes.io/projected/6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a-kube-api-access-rs5sx\") pod \"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a\" (UID: \"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a\") " Nov 26 02:39:29 crc kubenswrapper[4766]: I1126 02:39:29.889932 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a-utilities\") pod \"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a\" (UID: \"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a\") " Nov 26 02:39:29 crc kubenswrapper[4766]: I1126 02:39:29.891140 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a-utilities" (OuterVolumeSpecName: "utilities") pod "6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a" (UID: "6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:39:29 crc kubenswrapper[4766]: I1126 02:39:29.903569 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a-kube-api-access-rs5sx" (OuterVolumeSpecName: "kube-api-access-rs5sx") pod "6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a" (UID: "6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a"). InnerVolumeSpecName "kube-api-access-rs5sx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:39:29 crc kubenswrapper[4766]: I1126 02:39:29.934810 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a" (UID: "6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:39:29 crc kubenswrapper[4766]: I1126 02:39:29.992788 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 02:39:29 crc kubenswrapper[4766]: I1126 02:39:29.992820 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs5sx\" (UniqueName: \"kubernetes.io/projected/6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a-kube-api-access-rs5sx\") on node \"crc\" DevicePath \"\"" Nov 26 02:39:29 crc kubenswrapper[4766]: I1126 02:39:29.992830 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 02:39:30 crc kubenswrapper[4766]: I1126 02:39:30.215547 4766 generic.go:334] "Generic (PLEG): container finished" podID="6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a" containerID="abf79ff2a792624cad3be8255052c5dcef53f2e9a477face31d8a9449fd6b1da" exitCode=0 Nov 26 02:39:30 crc kubenswrapper[4766]: I1126 02:39:30.215716 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7qcj" event={"ID":"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a","Type":"ContainerDied","Data":"abf79ff2a792624cad3be8255052c5dcef53f2e9a477face31d8a9449fd6b1da"} Nov 26 02:39:30 crc kubenswrapper[4766]: I1126 02:39:30.215904 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7qcj" event={"ID":"6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a","Type":"ContainerDied","Data":"426b1b062719da82686a7c7c18634487b68a69429625a54d058d797957fa4267"} Nov 26 02:39:30 crc kubenswrapper[4766]: I1126 02:39:30.215929 4766 scope.go:117] "RemoveContainer" containerID="abf79ff2a792624cad3be8255052c5dcef53f2e9a477face31d8a9449fd6b1da" Nov 26 02:39:30 crc kubenswrapper[4766]: I1126 02:39:30.215786 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k7qcj" Nov 26 02:39:30 crc kubenswrapper[4766]: I1126 02:39:30.239738 4766 scope.go:117] "RemoveContainer" containerID="26fd8c74148262fd3db7f1b53060d3f19b64676e63d3a0def1e18c69b41f5a8e" Nov 26 02:39:30 crc kubenswrapper[4766]: I1126 02:39:30.266149 4766 scope.go:117] "RemoveContainer" containerID="e706490e40d0739c2e74c127fb7b60f004b24cc326c38a76a1a626a0b39206ae" Nov 26 02:39:30 crc kubenswrapper[4766]: I1126 02:39:30.279050 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7qcj"] Nov 26 02:39:30 crc kubenswrapper[4766]: I1126 02:39:30.294640 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7qcj"] Nov 26 02:39:30 crc kubenswrapper[4766]: I1126 02:39:30.341642 4766 scope.go:117] "RemoveContainer" containerID="abf79ff2a792624cad3be8255052c5dcef53f2e9a477face31d8a9449fd6b1da" Nov 26 02:39:30 crc kubenswrapper[4766]: E1126 02:39:30.342069 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abf79ff2a792624cad3be8255052c5dcef53f2e9a477face31d8a9449fd6b1da\": container with ID starting with abf79ff2a792624cad3be8255052c5dcef53f2e9a477face31d8a9449fd6b1da not found: ID does not exist" containerID="abf79ff2a792624cad3be8255052c5dcef53f2e9a477face31d8a9449fd6b1da" Nov 26 02:39:30 crc kubenswrapper[4766]: I1126 02:39:30.342103 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abf79ff2a792624cad3be8255052c5dcef53f2e9a477face31d8a9449fd6b1da"} err="failed to get container status \"abf79ff2a792624cad3be8255052c5dcef53f2e9a477face31d8a9449fd6b1da\": rpc error: code = NotFound desc = could not find container \"abf79ff2a792624cad3be8255052c5dcef53f2e9a477face31d8a9449fd6b1da\": container with ID starting with abf79ff2a792624cad3be8255052c5dcef53f2e9a477face31d8a9449fd6b1da not found: ID does not exist" Nov 26 02:39:30 crc kubenswrapper[4766]: I1126 02:39:30.342124 4766 scope.go:117] "RemoveContainer" containerID="26fd8c74148262fd3db7f1b53060d3f19b64676e63d3a0def1e18c69b41f5a8e" Nov 26 02:39:30 crc kubenswrapper[4766]: E1126 02:39:30.342461 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26fd8c74148262fd3db7f1b53060d3f19b64676e63d3a0def1e18c69b41f5a8e\": container with ID starting with 26fd8c74148262fd3db7f1b53060d3f19b64676e63d3a0def1e18c69b41f5a8e not found: ID does not exist" containerID="26fd8c74148262fd3db7f1b53060d3f19b64676e63d3a0def1e18c69b41f5a8e" Nov 26 02:39:30 crc kubenswrapper[4766]: I1126 02:39:30.342479 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26fd8c74148262fd3db7f1b53060d3f19b64676e63d3a0def1e18c69b41f5a8e"} err="failed to get container status \"26fd8c74148262fd3db7f1b53060d3f19b64676e63d3a0def1e18c69b41f5a8e\": rpc error: code = NotFound desc = could not find container \"26fd8c74148262fd3db7f1b53060d3f19b64676e63d3a0def1e18c69b41f5a8e\": container with ID starting with 26fd8c74148262fd3db7f1b53060d3f19b64676e63d3a0def1e18c69b41f5a8e not found: ID does not exist" Nov 26 02:39:30 crc kubenswrapper[4766]: I1126 02:39:30.342492 4766 scope.go:117] "RemoveContainer" containerID="e706490e40d0739c2e74c127fb7b60f004b24cc326c38a76a1a626a0b39206ae" Nov 26 02:39:30 crc kubenswrapper[4766]: E1126 02:39:30.342720 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e706490e40d0739c2e74c127fb7b60f004b24cc326c38a76a1a626a0b39206ae\": container with ID starting with e706490e40d0739c2e74c127fb7b60f004b24cc326c38a76a1a626a0b39206ae not found: ID does not exist" containerID="e706490e40d0739c2e74c127fb7b60f004b24cc326c38a76a1a626a0b39206ae" Nov 26 02:39:30 crc kubenswrapper[4766]: I1126 02:39:30.342738 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e706490e40d0739c2e74c127fb7b60f004b24cc326c38a76a1a626a0b39206ae"} err="failed to get container status \"e706490e40d0739c2e74c127fb7b60f004b24cc326c38a76a1a626a0b39206ae\": rpc error: code = NotFound desc = could not find container \"e706490e40d0739c2e74c127fb7b60f004b24cc326c38a76a1a626a0b39206ae\": container with ID starting with e706490e40d0739c2e74c127fb7b60f004b24cc326c38a76a1a626a0b39206ae not found: ID does not exist" Nov 26 02:39:31 crc kubenswrapper[4766]: I1126 02:39:31.849436 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a" path="/var/lib/kubelet/pods/6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a/volumes" Nov 26 02:39:42 crc kubenswrapper[4766]: I1126 02:39:42.963635 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_54262828-5166-42d6-934e-5eab8683ecbd/aodh-api/0.log" Nov 26 02:39:43 crc kubenswrapper[4766]: I1126 02:39:43.221402 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_54262828-5166-42d6-934e-5eab8683ecbd/aodh-listener/0.log" Nov 26 02:39:43 crc kubenswrapper[4766]: I1126 02:39:43.225840 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_54262828-5166-42d6-934e-5eab8683ecbd/aodh-evaluator/0.log" Nov 26 02:39:43 crc kubenswrapper[4766]: I1126 02:39:43.247193 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_54262828-5166-42d6-934e-5eab8683ecbd/aodh-notifier/0.log" Nov 26 02:39:43 crc kubenswrapper[4766]: I1126 02:39:43.438802 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6b7ccbcb5b-rf9v9_5655d266-3ea5-404f-901e-80da609b0319/barbican-api/0.log" Nov 26 02:39:43 crc kubenswrapper[4766]: I1126 02:39:43.446947 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6b7ccbcb5b-rf9v9_5655d266-3ea5-404f-901e-80da609b0319/barbican-api-log/0.log" Nov 26 02:39:43 crc kubenswrapper[4766]: I1126 02:39:43.682134 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7fd5c59d54-6ccpg_81836728-9333-4234-8175-954682305e4e/barbican-keystone-listener/0.log" Nov 26 02:39:43 crc kubenswrapper[4766]: I1126 02:39:43.844382 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7fd5c59d54-6ccpg_81836728-9333-4234-8175-954682305e4e/barbican-keystone-listener-log/0.log" Nov 26 02:39:43 crc kubenswrapper[4766]: I1126 02:39:43.894146 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-68b9476655-7svbz_01cfcf71-49d4-498f-a9cc-5e9c65c08709/barbican-worker/0.log" Nov 26 02:39:43 crc kubenswrapper[4766]: I1126 02:39:43.959275 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-68b9476655-7svbz_01cfcf71-49d4-498f-a9cc-5e9c65c08709/barbican-worker-log/0.log" Nov 26 02:39:44 crc kubenswrapper[4766]: I1126 02:39:44.175517 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-xs5dl_973efc5a-bdcf-43af-baf5-11ae375d3494/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:39:44 crc kubenswrapper[4766]: I1126 02:39:44.258097 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2c82c6d8-c395-4177-930f-f840d902df28/ceilometer-central-agent/0.log" Nov 26 02:39:44 crc kubenswrapper[4766]: I1126 02:39:44.404113 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2c82c6d8-c395-4177-930f-f840d902df28/ceilometer-notification-agent/0.log" Nov 26 02:39:44 crc kubenswrapper[4766]: I1126 02:39:44.406260 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2c82c6d8-c395-4177-930f-f840d902df28/proxy-httpd/0.log" Nov 26 02:39:44 crc kubenswrapper[4766]: I1126 02:39:44.453095 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2c82c6d8-c395-4177-930f-f840d902df28/sg-core/0.log" Nov 26 02:39:44 crc kubenswrapper[4766]: I1126 02:39:44.623749 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-6vbwl_91756acc-c3e0-471b-b79d-4eb1cf2a80df/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:39:44 crc kubenswrapper[4766]: I1126 02:39:44.645335 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-dv4d5_9f63fa32-bcb4-490f-947b-ec7ab639505c/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:39:44 crc kubenswrapper[4766]: I1126 02:39:44.861751 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_d4e1503f-0e08-4e3a-a578-fcbd3b51989a/cinder-api-log/0.log" Nov 26 02:39:44 crc kubenswrapper[4766]: I1126 02:39:44.974843 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_d4e1503f-0e08-4e3a-a578-fcbd3b51989a/cinder-api/0.log" Nov 26 02:39:45 crc kubenswrapper[4766]: I1126 02:39:45.156781 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_2c89caf0-3bca-4172-9a9b-27657d46a1c5/probe/0.log" Nov 26 02:39:45 crc kubenswrapper[4766]: I1126 02:39:45.163428 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_2c89caf0-3bca-4172-9a9b-27657d46a1c5/cinder-backup/0.log" Nov 26 02:39:45 crc kubenswrapper[4766]: I1126 02:39:45.257374 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_92d51b16-e82c-4fec-a095-53e3f0608fad/cinder-scheduler/0.log" Nov 26 02:39:45 crc kubenswrapper[4766]: I1126 02:39:45.444270 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_92d51b16-e82c-4fec-a095-53e3f0608fad/probe/0.log" Nov 26 02:39:45 crc kubenswrapper[4766]: I1126 02:39:45.517981 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_5687a3ac-6850-4533-9c40-b0954d98c628/cinder-volume/0.log" Nov 26 02:39:45 crc kubenswrapper[4766]: I1126 02:39:45.555473 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_5687a3ac-6850-4533-9c40-b0954d98c628/probe/0.log" Nov 26 02:39:45 crc kubenswrapper[4766]: I1126 02:39:45.724291 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-4bmds_a7ae4f62-184e-4093-9a0a-12d6188ee301/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:39:45 crc kubenswrapper[4766]: I1126 02:39:45.803144 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-9t59g_511ab0e6-b14c-4249-92e9-184b57a03147/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:39:45 crc kubenswrapper[4766]: I1126 02:39:45.963496 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-c8d8d886c-9lgvr_316b23d8-2140-4ac5-8e26-d63171814293/init/0.log" Nov 26 02:39:46 crc kubenswrapper[4766]: I1126 02:39:46.235237 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-c8d8d886c-9lgvr_316b23d8-2140-4ac5-8e26-d63171814293/init/0.log" Nov 26 02:39:46 crc kubenswrapper[4766]: I1126 02:39:46.248880 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-c8d8d886c-9lgvr_316b23d8-2140-4ac5-8e26-d63171814293/dnsmasq-dns/0.log" Nov 26 02:39:46 crc kubenswrapper[4766]: I1126 02:39:46.303822 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_5ea31542-fdc7-481a-853a-5935e39fca92/glance-httpd/0.log" Nov 26 02:39:46 crc kubenswrapper[4766]: I1126 02:39:46.419831 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_5ea31542-fdc7-481a-853a-5935e39fca92/glance-log/0.log" Nov 26 02:39:46 crc kubenswrapper[4766]: I1126 02:39:46.486239 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f5f6c81c-04d9-4542-9ab1-b932737f388c/glance-httpd/0.log" Nov 26 02:39:46 crc kubenswrapper[4766]: I1126 02:39:46.501263 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f5f6c81c-04d9-4542-9ab1-b932737f388c/glance-log/0.log" Nov 26 02:39:47 crc kubenswrapper[4766]: I1126 02:39:47.245609 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-7f6d8f97f4-szt84_65159db0-64b4-4cce-ba5e-aa6f889f7f58/heat-api/0.log" Nov 26 02:39:47 crc kubenswrapper[4766]: I1126 02:39:47.452007 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-6ff8c446-6wzh8_775687b1-1657-4115-b482-f7e418e77727/heat-engine/0.log" Nov 26 02:39:47 crc kubenswrapper[4766]: I1126 02:39:47.699676 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-cf46f664-8k6s8_778430ca-2e98-4156-b1d9-fdf05b76ef7e/horizon/0.log" Nov 26 02:39:47 crc kubenswrapper[4766]: I1126 02:39:47.710557 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-l279n_7509ec1a-b269-4bbe-bfca-0dcb052360ea/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:39:47 crc kubenswrapper[4766]: I1126 02:39:47.941392 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-5ff994f754-l2ctn_5f65a888-def8-495b-915a-0b6556aff6de/heat-cfnapi/0.log" Nov 26 02:39:47 crc kubenswrapper[4766]: I1126 02:39:47.973047 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-hvdtg_5d516c45-cb2e-4920-8e14-3a47063212a1/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:39:48 crc kubenswrapper[4766]: I1126 02:39:48.077126 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-cf46f664-8k6s8_778430ca-2e98-4156-b1d9-fdf05b76ef7e/horizon-log/0.log" Nov 26 02:39:48 crc kubenswrapper[4766]: I1126 02:39:48.143643 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29401981-22f5z_5e25beeb-e89d-4aed-9353-3755a0ef2efb/keystone-cron/0.log" Nov 26 02:39:48 crc kubenswrapper[4766]: I1126 02:39:48.338010 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29402041-bjg4c_bd82c837-925c-4835-a90f-3d5a873dfbae/keystone-cron/0.log" Nov 26 02:39:48 crc kubenswrapper[4766]: I1126 02:39:48.489882 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a700ac64-3ee9-4e7b-8da8-1cb7537fee78/kube-state-metrics/0.log" Nov 26 02:39:48 crc kubenswrapper[4766]: I1126 02:39:48.543640 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6768b6c6f7-lm9pw_3fdd0908-10b2-437d-92bf-ccdc345a0d80/keystone-api/0.log" Nov 26 02:39:48 crc kubenswrapper[4766]: I1126 02:39:48.647398 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-nswmp_64255763-3ccc-42d2-88ae-3c48ecea8c07/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:39:48 crc kubenswrapper[4766]: I1126 02:39:48.732277 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_logging-edpm-deployment-openstack-edpm-ipam-25nr7_85810b6d-fde1-48a2-b55d-b6c9c72d6b6a/logging-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:39:48 crc kubenswrapper[4766]: I1126 02:39:48.864767 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_8cb6838c-a086-4990-82f0-a3c36675be67/manila-api-log/0.log" Nov 26 02:39:49 crc kubenswrapper[4766]: I1126 02:39:49.036139 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_8cb6838c-a086-4990-82f0-a3c36675be67/manila-api/0.log" Nov 26 02:39:49 crc kubenswrapper[4766]: I1126 02:39:49.132382 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_a487b55f-cdaa-4351-95ed-7fb4a8ce382d/probe/0.log" Nov 26 02:39:49 crc kubenswrapper[4766]: I1126 02:39:49.201725 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_a487b55f-cdaa-4351-95ed-7fb4a8ce382d/manila-scheduler/0.log" Nov 26 02:39:49 crc kubenswrapper[4766]: I1126 02:39:49.263535 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_91ee515c-0d11-436f-b0c9-184f065d52eb/probe/0.log" Nov 26 02:39:49 crc kubenswrapper[4766]: I1126 02:39:49.319037 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_91ee515c-0d11-436f-b0c9-184f065d52eb/manila-share/0.log" Nov 26 02:39:49 crc kubenswrapper[4766]: I1126 02:39:49.518059 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mysqld-exporter-0_51bb050a-0106-4f7c-acf2-8c6c19f93d39/mysqld-exporter/0.log" Nov 26 02:39:49 crc kubenswrapper[4766]: I1126 02:39:49.885290 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-756fcfc485-ktbfm_8d171003-64ac-4108-957b-70f4225359c5/neutron-api/0.log" Nov 26 02:39:49 crc kubenswrapper[4766]: I1126 02:39:49.906722 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-jm96h_3efd579d-7c56-4624-a920-1e40cea0b951/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:39:49 crc kubenswrapper[4766]: I1126 02:39:49.964581 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-756fcfc485-ktbfm_8d171003-64ac-4108-957b-70f4225359c5/neutron-httpd/0.log" Nov 26 02:39:50 crc kubenswrapper[4766]: I1126 02:39:50.791844 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a93e0b93-3cde-4098-a853-ea1b3c86c323/nova-api-log/0.log" Nov 26 02:39:50 crc kubenswrapper[4766]: I1126 02:39:50.881824 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_b33aa993-e31f-402f-9574-33bd30a942c4/nova-cell0-conductor-conductor/0.log" Nov 26 02:39:51 crc kubenswrapper[4766]: I1126 02:39:51.013030 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_f2ec2417-bd6f-4166-b42c-407424769274/nova-cell1-conductor-conductor/0.log" Nov 26 02:39:51 crc kubenswrapper[4766]: I1126 02:39:51.338744 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6t964_8da0c4a5-4c26-46de-8408-a8b562ba8f76/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:39:51 crc kubenswrapper[4766]: I1126 02:39:51.374222 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_4cf2f787-c351-49ce-a347-dc0975ea8bc4/nova-cell1-novncproxy-novncproxy/0.log" Nov 26 02:39:51 crc kubenswrapper[4766]: I1126 02:39:51.631776 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1f172da7-f357-44b5-a4c0-5a2cc493f906/nova-metadata-log/0.log" Nov 26 02:39:51 crc kubenswrapper[4766]: I1126 02:39:51.694055 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a93e0b93-3cde-4098-a853-ea1b3c86c323/nova-api-api/0.log" Nov 26 02:39:52 crc kubenswrapper[4766]: I1126 02:39:52.099527 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_44f36c8b-51b8-4bce-afbb-1962eb317dec/mysql-bootstrap/0.log" Nov 26 02:39:52 crc kubenswrapper[4766]: I1126 02:39:52.119914 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_64dcd71b-3e04-470c-ab99-37105bd225dc/nova-scheduler-scheduler/0.log" Nov 26 02:39:52 crc kubenswrapper[4766]: I1126 02:39:52.290260 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_44f36c8b-51b8-4bce-afbb-1962eb317dec/mysql-bootstrap/0.log" Nov 26 02:39:52 crc kubenswrapper[4766]: I1126 02:39:52.348226 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_44f36c8b-51b8-4bce-afbb-1962eb317dec/galera/0.log" Nov 26 02:39:52 crc kubenswrapper[4766]: I1126 02:39:52.514644 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_54c97191-0bfc-4a65-8168-10833aab2814/mysql-bootstrap/0.log" Nov 26 02:39:52 crc kubenswrapper[4766]: I1126 02:39:52.673211 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_54c97191-0bfc-4a65-8168-10833aab2814/mysql-bootstrap/0.log" Nov 26 02:39:52 crc kubenswrapper[4766]: I1126 02:39:52.730505 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_54c97191-0bfc-4a65-8168-10833aab2814/galera/0.log" Nov 26 02:39:52 crc kubenswrapper[4766]: I1126 02:39:52.917522 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_d8551b13-7602-4c7c-bfb5-0bcf7e64954d/openstackclient/0.log" Nov 26 02:39:52 crc kubenswrapper[4766]: I1126 02:39:52.963883 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-qnhgk_342b8104-66eb-4fe1-88ad-ef5afc617d65/openstack-network-exporter/0.log" Nov 26 02:39:53 crc kubenswrapper[4766]: I1126 02:39:53.194269 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6ghtc_71963f95-01c9-473a-be8a-7798f6038ffb/ovsdb-server-init/0.log" Nov 26 02:39:53 crc kubenswrapper[4766]: I1126 02:39:53.405727 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6ghtc_71963f95-01c9-473a-be8a-7798f6038ffb/ovs-vswitchd/0.log" Nov 26 02:39:53 crc kubenswrapper[4766]: I1126 02:39:53.435218 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6ghtc_71963f95-01c9-473a-be8a-7798f6038ffb/ovsdb-server-init/0.log" Nov 26 02:39:53 crc kubenswrapper[4766]: I1126 02:39:53.452477 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6ghtc_71963f95-01c9-473a-be8a-7798f6038ffb/ovsdb-server/0.log" Nov 26 02:39:53 crc kubenswrapper[4766]: I1126 02:39:53.684008 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-p42tr_22168318-cef8-4363-8b3e-3042e82037ea/ovn-controller/0.log" Nov 26 02:39:53 crc kubenswrapper[4766]: I1126 02:39:53.934038 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-8xf9z_45de00ed-b796-4ac1-8e38-ee601eab6dc0/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:39:54 crc kubenswrapper[4766]: I1126 02:39:54.151900 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_64b18a9d-088b-4a4a-9c1a-4b63cbb1d642/openstack-network-exporter/0.log" Nov 26 02:39:54 crc kubenswrapper[4766]: I1126 02:39:54.317423 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_64b18a9d-088b-4a4a-9c1a-4b63cbb1d642/ovn-northd/0.log" Nov 26 02:39:54 crc kubenswrapper[4766]: I1126 02:39:54.378235 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d346209c-ec54-4719-9b56-7a079c603ebc/openstack-network-exporter/0.log" Nov 26 02:39:54 crc kubenswrapper[4766]: I1126 02:39:54.527510 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d346209c-ec54-4719-9b56-7a079c603ebc/ovsdbserver-nb/0.log" Nov 26 02:39:54 crc kubenswrapper[4766]: I1126 02:39:54.595038 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_954395e4-178d-46ef-a695-8895ebe444d8/openstack-network-exporter/0.log" Nov 26 02:39:54 crc kubenswrapper[4766]: I1126 02:39:54.719243 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1f172da7-f357-44b5-a4c0-5a2cc493f906/nova-metadata-metadata/0.log" Nov 26 02:39:54 crc kubenswrapper[4766]: I1126 02:39:54.783586 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_954395e4-178d-46ef-a695-8895ebe444d8/ovsdbserver-sb/0.log" Nov 26 02:39:55 crc kubenswrapper[4766]: I1126 02:39:55.097698 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5fd6484f46-cqsxg_be4073df-dee9-4a70-8911-b23a317a0a57/placement-api/0.log" Nov 26 02:39:55 crc kubenswrapper[4766]: I1126 02:39:55.156630 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5fd6484f46-cqsxg_be4073df-dee9-4a70-8911-b23a317a0a57/placement-log/0.log" Nov 26 02:39:55 crc kubenswrapper[4766]: I1126 02:39:55.178261 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0c404609-2299-4fe5-8c43-da5614df251e/init-config-reloader/0.log" Nov 26 02:39:55 crc kubenswrapper[4766]: I1126 02:39:55.433025 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0c404609-2299-4fe5-8c43-da5614df251e/init-config-reloader/0.log" Nov 26 02:39:55 crc kubenswrapper[4766]: I1126 02:39:55.436974 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0c404609-2299-4fe5-8c43-da5614df251e/prometheus/0.log" Nov 26 02:39:55 crc kubenswrapper[4766]: I1126 02:39:55.447555 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0c404609-2299-4fe5-8c43-da5614df251e/thanos-sidecar/0.log" Nov 26 02:39:55 crc kubenswrapper[4766]: I1126 02:39:55.455402 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0c404609-2299-4fe5-8c43-da5614df251e/config-reloader/0.log" Nov 26 02:39:55 crc kubenswrapper[4766]: I1126 02:39:55.691692 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_64400ff7-9c3a-42c3-b1fc-2f9696f1206c/setup-container/0.log" Nov 26 02:39:55 crc kubenswrapper[4766]: I1126 02:39:55.895269 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_64400ff7-9c3a-42c3-b1fc-2f9696f1206c/setup-container/0.log" Nov 26 02:39:55 crc kubenswrapper[4766]: I1126 02:39:55.900908 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_64400ff7-9c3a-42c3-b1fc-2f9696f1206c/rabbitmq/0.log" Nov 26 02:39:56 crc kubenswrapper[4766]: I1126 02:39:56.008250 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_797e86ef-8e0d-499e-859f-6f6732826d7e/setup-container/0.log" Nov 26 02:39:56 crc kubenswrapper[4766]: I1126 02:39:56.181879 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_797e86ef-8e0d-499e-859f-6f6732826d7e/setup-container/0.log" Nov 26 02:39:56 crc kubenswrapper[4766]: I1126 02:39:56.209097 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-f5dqj_b59fb3c0-2626-47c2-b2f4-f13c13b011a5/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:39:56 crc kubenswrapper[4766]: I1126 02:39:56.235541 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_797e86ef-8e0d-499e-859f-6f6732826d7e/rabbitmq/0.log" Nov 26 02:39:56 crc kubenswrapper[4766]: I1126 02:39:56.418562 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-wnmdl_7ce6b646-f313-478b-b425-bc3b56504c3f/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:39:56 crc kubenswrapper[4766]: I1126 02:39:56.473314 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-tzd58_14ea5f7a-f044-426a-919e-a05cdcfe900c/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:39:56 crc kubenswrapper[4766]: I1126 02:39:56.697335 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-qlskl_57e4c16c-215b-4a8f-8c6c-a8d12e0b04fa/ssh-known-hosts-edpm-deployment/0.log" Nov 26 02:39:56 crc kubenswrapper[4766]: I1126 02:39:56.944665 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-74d678b987-q89lp_3d4e6997-4bbc-4cb2-a0f2-abf3b940090c/proxy-server/0.log" Nov 26 02:39:57 crc kubenswrapper[4766]: I1126 02:39:57.092441 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-tmfzn_6df26511-75fd-4bcb-ab3e-65cde68c9f46/swift-ring-rebalance/0.log" Nov 26 02:39:57 crc kubenswrapper[4766]: I1126 02:39:57.144743 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-74d678b987-q89lp_3d4e6997-4bbc-4cb2-a0f2-abf3b940090c/proxy-httpd/0.log" Nov 26 02:39:57 crc kubenswrapper[4766]: I1126 02:39:57.190199 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/account-auditor/0.log" Nov 26 02:39:57 crc kubenswrapper[4766]: I1126 02:39:57.335345 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/account-reaper/0.log" Nov 26 02:39:57 crc kubenswrapper[4766]: I1126 02:39:57.678133 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/account-server/0.log" Nov 26 02:39:57 crc kubenswrapper[4766]: I1126 02:39:57.709755 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/container-auditor/0.log" Nov 26 02:39:57 crc kubenswrapper[4766]: I1126 02:39:57.718489 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/account-replicator/0.log" Nov 26 02:39:57 crc kubenswrapper[4766]: I1126 02:39:57.816326 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/container-replicator/0.log" Nov 26 02:39:57 crc kubenswrapper[4766]: I1126 02:39:57.925012 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/container-updater/0.log" Nov 26 02:39:57 crc kubenswrapper[4766]: I1126 02:39:57.926159 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/container-server/0.log" Nov 26 02:39:58 crc kubenswrapper[4766]: I1126 02:39:58.039437 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/object-expirer/0.log" Nov 26 02:39:58 crc kubenswrapper[4766]: I1126 02:39:58.047740 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/object-auditor/0.log" Nov 26 02:39:58 crc kubenswrapper[4766]: I1126 02:39:58.163573 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/object-server/0.log" Nov 26 02:39:58 crc kubenswrapper[4766]: I1126 02:39:58.180029 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/object-replicator/0.log" Nov 26 02:39:58 crc kubenswrapper[4766]: I1126 02:39:58.283425 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/object-updater/0.log" Nov 26 02:39:58 crc kubenswrapper[4766]: I1126 02:39:58.286951 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/rsync/0.log" Nov 26 02:39:58 crc kubenswrapper[4766]: I1126 02:39:58.337814 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a1caa1fb-917a-4267-a947-1194557347eb/swift-recon-cron/0.log" Nov 26 02:39:58 crc kubenswrapper[4766]: I1126 02:39:58.575697 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-g6znx_eef488e4-2c28-4247-a8f4-b8e9008d518e/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:39:58 crc kubenswrapper[4766]: I1126 02:39:58.619183 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-power-monitoring-edpm-deployment-openstack-edpm-zv8s8_0667c3b6-bdae-4a36-acce-947435c51d91/telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:39:58 crc kubenswrapper[4766]: I1126 02:39:58.888324 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_f4a47960-5a63-4a92-a4e1-89d48aaa6515/test-operator-logs-container/0.log" Nov 26 02:39:59 crc kubenswrapper[4766]: I1126 02:39:59.094758 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-s9vgs_d2efe136-b7e0-475d-9204-f0e36e720e96/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 02:39:59 crc kubenswrapper[4766]: I1126 02:39:59.521157 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_edfc5f7f-1ff2-4147-b68e-b9be01ff6a99/tempest-tests-tempest-tests-runner/0.log" Nov 26 02:40:10 crc kubenswrapper[4766]: I1126 02:40:10.209981 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_f3679589-d23a-48c7-8b4a-712935d229f1/memcached/0.log" Nov 26 02:40:27 crc kubenswrapper[4766]: I1126 02:40:27.935152 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b_28a6a73a-8156-4df1-a41e-d2fe03ce2b25/util/0.log" Nov 26 02:40:28 crc kubenswrapper[4766]: I1126 02:40:28.087392 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b_28a6a73a-8156-4df1-a41e-d2fe03ce2b25/util/0.log" Nov 26 02:40:28 crc kubenswrapper[4766]: I1126 02:40:28.103761 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b_28a6a73a-8156-4df1-a41e-d2fe03ce2b25/pull/0.log" Nov 26 02:40:28 crc kubenswrapper[4766]: I1126 02:40:28.144218 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b_28a6a73a-8156-4df1-a41e-d2fe03ce2b25/pull/0.log" Nov 26 02:40:28 crc kubenswrapper[4766]: I1126 02:40:28.273018 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b_28a6a73a-8156-4df1-a41e-d2fe03ce2b25/util/0.log" Nov 26 02:40:28 crc kubenswrapper[4766]: I1126 02:40:28.290402 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b_28a6a73a-8156-4df1-a41e-d2fe03ce2b25/extract/0.log" Nov 26 02:40:28 crc kubenswrapper[4766]: I1126 02:40:28.344147 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_afdf394cba99fd80050b3a30ea222e2ab904ac1fc3df83873fd92ab55dftd6b_28a6a73a-8156-4df1-a41e-d2fe03ce2b25/pull/0.log" Nov 26 02:40:28 crc kubenswrapper[4766]: I1126 02:40:28.426207 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-gbdhz_cc87f631-ba4e-4943-b51d-bf1012cca754/kube-rbac-proxy/0.log" Nov 26 02:40:28 crc kubenswrapper[4766]: I1126 02:40:28.465999 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-gbdhz_cc87f631-ba4e-4943-b51d-bf1012cca754/manager/3.log" Nov 26 02:40:28 crc kubenswrapper[4766]: I1126 02:40:28.525276 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-gbdhz_cc87f631-ba4e-4943-b51d-bf1012cca754/manager/2.log" Nov 26 02:40:28 crc kubenswrapper[4766]: I1126 02:40:28.588189 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-jgz6c_31b1b73b-1d10-49ac-8549-b202c965b63e/kube-rbac-proxy/0.log" Nov 26 02:40:28 crc kubenswrapper[4766]: I1126 02:40:28.706168 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-jgz6c_31b1b73b-1d10-49ac-8549-b202c965b63e/manager/2.log" Nov 26 02:40:28 crc kubenswrapper[4766]: I1126 02:40:28.719925 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-jgz6c_31b1b73b-1d10-49ac-8549-b202c965b63e/manager/3.log" Nov 26 02:40:28 crc kubenswrapper[4766]: I1126 02:40:28.801512 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-dqtcj_0e6cfef3-86af-4cf0-ab06-b3eef687836b/kube-rbac-proxy/0.log" Nov 26 02:40:29 crc kubenswrapper[4766]: I1126 02:40:29.011645 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-dqtcj_0e6cfef3-86af-4cf0-ab06-b3eef687836b/manager/3.log" Nov 26 02:40:29 crc kubenswrapper[4766]: I1126 02:40:29.026111 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-dqtcj_0e6cfef3-86af-4cf0-ab06-b3eef687836b/manager/2.log" Nov 26 02:40:29 crc kubenswrapper[4766]: I1126 02:40:29.063108 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-k68q9_45353fc4-9955-460b-9276-9792dfdaf8f4/kube-rbac-proxy/0.log" Nov 26 02:40:29 crc kubenswrapper[4766]: I1126 02:40:29.184713 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-k68q9_45353fc4-9955-460b-9276-9792dfdaf8f4/manager/1.log" Nov 26 02:40:29 crc kubenswrapper[4766]: I1126 02:40:29.211325 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-k68q9_45353fc4-9955-460b-9276-9792dfdaf8f4/manager/2.log" Nov 26 02:40:29 crc kubenswrapper[4766]: I1126 02:40:29.256589 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-tkj9b_1ecbd1d2-cb7f-432f-bc75-903d0643ea29/kube-rbac-proxy/0.log" Nov 26 02:40:29 crc kubenswrapper[4766]: I1126 02:40:29.420365 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-tkj9b_1ecbd1d2-cb7f-432f-bc75-903d0643ea29/manager/2.log" Nov 26 02:40:29 crc kubenswrapper[4766]: I1126 02:40:29.446490 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-tkj9b_1ecbd1d2-cb7f-432f-bc75-903d0643ea29/manager/3.log" Nov 26 02:40:29 crc kubenswrapper[4766]: I1126 02:40:29.509376 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-h4mrn_5284ad70-a36c-4666-8202-84633d83f50b/kube-rbac-proxy/0.log" Nov 26 02:40:29 crc kubenswrapper[4766]: I1126 02:40:29.658115 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-h4mrn_5284ad70-a36c-4666-8202-84633d83f50b/manager/3.log" Nov 26 02:40:29 crc kubenswrapper[4766]: I1126 02:40:29.660747 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-h4mrn_5284ad70-a36c-4666-8202-84633d83f50b/manager/2.log" Nov 26 02:40:29 crc kubenswrapper[4766]: I1126 02:40:29.764107 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-x6h4v_c3cec209-fdef-44a6-90d0-2d06cfd7e5cc/kube-rbac-proxy/0.log" Nov 26 02:40:29 crc kubenswrapper[4766]: I1126 02:40:29.851038 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-x6h4v_c3cec209-fdef-44a6-90d0-2d06cfd7e5cc/manager/2.log" Nov 26 02:40:29 crc kubenswrapper[4766]: I1126 02:40:29.958002 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-qtnxq_08bd3753-3b70-44f8-92ff-98bedd88f662/kube-rbac-proxy/0.log" Nov 26 02:40:29 crc kubenswrapper[4766]: I1126 02:40:29.970780 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-x6h4v_c3cec209-fdef-44a6-90d0-2d06cfd7e5cc/manager/3.log" Nov 26 02:40:30 crc kubenswrapper[4766]: I1126 02:40:30.067291 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-qtnxq_08bd3753-3b70-44f8-92ff-98bedd88f662/manager/3.log" Nov 26 02:40:30 crc kubenswrapper[4766]: I1126 02:40:30.158682 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-fcr76_3d0d59b2-0486-40a2-ae60-b738a4f76dc4/kube-rbac-proxy/0.log" Nov 26 02:40:30 crc kubenswrapper[4766]: I1126 02:40:30.188425 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-qtnxq_08bd3753-3b70-44f8-92ff-98bedd88f662/manager/2.log" Nov 26 02:40:30 crc kubenswrapper[4766]: I1126 02:40:30.244038 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-fcr76_3d0d59b2-0486-40a2-ae60-b738a4f76dc4/manager/3.log" Nov 26 02:40:30 crc kubenswrapper[4766]: I1126 02:40:30.330892 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-fcr76_3d0d59b2-0486-40a2-ae60-b738a4f76dc4/manager/2.log" Nov 26 02:40:30 crc kubenswrapper[4766]: I1126 02:40:30.417503 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-rs4kn_a16c5563-750f-4e91-93bd-4044c656cb88/kube-rbac-proxy/0.log" Nov 26 02:40:30 crc kubenswrapper[4766]: I1126 02:40:30.530298 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-rs4kn_a16c5563-750f-4e91-93bd-4044c656cb88/manager/2.log" Nov 26 02:40:30 crc kubenswrapper[4766]: I1126 02:40:30.633872 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-rs4kn_a16c5563-750f-4e91-93bd-4044c656cb88/manager/1.log" Nov 26 02:40:30 crc kubenswrapper[4766]: I1126 02:40:30.653218 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-z865g_749a25e6-23eb-4fdd-8e15-ac2874cdee7f/kube-rbac-proxy/0.log" Nov 26 02:40:30 crc kubenswrapper[4766]: I1126 02:40:30.707494 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-z865g_749a25e6-23eb-4fdd-8e15-ac2874cdee7f/manager/3.log" Nov 26 02:40:30 crc kubenswrapper[4766]: I1126 02:40:30.819966 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-z865g_749a25e6-23eb-4fdd-8e15-ac2874cdee7f/manager/2.log" Nov 26 02:40:30 crc kubenswrapper[4766]: I1126 02:40:30.849406 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-wdzgw_2bb23a3b-1f4a-409f-9d9c-0afe73a33352/kube-rbac-proxy/0.log" Nov 26 02:40:30 crc kubenswrapper[4766]: I1126 02:40:30.913818 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-wdzgw_2bb23a3b-1f4a-409f-9d9c-0afe73a33352/manager/3.log" Nov 26 02:40:31 crc kubenswrapper[4766]: I1126 02:40:31.027109 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-wdzgw_2bb23a3b-1f4a-409f-9d9c-0afe73a33352/manager/2.log" Nov 26 02:40:31 crc kubenswrapper[4766]: I1126 02:40:31.120377 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-dm7xb_8969652f-14c9-4618-ad2d-1b3fd1ad3627/kube-rbac-proxy/0.log" Nov 26 02:40:31 crc kubenswrapper[4766]: I1126 02:40:31.155446 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-dm7xb_8969652f-14c9-4618-ad2d-1b3fd1ad3627/manager/3.log" Nov 26 02:40:31 crc kubenswrapper[4766]: I1126 02:40:31.227944 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-dm7xb_8969652f-14c9-4618-ad2d-1b3fd1ad3627/manager/2.log" Nov 26 02:40:31 crc kubenswrapper[4766]: I1126 02:40:31.307320 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-bfw5f_91d3e9c8-8bbc-43b9-947b-71c4bb474b21/kube-rbac-proxy/0.log" Nov 26 02:40:31 crc kubenswrapper[4766]: I1126 02:40:31.381406 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-bfw5f_91d3e9c8-8bbc-43b9-947b-71c4bb474b21/manager/3.log" Nov 26 02:40:31 crc kubenswrapper[4766]: I1126 02:40:31.395626 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-bfw5f_91d3e9c8-8bbc-43b9-947b-71c4bb474b21/manager/2.log" Nov 26 02:40:31 crc kubenswrapper[4766]: I1126 02:40:31.484379 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t_150d5aa9-549d-4b43-b5c3-e6c81e942146/kube-rbac-proxy/0.log" Nov 26 02:40:31 crc kubenswrapper[4766]: I1126 02:40:31.522350 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t_150d5aa9-549d-4b43-b5c3-e6c81e942146/manager/1.log" Nov 26 02:40:31 crc kubenswrapper[4766]: I1126 02:40:31.592498 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6b4bn8t_150d5aa9-549d-4b43-b5c3-e6c81e942146/manager/0.log" Nov 26 02:40:31 crc kubenswrapper[4766]: I1126 02:40:31.714765 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-86674bdb4b-vq98m_d00b9c60-0fb7-49c6-944a-fac598cc22a4/manager/2.log" Nov 26 02:40:31 crc kubenswrapper[4766]: I1126 02:40:31.877980 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7d5fdccdc4-5sh97_62c72069-cf63-4244-a55e-0028aaaed380/operator/1.log" Nov 26 02:40:31 crc kubenswrapper[4766]: I1126 02:40:31.991746 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-sp492_5244fd2a-7bec-4649-af25-c77d4ec96ed0/registry-server/0.log" Nov 26 02:40:32 crc kubenswrapper[4766]: I1126 02:40:32.047850 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7d5fdccdc4-5sh97_62c72069-cf63-4244-a55e-0028aaaed380/operator/0.log" Nov 26 02:40:32 crc kubenswrapper[4766]: I1126 02:40:32.147560 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-9zhql_a2e7c098-bf9d-4d99-84bf-adfb49ee6b12/kube-rbac-proxy/0.log" Nov 26 02:40:32 crc kubenswrapper[4766]: I1126 02:40:32.249427 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-9zhql_a2e7c098-bf9d-4d99-84bf-adfb49ee6b12/manager/2.log" Nov 26 02:40:32 crc kubenswrapper[4766]: I1126 02:40:32.260243 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-9zhql_a2e7c098-bf9d-4d99-84bf-adfb49ee6b12/manager/3.log" Nov 26 02:40:32 crc kubenswrapper[4766]: I1126 02:40:32.397567 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-xlxgj_97419f32-8768-495f-a652-8a1e8701546f/kube-rbac-proxy/0.log" Nov 26 02:40:32 crc kubenswrapper[4766]: I1126 02:40:32.484038 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-xlxgj_97419f32-8768-495f-a652-8a1e8701546f/manager/1.log" Nov 26 02:40:32 crc kubenswrapper[4766]: I1126 02:40:32.486775 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-xlxgj_97419f32-8768-495f-a652-8a1e8701546f/manager/2.log" Nov 26 02:40:32 crc kubenswrapper[4766]: I1126 02:40:32.631278 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-mx9vw_91ee52b5-0e2e-4732-87bf-809e378d33d3/operator/3.log" Nov 26 02:40:32 crc kubenswrapper[4766]: I1126 02:40:32.647903 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-mx9vw_91ee52b5-0e2e-4732-87bf-809e378d33d3/operator/2.log" Nov 26 02:40:32 crc kubenswrapper[4766]: I1126 02:40:32.682862 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-86674bdb4b-vq98m_d00b9c60-0fb7-49c6-944a-fac598cc22a4/manager/3.log" Nov 26 02:40:32 crc kubenswrapper[4766]: I1126 02:40:32.810737 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-2t9w2_00a32e6c-edcc-4625-ac35-26fed19f1638/manager/3.log" Nov 26 02:40:32 crc kubenswrapper[4766]: I1126 02:40:32.815111 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-2t9w2_00a32e6c-edcc-4625-ac35-26fed19f1638/kube-rbac-proxy/0.log" Nov 26 02:40:32 crc kubenswrapper[4766]: I1126 02:40:32.872442 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-2t9w2_00a32e6c-edcc-4625-ac35-26fed19f1638/manager/2.log" Nov 26 02:40:32 crc kubenswrapper[4766]: I1126 02:40:32.874535 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5ccf99cfdd-9cfft_04e53d09-de84-4682-b17b-61f9ea7bf152/kube-rbac-proxy/0.log" Nov 26 02:40:33 crc kubenswrapper[4766]: I1126 02:40:33.006745 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5ccf99cfdd-9cfft_04e53d09-de84-4682-b17b-61f9ea7bf152/manager/2.log" Nov 26 02:40:33 crc kubenswrapper[4766]: I1126 02:40:33.070917 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-glkl6_40780971-20ed-4b4c-8d19-e74f6582ef8c/kube-rbac-proxy/0.log" Nov 26 02:40:33 crc kubenswrapper[4766]: I1126 02:40:33.097854 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5ccf99cfdd-9cfft_04e53d09-de84-4682-b17b-61f9ea7bf152/manager/3.log" Nov 26 02:40:33 crc kubenswrapper[4766]: I1126 02:40:33.119475 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-glkl6_40780971-20ed-4b4c-8d19-e74f6582ef8c/manager/1.log" Nov 26 02:40:33 crc kubenswrapper[4766]: I1126 02:40:33.235055 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-glkl6_40780971-20ed-4b4c-8d19-e74f6582ef8c/manager/0.log" Nov 26 02:40:33 crc kubenswrapper[4766]: I1126 02:40:33.264267 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-2hpg8_751b8e60-46d4-4f5e-b62d-800f641ccd40/kube-rbac-proxy/0.log" Nov 26 02:40:33 crc kubenswrapper[4766]: I1126 02:40:33.307166 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-2hpg8_751b8e60-46d4-4f5e-b62d-800f641ccd40/manager/2.log" Nov 26 02:40:33 crc kubenswrapper[4766]: I1126 02:40:33.331894 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-2hpg8_751b8e60-46d4-4f5e-b62d-800f641ccd40/manager/3.log" Nov 26 02:40:41 crc kubenswrapper[4766]: I1126 02:40:41.480088 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:40:41 crc kubenswrapper[4766]: I1126 02:40:41.480644 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:40:51 crc kubenswrapper[4766]: I1126 02:40:51.132172 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-pm5j5_752819e1-d0ad-4820-979a-bf6ebfc59b4d/control-plane-machine-set-operator/0.log" Nov 26 02:40:51 crc kubenswrapper[4766]: I1126 02:40:51.292363 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-x2vlb_c8260737-caa6-42f1-8854-ad2b97a1ee42/kube-rbac-proxy/0.log" Nov 26 02:40:51 crc kubenswrapper[4766]: I1126 02:40:51.314015 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-x2vlb_c8260737-caa6-42f1-8854-ad2b97a1ee42/machine-api-operator/0.log" Nov 26 02:41:04 crc kubenswrapper[4766]: I1126 02:41:04.906985 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-nd9m2_954f9d95-f1ab-4e36-9b7a-d16fbe282e53/cert-manager-controller/1.log" Nov 26 02:41:04 crc kubenswrapper[4766]: I1126 02:41:04.984827 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-nd9m2_954f9d95-f1ab-4e36-9b7a-d16fbe282e53/cert-manager-controller/0.log" Nov 26 02:41:05 crc kubenswrapper[4766]: I1126 02:41:05.138412 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-qjmjr_e6ee9725-56d1-426a-afeb-c00b1abb9e2a/cert-manager-cainjector/0.log" Nov 26 02:41:05 crc kubenswrapper[4766]: I1126 02:41:05.242554 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-btczr_403aadd6-8bd1-4446-b0e7-f342f31e54c9/cert-manager-webhook/0.log" Nov 26 02:41:07 crc kubenswrapper[4766]: I1126 02:41:07.294764 4766 scope.go:117] "RemoveContainer" containerID="6d9f6e9757d90f17eba19feba3f6801c06dca432db823fa085eda6913b44d04a" Nov 26 02:41:07 crc kubenswrapper[4766]: I1126 02:41:07.324646 4766 scope.go:117] "RemoveContainer" containerID="065d0d3a755ab9e468d894c736971bc9103af510ea09537839d2185983023261" Nov 26 02:41:07 crc kubenswrapper[4766]: I1126 02:41:07.389820 4766 scope.go:117] "RemoveContainer" containerID="4b683313ddaf5a330416b8bdf43dee1e2ac479b33f0bc46dc6898ca2a4028b4b" Nov 26 02:41:11 crc kubenswrapper[4766]: I1126 02:41:11.480304 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:41:11 crc kubenswrapper[4766]: I1126 02:41:11.480917 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:41:19 crc kubenswrapper[4766]: I1126 02:41:19.103376 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-nzhrs_d5bae28b-f4e5-4d7c-9f46-2b6cb7c9f130/nmstate-console-plugin/0.log" Nov 26 02:41:19 crc kubenswrapper[4766]: I1126 02:41:19.249728 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-q4sb9_58a75814-134a-4a52-aae0-5c21109f658f/nmstate-handler/0.log" Nov 26 02:41:19 crc kubenswrapper[4766]: I1126 02:41:19.258288 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-f58l6_8f6f1e2d-6ebd-4f98-bc08-8cba715f33ae/kube-rbac-proxy/0.log" Nov 26 02:41:19 crc kubenswrapper[4766]: I1126 02:41:19.338078 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-f58l6_8f6f1e2d-6ebd-4f98-bc08-8cba715f33ae/nmstate-metrics/0.log" Nov 26 02:41:19 crc kubenswrapper[4766]: I1126 02:41:19.442571 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-ztjbb_61f41116-4bc2-47c5-b805-d0a0990c6024/nmstate-operator/0.log" Nov 26 02:41:19 crc kubenswrapper[4766]: I1126 02:41:19.572398 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-mn64d_d874f3cf-9f30-431e-81e1-6334a0cf4351/nmstate-webhook/0.log" Nov 26 02:41:33 crc kubenswrapper[4766]: I1126 02:41:33.266393 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-7bbf4564c5-m77vx_b90f7164-a5e9-452f-8733-f549f15e17ec/kube-rbac-proxy/0.log" Nov 26 02:41:33 crc kubenswrapper[4766]: I1126 02:41:33.287716 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-7bbf4564c5-m77vx_b90f7164-a5e9-452f-8733-f549f15e17ec/manager/1.log" Nov 26 02:41:33 crc kubenswrapper[4766]: I1126 02:41:33.466298 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-7bbf4564c5-m77vx_b90f7164-a5e9-452f-8733-f549f15e17ec/manager/0.log" Nov 26 02:41:36 crc kubenswrapper[4766]: I1126 02:41:36.079120 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-68nzx"] Nov 26 02:41:36 crc kubenswrapper[4766]: E1126 02:41:36.079943 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a" containerName="registry-server" Nov 26 02:41:36 crc kubenswrapper[4766]: I1126 02:41:36.079955 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a" containerName="registry-server" Nov 26 02:41:36 crc kubenswrapper[4766]: E1126 02:41:36.079965 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a" containerName="extract-utilities" Nov 26 02:41:36 crc kubenswrapper[4766]: I1126 02:41:36.079971 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a" containerName="extract-utilities" Nov 26 02:41:36 crc kubenswrapper[4766]: E1126 02:41:36.079986 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a" containerName="extract-content" Nov 26 02:41:36 crc kubenswrapper[4766]: I1126 02:41:36.079991 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a" containerName="extract-content" Nov 26 02:41:36 crc kubenswrapper[4766]: I1126 02:41:36.080218 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b0a5e8d-17b2-4cab-9143-a7d2d2498f6a" containerName="registry-server" Nov 26 02:41:36 crc kubenswrapper[4766]: I1126 02:41:36.081865 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-68nzx" Nov 26 02:41:36 crc kubenswrapper[4766]: I1126 02:41:36.101783 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-68nzx"] Nov 26 02:41:36 crc kubenswrapper[4766]: I1126 02:41:36.221997 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8e40b61-1f16-47cf-9e5e-b260721f3ad9-catalog-content\") pod \"redhat-operators-68nzx\" (UID: \"e8e40b61-1f16-47cf-9e5e-b260721f3ad9\") " pod="openshift-marketplace/redhat-operators-68nzx" Nov 26 02:41:36 crc kubenswrapper[4766]: I1126 02:41:36.222094 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvmg2\" (UniqueName: \"kubernetes.io/projected/e8e40b61-1f16-47cf-9e5e-b260721f3ad9-kube-api-access-jvmg2\") pod \"redhat-operators-68nzx\" (UID: \"e8e40b61-1f16-47cf-9e5e-b260721f3ad9\") " pod="openshift-marketplace/redhat-operators-68nzx" Nov 26 02:41:36 crc kubenswrapper[4766]: I1126 02:41:36.222123 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8e40b61-1f16-47cf-9e5e-b260721f3ad9-utilities\") pod \"redhat-operators-68nzx\" (UID: \"e8e40b61-1f16-47cf-9e5e-b260721f3ad9\") " pod="openshift-marketplace/redhat-operators-68nzx" Nov 26 02:41:36 crc kubenswrapper[4766]: I1126 02:41:36.324418 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8e40b61-1f16-47cf-9e5e-b260721f3ad9-catalog-content\") pod \"redhat-operators-68nzx\" (UID: \"e8e40b61-1f16-47cf-9e5e-b260721f3ad9\") " pod="openshift-marketplace/redhat-operators-68nzx" Nov 26 02:41:36 crc kubenswrapper[4766]: I1126 02:41:36.325286 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvmg2\" (UniqueName: \"kubernetes.io/projected/e8e40b61-1f16-47cf-9e5e-b260721f3ad9-kube-api-access-jvmg2\") pod \"redhat-operators-68nzx\" (UID: \"e8e40b61-1f16-47cf-9e5e-b260721f3ad9\") " pod="openshift-marketplace/redhat-operators-68nzx" Nov 26 02:41:36 crc kubenswrapper[4766]: I1126 02:41:36.325439 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8e40b61-1f16-47cf-9e5e-b260721f3ad9-utilities\") pod \"redhat-operators-68nzx\" (UID: \"e8e40b61-1f16-47cf-9e5e-b260721f3ad9\") " pod="openshift-marketplace/redhat-operators-68nzx" Nov 26 02:41:36 crc kubenswrapper[4766]: I1126 02:41:36.325138 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8e40b61-1f16-47cf-9e5e-b260721f3ad9-catalog-content\") pod \"redhat-operators-68nzx\" (UID: \"e8e40b61-1f16-47cf-9e5e-b260721f3ad9\") " pod="openshift-marketplace/redhat-operators-68nzx" Nov 26 02:41:36 crc kubenswrapper[4766]: I1126 02:41:36.325995 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8e40b61-1f16-47cf-9e5e-b260721f3ad9-utilities\") pod \"redhat-operators-68nzx\" (UID: \"e8e40b61-1f16-47cf-9e5e-b260721f3ad9\") " pod="openshift-marketplace/redhat-operators-68nzx" Nov 26 02:41:36 crc kubenswrapper[4766]: I1126 02:41:36.346573 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvmg2\" (UniqueName: \"kubernetes.io/projected/e8e40b61-1f16-47cf-9e5e-b260721f3ad9-kube-api-access-jvmg2\") pod \"redhat-operators-68nzx\" (UID: \"e8e40b61-1f16-47cf-9e5e-b260721f3ad9\") " pod="openshift-marketplace/redhat-operators-68nzx" Nov 26 02:41:36 crc kubenswrapper[4766]: I1126 02:41:36.403453 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-68nzx" Nov 26 02:41:37 crc kubenswrapper[4766]: I1126 02:41:37.192037 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-68nzx"] Nov 26 02:41:37 crc kubenswrapper[4766]: I1126 02:41:37.710505 4766 generic.go:334] "Generic (PLEG): container finished" podID="e8e40b61-1f16-47cf-9e5e-b260721f3ad9" containerID="5d64cae43f5a362ab1164231215086b150a2cb8e53ae238b9080d1207690756c" exitCode=0 Nov 26 02:41:37 crc kubenswrapper[4766]: I1126 02:41:37.710614 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68nzx" event={"ID":"e8e40b61-1f16-47cf-9e5e-b260721f3ad9","Type":"ContainerDied","Data":"5d64cae43f5a362ab1164231215086b150a2cb8e53ae238b9080d1207690756c"} Nov 26 02:41:37 crc kubenswrapper[4766]: I1126 02:41:37.710833 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68nzx" event={"ID":"e8e40b61-1f16-47cf-9e5e-b260721f3ad9","Type":"ContainerStarted","Data":"d3ae93bf8ba1f566f062025e8235dca02adb04d016b45fd8976bf442d9cb9a8b"} Nov 26 02:41:38 crc kubenswrapper[4766]: I1126 02:41:38.723231 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68nzx" event={"ID":"e8e40b61-1f16-47cf-9e5e-b260721f3ad9","Type":"ContainerStarted","Data":"412e7b3d4b35d9aaa9db72a34610e676dbc92fbdc48da87a7b00eeed2da47677"} Nov 26 02:41:41 crc kubenswrapper[4766]: I1126 02:41:41.479832 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:41:41 crc kubenswrapper[4766]: I1126 02:41:41.480434 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:41:41 crc kubenswrapper[4766]: I1126 02:41:41.480508 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 02:41:41 crc kubenswrapper[4766]: I1126 02:41:41.481954 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"93d6a4a6ec497ff7bc4cee7766539719d24858cdc0bc941ecde4de3cf6ff02cd"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 02:41:41 crc kubenswrapper[4766]: I1126 02:41:41.482061 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://93d6a4a6ec497ff7bc4cee7766539719d24858cdc0bc941ecde4de3cf6ff02cd" gracePeriod=600 Nov 26 02:41:42 crc kubenswrapper[4766]: I1126 02:41:42.761260 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="93d6a4a6ec497ff7bc4cee7766539719d24858cdc0bc941ecde4de3cf6ff02cd" exitCode=0 Nov 26 02:41:42 crc kubenswrapper[4766]: I1126 02:41:42.761381 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"93d6a4a6ec497ff7bc4cee7766539719d24858cdc0bc941ecde4de3cf6ff02cd"} Nov 26 02:41:42 crc kubenswrapper[4766]: I1126 02:41:42.761801 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerStarted","Data":"fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3"} Nov 26 02:41:42 crc kubenswrapper[4766]: I1126 02:41:42.761834 4766 scope.go:117] "RemoveContainer" containerID="a7d6e748576ae86e244e4ef4ea393725046b64f3769b59d942f31b1f8e33503a" Nov 26 02:41:42 crc kubenswrapper[4766]: I1126 02:41:42.766018 4766 generic.go:334] "Generic (PLEG): container finished" podID="e8e40b61-1f16-47cf-9e5e-b260721f3ad9" containerID="412e7b3d4b35d9aaa9db72a34610e676dbc92fbdc48da87a7b00eeed2da47677" exitCode=0 Nov 26 02:41:42 crc kubenswrapper[4766]: I1126 02:41:42.766053 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68nzx" event={"ID":"e8e40b61-1f16-47cf-9e5e-b260721f3ad9","Type":"ContainerDied","Data":"412e7b3d4b35d9aaa9db72a34610e676dbc92fbdc48da87a7b00eeed2da47677"} Nov 26 02:41:43 crc kubenswrapper[4766]: I1126 02:41:43.785094 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68nzx" event={"ID":"e8e40b61-1f16-47cf-9e5e-b260721f3ad9","Type":"ContainerStarted","Data":"8100c3e8efe482a5e07eca96e5d7788cc282ce31fbbc71e4eb27c449783b662e"} Nov 26 02:41:43 crc kubenswrapper[4766]: I1126 02:41:43.824681 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-68nzx" podStartSLOduration=2.352062382 podStartE2EDuration="7.823211094s" podCreationTimestamp="2025-11-26 02:41:36 +0000 UTC" firstStartedPulling="2025-11-26 02:41:37.712420643 +0000 UTC m=+8278.561191073" lastFinishedPulling="2025-11-26 02:41:43.183569345 +0000 UTC m=+8284.032339785" observedRunningTime="2025-11-26 02:41:43.807803247 +0000 UTC m=+8284.656573687" watchObservedRunningTime="2025-11-26 02:41:43.823211094 +0000 UTC m=+8284.671981534" Nov 26 02:41:46 crc kubenswrapper[4766]: I1126 02:41:46.404267 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-68nzx" Nov 26 02:41:46 crc kubenswrapper[4766]: I1126 02:41:46.405728 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-68nzx" Nov 26 02:41:47 crc kubenswrapper[4766]: I1126 02:41:47.460083 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-68nzx" podUID="e8e40b61-1f16-47cf-9e5e-b260721f3ad9" containerName="registry-server" probeResult="failure" output=< Nov 26 02:41:47 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 02:41:47 crc kubenswrapper[4766]: > Nov 26 02:41:51 crc kubenswrapper[4766]: I1126 02:41:51.194360 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-ff9846bd-mq2w5_2850bfaf-5303-44e6-8fc6-616612cb68e1/cluster-logging-operator/0.log" Nov 26 02:41:51 crc kubenswrapper[4766]: I1126 02:41:51.374276 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-zzrrp_bed1e93e-177c-4e72-99b7-e35f6d8f6cf9/collector/0.log" Nov 26 02:41:51 crc kubenswrapper[4766]: I1126 02:41:51.377360 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_4a6f0b48-aeb6-4df1-ba96-6fcc9c31b08e/loki-compactor/0.log" Nov 26 02:41:51 crc kubenswrapper[4766]: I1126 02:41:51.596014 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-76cc67bf56-7khhv_bfb11063-6ad6-463b-87e6-eaf9333e2d4a/loki-distributor/0.log" Nov 26 02:41:51 crc kubenswrapper[4766]: I1126 02:41:51.646675 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-5d9974794d-bbcbk_261b6830-cc71-4704-ab0d-fd3a9708a786/gateway/0.log" Nov 26 02:41:51 crc kubenswrapper[4766]: I1126 02:41:51.738672 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-5d9974794d-bbcbk_261b6830-cc71-4704-ab0d-fd3a9708a786/opa/0.log" Nov 26 02:41:51 crc kubenswrapper[4766]: I1126 02:41:51.774260 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-5d9974794d-hgkj4_bc3b1ab0-39c8-42a1-b0d1-c45baa43df87/gateway/0.log" Nov 26 02:41:51 crc kubenswrapper[4766]: I1126 02:41:51.866241 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-5d9974794d-hgkj4_bc3b1ab0-39c8-42a1-b0d1-c45baa43df87/opa/0.log" Nov 26 02:41:51 crc kubenswrapper[4766]: I1126 02:41:51.998292 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_87bee09b-03e7-4b6c-a28f-7e12aab1eb45/loki-index-gateway/0.log" Nov 26 02:41:52 crc kubenswrapper[4766]: I1126 02:41:52.124558 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_aae58301-788d-430a-b020-a5f530d70c5a/loki-ingester/0.log" Nov 26 02:41:52 crc kubenswrapper[4766]: I1126 02:41:52.385250 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-5895d59bb8-tv6v8_036dedff-92b9-4ad9-89b4-15dfb60e73c2/loki-querier/0.log" Nov 26 02:41:52 crc kubenswrapper[4766]: I1126 02:41:52.476546 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-84558f7c9f-tltn8_7fb68151-644a-4d8b-9274-18bb7073cc41/loki-query-frontend/0.log" Nov 26 02:41:57 crc kubenswrapper[4766]: I1126 02:41:57.459520 4766 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-68nzx" podUID="e8e40b61-1f16-47cf-9e5e-b260721f3ad9" containerName="registry-server" probeResult="failure" output=< Nov 26 02:41:57 crc kubenswrapper[4766]: timeout: failed to connect service ":50051" within 1s Nov 26 02:41:57 crc kubenswrapper[4766]: > Nov 26 02:42:06 crc kubenswrapper[4766]: I1126 02:42:06.479502 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-68nzx" Nov 26 02:42:06 crc kubenswrapper[4766]: I1126 02:42:06.572043 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-68nzx" Nov 26 02:42:06 crc kubenswrapper[4766]: I1126 02:42:06.728526 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-68nzx"] Nov 26 02:42:08 crc kubenswrapper[4766]: I1126 02:42:08.065803 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-68nzx" podUID="e8e40b61-1f16-47cf-9e5e-b260721f3ad9" containerName="registry-server" containerID="cri-o://8100c3e8efe482a5e07eca96e5d7788cc282ce31fbbc71e4eb27c449783b662e" gracePeriod=2 Nov 26 02:42:08 crc kubenswrapper[4766]: I1126 02:42:08.699043 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-9dzn2_0c9c337a-681d-4a68-8e51-5d690a326fe7/kube-rbac-proxy/0.log" Nov 26 02:42:08 crc kubenswrapper[4766]: I1126 02:42:08.708162 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-68nzx" Nov 26 02:42:08 crc kubenswrapper[4766]: I1126 02:42:08.801446 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8e40b61-1f16-47cf-9e5e-b260721f3ad9-catalog-content\") pod \"e8e40b61-1f16-47cf-9e5e-b260721f3ad9\" (UID: \"e8e40b61-1f16-47cf-9e5e-b260721f3ad9\") " Nov 26 02:42:08 crc kubenswrapper[4766]: I1126 02:42:08.803154 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8e40b61-1f16-47cf-9e5e-b260721f3ad9-utilities\") pod \"e8e40b61-1f16-47cf-9e5e-b260721f3ad9\" (UID: \"e8e40b61-1f16-47cf-9e5e-b260721f3ad9\") " Nov 26 02:42:08 crc kubenswrapper[4766]: I1126 02:42:08.803271 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvmg2\" (UniqueName: \"kubernetes.io/projected/e8e40b61-1f16-47cf-9e5e-b260721f3ad9-kube-api-access-jvmg2\") pod \"e8e40b61-1f16-47cf-9e5e-b260721f3ad9\" (UID: \"e8e40b61-1f16-47cf-9e5e-b260721f3ad9\") " Nov 26 02:42:08 crc kubenswrapper[4766]: I1126 02:42:08.804784 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8e40b61-1f16-47cf-9e5e-b260721f3ad9-utilities" (OuterVolumeSpecName: "utilities") pod "e8e40b61-1f16-47cf-9e5e-b260721f3ad9" (UID: "e8e40b61-1f16-47cf-9e5e-b260721f3ad9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:42:08 crc kubenswrapper[4766]: I1126 02:42:08.806818 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8e40b61-1f16-47cf-9e5e-b260721f3ad9-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 02:42:08 crc kubenswrapper[4766]: I1126 02:42:08.812035 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8e40b61-1f16-47cf-9e5e-b260721f3ad9-kube-api-access-jvmg2" (OuterVolumeSpecName: "kube-api-access-jvmg2") pod "e8e40b61-1f16-47cf-9e5e-b260721f3ad9" (UID: "e8e40b61-1f16-47cf-9e5e-b260721f3ad9"). InnerVolumeSpecName "kube-api-access-jvmg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:42:08 crc kubenswrapper[4766]: I1126 02:42:08.898317 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8e40b61-1f16-47cf-9e5e-b260721f3ad9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e8e40b61-1f16-47cf-9e5e-b260721f3ad9" (UID: "e8e40b61-1f16-47cf-9e5e-b260721f3ad9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:42:08 crc kubenswrapper[4766]: I1126 02:42:08.909322 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8e40b61-1f16-47cf-9e5e-b260721f3ad9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 02:42:08 crc kubenswrapper[4766]: I1126 02:42:08.909687 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvmg2\" (UniqueName: \"kubernetes.io/projected/e8e40b61-1f16-47cf-9e5e-b260721f3ad9-kube-api-access-jvmg2\") on node \"crc\" DevicePath \"\"" Nov 26 02:42:08 crc kubenswrapper[4766]: I1126 02:42:08.920627 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-9dzn2_0c9c337a-681d-4a68-8e51-5d690a326fe7/controller/0.log" Nov 26 02:42:08 crc kubenswrapper[4766]: I1126 02:42:08.982519 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-frr-files/0.log" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.077487 4766 generic.go:334] "Generic (PLEG): container finished" podID="e8e40b61-1f16-47cf-9e5e-b260721f3ad9" containerID="8100c3e8efe482a5e07eca96e5d7788cc282ce31fbbc71e4eb27c449783b662e" exitCode=0 Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.077526 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68nzx" event={"ID":"e8e40b61-1f16-47cf-9e5e-b260721f3ad9","Type":"ContainerDied","Data":"8100c3e8efe482a5e07eca96e5d7788cc282ce31fbbc71e4eb27c449783b662e"} Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.077554 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68nzx" event={"ID":"e8e40b61-1f16-47cf-9e5e-b260721f3ad9","Type":"ContainerDied","Data":"d3ae93bf8ba1f566f062025e8235dca02adb04d016b45fd8976bf442d9cb9a8b"} Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.077556 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-68nzx" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.077572 4766 scope.go:117] "RemoveContainer" containerID="8100c3e8efe482a5e07eca96e5d7788cc282ce31fbbc71e4eb27c449783b662e" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.124357 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-68nzx"] Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.128363 4766 scope.go:117] "RemoveContainer" containerID="412e7b3d4b35d9aaa9db72a34610e676dbc92fbdc48da87a7b00eeed2da47677" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.144217 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-68nzx"] Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.189526 4766 scope.go:117] "RemoveContainer" containerID="5d64cae43f5a362ab1164231215086b150a2cb8e53ae238b9080d1207690756c" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.247350 4766 scope.go:117] "RemoveContainer" containerID="8100c3e8efe482a5e07eca96e5d7788cc282ce31fbbc71e4eb27c449783b662e" Nov 26 02:42:09 crc kubenswrapper[4766]: E1126 02:42:09.248862 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8100c3e8efe482a5e07eca96e5d7788cc282ce31fbbc71e4eb27c449783b662e\": container with ID starting with 8100c3e8efe482a5e07eca96e5d7788cc282ce31fbbc71e4eb27c449783b662e not found: ID does not exist" containerID="8100c3e8efe482a5e07eca96e5d7788cc282ce31fbbc71e4eb27c449783b662e" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.248897 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8100c3e8efe482a5e07eca96e5d7788cc282ce31fbbc71e4eb27c449783b662e"} err="failed to get container status \"8100c3e8efe482a5e07eca96e5d7788cc282ce31fbbc71e4eb27c449783b662e\": rpc error: code = NotFound desc = could not find container \"8100c3e8efe482a5e07eca96e5d7788cc282ce31fbbc71e4eb27c449783b662e\": container with ID starting with 8100c3e8efe482a5e07eca96e5d7788cc282ce31fbbc71e4eb27c449783b662e not found: ID does not exist" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.248916 4766 scope.go:117] "RemoveContainer" containerID="412e7b3d4b35d9aaa9db72a34610e676dbc92fbdc48da87a7b00eeed2da47677" Nov 26 02:42:09 crc kubenswrapper[4766]: E1126 02:42:09.249314 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"412e7b3d4b35d9aaa9db72a34610e676dbc92fbdc48da87a7b00eeed2da47677\": container with ID starting with 412e7b3d4b35d9aaa9db72a34610e676dbc92fbdc48da87a7b00eeed2da47677 not found: ID does not exist" containerID="412e7b3d4b35d9aaa9db72a34610e676dbc92fbdc48da87a7b00eeed2da47677" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.249353 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"412e7b3d4b35d9aaa9db72a34610e676dbc92fbdc48da87a7b00eeed2da47677"} err="failed to get container status \"412e7b3d4b35d9aaa9db72a34610e676dbc92fbdc48da87a7b00eeed2da47677\": rpc error: code = NotFound desc = could not find container \"412e7b3d4b35d9aaa9db72a34610e676dbc92fbdc48da87a7b00eeed2da47677\": container with ID starting with 412e7b3d4b35d9aaa9db72a34610e676dbc92fbdc48da87a7b00eeed2da47677 not found: ID does not exist" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.249381 4766 scope.go:117] "RemoveContainer" containerID="5d64cae43f5a362ab1164231215086b150a2cb8e53ae238b9080d1207690756c" Nov 26 02:42:09 crc kubenswrapper[4766]: E1126 02:42:09.255750 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d64cae43f5a362ab1164231215086b150a2cb8e53ae238b9080d1207690756c\": container with ID starting with 5d64cae43f5a362ab1164231215086b150a2cb8e53ae238b9080d1207690756c not found: ID does not exist" containerID="5d64cae43f5a362ab1164231215086b150a2cb8e53ae238b9080d1207690756c" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.255785 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d64cae43f5a362ab1164231215086b150a2cb8e53ae238b9080d1207690756c"} err="failed to get container status \"5d64cae43f5a362ab1164231215086b150a2cb8e53ae238b9080d1207690756c\": rpc error: code = NotFound desc = could not find container \"5d64cae43f5a362ab1164231215086b150a2cb8e53ae238b9080d1207690756c\": container with ID starting with 5d64cae43f5a362ab1164231215086b150a2cb8e53ae238b9080d1207690756c not found: ID does not exist" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.332684 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-reloader/0.log" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.337804 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-frr-files/0.log" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.365620 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-reloader/0.log" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.370707 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-metrics/0.log" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.527805 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-metrics/0.log" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.577570 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-frr-files/0.log" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.596680 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-reloader/0.log" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.616000 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-metrics/0.log" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.741565 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-frr-files/0.log" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.753869 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-reloader/0.log" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.769293 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/cp-metrics/0.log" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.825604 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/controller/0.log" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.837164 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8e40b61-1f16-47cf-9e5e-b260721f3ad9" path="/var/lib/kubelet/pods/e8e40b61-1f16-47cf-9e5e-b260721f3ad9/volumes" Nov 26 02:42:09 crc kubenswrapper[4766]: I1126 02:42:09.935026 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/kube-rbac-proxy/0.log" Nov 26 02:42:10 crc kubenswrapper[4766]: I1126 02:42:10.002184 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/frr-metrics/0.log" Nov 26 02:42:10 crc kubenswrapper[4766]: I1126 02:42:10.087669 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/kube-rbac-proxy-frr/0.log" Nov 26 02:42:10 crc kubenswrapper[4766]: I1126 02:42:10.194541 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/reloader/0.log" Nov 26 02:42:10 crc kubenswrapper[4766]: I1126 02:42:10.320688 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-k6vct_88faf6a0-1824-43f6-b711-09cb78c54a5d/frr-k8s-webhook-server/0.log" Nov 26 02:42:10 crc kubenswrapper[4766]: I1126 02:42:10.489490 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-764c8cf56f-vrwpt_fd3a6275-9f47-4799-bc15-96a51b93f31b/manager/3.log" Nov 26 02:42:10 crc kubenswrapper[4766]: I1126 02:42:10.540184 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-764c8cf56f-vrwpt_fd3a6275-9f47-4799-bc15-96a51b93f31b/manager/2.log" Nov 26 02:42:10 crc kubenswrapper[4766]: I1126 02:42:10.793170 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6f5d8f677c-qjhpt_ce7c66d0-5857-46ce-a725-6531c37891cb/webhook-server/0.log" Nov 26 02:42:11 crc kubenswrapper[4766]: I1126 02:42:11.020953 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-d5tdc_76a9ad7d-db95-4177-839f-db3e7cc1ac81/kube-rbac-proxy/0.log" Nov 26 02:42:11 crc kubenswrapper[4766]: I1126 02:42:11.757097 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-d5tdc_76a9ad7d-db95-4177-839f-db3e7cc1ac81/speaker/0.log" Nov 26 02:42:12 crc kubenswrapper[4766]: I1126 02:42:12.370146 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rtx6_931efe26-0d06-4bcb-bbaa-f65225f02419/frr/0.log" Nov 26 02:42:26 crc kubenswrapper[4766]: I1126 02:42:26.110113 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58_64a27a6e-0ec0-42e2-8606-370bafc71c04/util/0.log" Nov 26 02:42:26 crc kubenswrapper[4766]: I1126 02:42:26.333208 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58_64a27a6e-0ec0-42e2-8606-370bafc71c04/util/0.log" Nov 26 02:42:26 crc kubenswrapper[4766]: I1126 02:42:26.363787 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58_64a27a6e-0ec0-42e2-8606-370bafc71c04/pull/0.log" Nov 26 02:42:26 crc kubenswrapper[4766]: I1126 02:42:26.366608 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58_64a27a6e-0ec0-42e2-8606-370bafc71c04/pull/0.log" Nov 26 02:42:26 crc kubenswrapper[4766]: I1126 02:42:26.564965 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58_64a27a6e-0ec0-42e2-8606-370bafc71c04/util/0.log" Nov 26 02:42:26 crc kubenswrapper[4766]: I1126 02:42:26.611138 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58_64a27a6e-0ec0-42e2-8606-370bafc71c04/pull/0.log" Nov 26 02:42:26 crc kubenswrapper[4766]: I1126 02:42:26.631036 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb87tk58_64a27a6e-0ec0-42e2-8606-370bafc71c04/extract/0.log" Nov 26 02:42:26 crc kubenswrapper[4766]: I1126 02:42:26.751396 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm_9e6e8bab-7232-4619-a55d-95125dfaf6d0/util/0.log" Nov 26 02:42:26 crc kubenswrapper[4766]: I1126 02:42:26.945451 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm_9e6e8bab-7232-4619-a55d-95125dfaf6d0/util/0.log" Nov 26 02:42:26 crc kubenswrapper[4766]: I1126 02:42:26.958801 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm_9e6e8bab-7232-4619-a55d-95125dfaf6d0/pull/0.log" Nov 26 02:42:26 crc kubenswrapper[4766]: I1126 02:42:26.974824 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm_9e6e8bab-7232-4619-a55d-95125dfaf6d0/pull/0.log" Nov 26 02:42:27 crc kubenswrapper[4766]: I1126 02:42:27.076932 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm_9e6e8bab-7232-4619-a55d-95125dfaf6d0/util/0.log" Nov 26 02:42:27 crc kubenswrapper[4766]: I1126 02:42:27.124744 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm_9e6e8bab-7232-4619-a55d-95125dfaf6d0/pull/0.log" Nov 26 02:42:27 crc kubenswrapper[4766]: I1126 02:42:27.180317 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ets4wm_9e6e8bab-7232-4619-a55d-95125dfaf6d0/extract/0.log" Nov 26 02:42:27 crc kubenswrapper[4766]: I1126 02:42:27.287709 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq_4d00d8e8-b83e-4585-9678-fb64e663fec3/util/0.log" Nov 26 02:42:27 crc kubenswrapper[4766]: I1126 02:42:27.466819 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq_4d00d8e8-b83e-4585-9678-fb64e663fec3/pull/0.log" Nov 26 02:42:27 crc kubenswrapper[4766]: I1126 02:42:27.487066 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq_4d00d8e8-b83e-4585-9678-fb64e663fec3/util/0.log" Nov 26 02:42:27 crc kubenswrapper[4766]: I1126 02:42:27.510257 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq_4d00d8e8-b83e-4585-9678-fb64e663fec3/pull/0.log" Nov 26 02:42:27 crc kubenswrapper[4766]: I1126 02:42:27.668842 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq_4d00d8e8-b83e-4585-9678-fb64e663fec3/extract/0.log" Nov 26 02:42:27 crc kubenswrapper[4766]: I1126 02:42:27.682949 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq_4d00d8e8-b83e-4585-9678-fb64e663fec3/pull/0.log" Nov 26 02:42:27 crc kubenswrapper[4766]: I1126 02:42:27.685253 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92108bhpq_4d00d8e8-b83e-4585-9678-fb64e663fec3/util/0.log" Nov 26 02:42:27 crc kubenswrapper[4766]: I1126 02:42:27.868624 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9_36c1a0d9-7873-4ce0-b926-17e535470807/util/0.log" Nov 26 02:42:28 crc kubenswrapper[4766]: I1126 02:42:28.253625 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9_36c1a0d9-7873-4ce0-b926-17e535470807/util/0.log" Nov 26 02:42:28 crc kubenswrapper[4766]: I1126 02:42:28.274184 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9_36c1a0d9-7873-4ce0-b926-17e535470807/pull/0.log" Nov 26 02:42:28 crc kubenswrapper[4766]: I1126 02:42:28.305107 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9_36c1a0d9-7873-4ce0-b926-17e535470807/pull/0.log" Nov 26 02:42:28 crc kubenswrapper[4766]: I1126 02:42:28.437182 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9_36c1a0d9-7873-4ce0-b926-17e535470807/util/0.log" Nov 26 02:42:28 crc kubenswrapper[4766]: I1126 02:42:28.463393 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9_36c1a0d9-7873-4ce0-b926-17e535470807/pull/0.log" Nov 26 02:42:28 crc kubenswrapper[4766]: I1126 02:42:28.523197 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsqcb9_36c1a0d9-7873-4ce0-b926-17e535470807/extract/0.log" Nov 26 02:42:28 crc kubenswrapper[4766]: I1126 02:42:28.629699 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvvwl_74038ed5-c727-40ae-bc0f-d779aa47613d/extract-utilities/0.log" Nov 26 02:42:28 crc kubenswrapper[4766]: I1126 02:42:28.853427 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvvwl_74038ed5-c727-40ae-bc0f-d779aa47613d/extract-content/0.log" Nov 26 02:42:28 crc kubenswrapper[4766]: I1126 02:42:28.870187 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvvwl_74038ed5-c727-40ae-bc0f-d779aa47613d/extract-content/0.log" Nov 26 02:42:28 crc kubenswrapper[4766]: I1126 02:42:28.905589 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvvwl_74038ed5-c727-40ae-bc0f-d779aa47613d/extract-utilities/0.log" Nov 26 02:42:29 crc kubenswrapper[4766]: I1126 02:42:29.108769 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvvwl_74038ed5-c727-40ae-bc0f-d779aa47613d/extract-content/0.log" Nov 26 02:42:29 crc kubenswrapper[4766]: I1126 02:42:29.129555 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvvwl_74038ed5-c727-40ae-bc0f-d779aa47613d/extract-utilities/0.log" Nov 26 02:42:29 crc kubenswrapper[4766]: I1126 02:42:29.295513 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6vj5j_e1f0994c-ef19-451f-bc61-fe7b66c3399b/extract-utilities/0.log" Nov 26 02:42:29 crc kubenswrapper[4766]: I1126 02:42:29.594935 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6vj5j_e1f0994c-ef19-451f-bc61-fe7b66c3399b/extract-utilities/0.log" Nov 26 02:42:29 crc kubenswrapper[4766]: I1126 02:42:29.651561 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6vj5j_e1f0994c-ef19-451f-bc61-fe7b66c3399b/extract-content/0.log" Nov 26 02:42:29 crc kubenswrapper[4766]: I1126 02:42:29.658948 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6vj5j_e1f0994c-ef19-451f-bc61-fe7b66c3399b/extract-content/0.log" Nov 26 02:42:29 crc kubenswrapper[4766]: I1126 02:42:29.707629 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvvwl_74038ed5-c727-40ae-bc0f-d779aa47613d/registry-server/0.log" Nov 26 02:42:29 crc kubenswrapper[4766]: I1126 02:42:29.775365 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6vj5j_e1f0994c-ef19-451f-bc61-fe7b66c3399b/extract-utilities/0.log" Nov 26 02:42:29 crc kubenswrapper[4766]: I1126 02:42:29.796030 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6vj5j_e1f0994c-ef19-451f-bc61-fe7b66c3399b/extract-content/0.log" Nov 26 02:42:29 crc kubenswrapper[4766]: I1126 02:42:29.927588 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk_631bfaa9-4706-44f7-9417-500418a18c73/util/0.log" Nov 26 02:42:30 crc kubenswrapper[4766]: I1126 02:42:30.215968 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk_631bfaa9-4706-44f7-9417-500418a18c73/pull/0.log" Nov 26 02:42:30 crc kubenswrapper[4766]: I1126 02:42:30.225798 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk_631bfaa9-4706-44f7-9417-500418a18c73/util/0.log" Nov 26 02:42:30 crc kubenswrapper[4766]: I1126 02:42:30.297446 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk_631bfaa9-4706-44f7-9417-500418a18c73/pull/0.log" Nov 26 02:42:30 crc kubenswrapper[4766]: I1126 02:42:30.349751 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6vj5j_e1f0994c-ef19-451f-bc61-fe7b66c3399b/registry-server/0.log" Nov 26 02:42:30 crc kubenswrapper[4766]: I1126 02:42:30.419592 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk_631bfaa9-4706-44f7-9417-500418a18c73/pull/0.log" Nov 26 02:42:30 crc kubenswrapper[4766]: I1126 02:42:30.434431 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk_631bfaa9-4706-44f7-9417-500418a18c73/util/0.log" Nov 26 02:42:30 crc kubenswrapper[4766]: I1126 02:42:30.466120 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6qztkk_631bfaa9-4706-44f7-9417-500418a18c73/extract/0.log" Nov 26 02:42:30 crc kubenswrapper[4766]: I1126 02:42:30.568404 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-2xsk5_cad040b4-5d74-4cd0-b726-93d9c6cb73e2/marketplace-operator/0.log" Nov 26 02:42:30 crc kubenswrapper[4766]: I1126 02:42:30.666352 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vhw86_f18c8ba9-685a-4cd8-92e9-797074b81278/extract-utilities/0.log" Nov 26 02:42:30 crc kubenswrapper[4766]: I1126 02:42:30.791528 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vhw86_f18c8ba9-685a-4cd8-92e9-797074b81278/extract-utilities/0.log" Nov 26 02:42:30 crc kubenswrapper[4766]: I1126 02:42:30.819080 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vhw86_f18c8ba9-685a-4cd8-92e9-797074b81278/extract-content/0.log" Nov 26 02:42:30 crc kubenswrapper[4766]: I1126 02:42:30.842858 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vhw86_f18c8ba9-685a-4cd8-92e9-797074b81278/extract-content/0.log" Nov 26 02:42:30 crc kubenswrapper[4766]: I1126 02:42:30.969544 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vhw86_f18c8ba9-685a-4cd8-92e9-797074b81278/extract-utilities/0.log" Nov 26 02:42:31 crc kubenswrapper[4766]: I1126 02:42:31.046836 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vhw86_f18c8ba9-685a-4cd8-92e9-797074b81278/extract-content/0.log" Nov 26 02:42:31 crc kubenswrapper[4766]: I1126 02:42:31.153355 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pjtsc_34c7586e-58a4-4e97-b559-cb805d91e9d8/extract-utilities/0.log" Nov 26 02:42:31 crc kubenswrapper[4766]: I1126 02:42:31.258432 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vhw86_f18c8ba9-685a-4cd8-92e9-797074b81278/registry-server/0.log" Nov 26 02:42:31 crc kubenswrapper[4766]: I1126 02:42:31.378028 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pjtsc_34c7586e-58a4-4e97-b559-cb805d91e9d8/extract-content/0.log" Nov 26 02:42:31 crc kubenswrapper[4766]: I1126 02:42:31.388131 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pjtsc_34c7586e-58a4-4e97-b559-cb805d91e9d8/extract-utilities/0.log" Nov 26 02:42:31 crc kubenswrapper[4766]: I1126 02:42:31.390615 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pjtsc_34c7586e-58a4-4e97-b559-cb805d91e9d8/extract-content/0.log" Nov 26 02:42:31 crc kubenswrapper[4766]: I1126 02:42:31.559019 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pjtsc_34c7586e-58a4-4e97-b559-cb805d91e9d8/extract-content/0.log" Nov 26 02:42:31 crc kubenswrapper[4766]: I1126 02:42:31.563859 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pjtsc_34c7586e-58a4-4e97-b559-cb805d91e9d8/extract-utilities/0.log" Nov 26 02:42:32 crc kubenswrapper[4766]: I1126 02:42:32.565040 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pjtsc_34c7586e-58a4-4e97-b559-cb805d91e9d8/registry-server/0.log" Nov 26 02:42:46 crc kubenswrapper[4766]: I1126 02:42:46.095761 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-bt977_bf6b72c0-e2e6-4f3d-89f8-c18157e1ca70/prometheus-operator/0.log" Nov 26 02:42:46 crc kubenswrapper[4766]: I1126 02:42:46.249433 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5b9d946888-cjxjs_1125f245-de34-4508-a4b0-2a468e73df36/prometheus-operator-admission-webhook/0.log" Nov 26 02:42:46 crc kubenswrapper[4766]: I1126 02:42:46.313845 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5b9d946888-d8rcx_490105d8-e5f5-4fe9-89b9-d67ee8e07b87/prometheus-operator-admission-webhook/0.log" Nov 26 02:42:46 crc kubenswrapper[4766]: I1126 02:42:46.442395 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-b8qgs_a6dba801-0cee-4c4b-856c-db3be51c0ed9/operator/0.log" Nov 26 02:42:46 crc kubenswrapper[4766]: I1126 02:42:46.493490 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-ui-dashboards-7d5fb4cbfb-cmzfk_468c5f0c-30d9-4493-a87c-2f4482a3b4d3/observability-ui-dashboards/0.log" Nov 26 02:42:46 crc kubenswrapper[4766]: I1126 02:42:46.639168 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-twhqz_0be2b756-c941-49e3-b8f0-0977eaaf59d0/perses-operator/0.log" Nov 26 02:43:00 crc kubenswrapper[4766]: I1126 02:43:00.292986 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-7bbf4564c5-m77vx_b90f7164-a5e9-452f-8733-f549f15e17ec/kube-rbac-proxy/0.log" Nov 26 02:43:00 crc kubenswrapper[4766]: I1126 02:43:00.329999 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-7bbf4564c5-m77vx_b90f7164-a5e9-452f-8733-f549f15e17ec/manager/1.log" Nov 26 02:43:00 crc kubenswrapper[4766]: I1126 02:43:00.378057 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-7bbf4564c5-m77vx_b90f7164-a5e9-452f-8733-f549f15e17ec/manager/0.log" Nov 26 02:43:25 crc kubenswrapper[4766]: E1126 02:43:25.561905 4766 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.32:45686->38.102.83.32:36993: read tcp 38.102.83.32:45686->38.102.83.32:36993: read: connection reset by peer Nov 26 02:44:11 crc kubenswrapper[4766]: I1126 02:44:11.483462 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:44:11 crc kubenswrapper[4766]: I1126 02:44:11.484425 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:44:41 crc kubenswrapper[4766]: I1126 02:44:41.479802 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:44:41 crc kubenswrapper[4766]: I1126 02:44:41.480573 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:44:57 crc kubenswrapper[4766]: I1126 02:44:57.211116 4766 generic.go:334] "Generic (PLEG): container finished" podID="e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb" containerID="24e3e8da6468bd4fd80a5ae54b258044de46c878b291d6c281dd72a64b80b501" exitCode=0 Nov 26 02:44:57 crc kubenswrapper[4766]: I1126 02:44:57.211215 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffdxw/must-gather-x859m" event={"ID":"e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb","Type":"ContainerDied","Data":"24e3e8da6468bd4fd80a5ae54b258044de46c878b291d6c281dd72a64b80b501"} Nov 26 02:44:57 crc kubenswrapper[4766]: I1126 02:44:57.212311 4766 scope.go:117] "RemoveContainer" containerID="24e3e8da6468bd4fd80a5ae54b258044de46c878b291d6c281dd72a64b80b501" Nov 26 02:44:57 crc kubenswrapper[4766]: I1126 02:44:57.672344 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ffdxw_must-gather-x859m_e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb/gather/0.log" Nov 26 02:45:00 crc kubenswrapper[4766]: I1126 02:45:00.172839 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402085-2gztd"] Nov 26 02:45:00 crc kubenswrapper[4766]: E1126 02:45:00.173919 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8e40b61-1f16-47cf-9e5e-b260721f3ad9" containerName="extract-utilities" Nov 26 02:45:00 crc kubenswrapper[4766]: I1126 02:45:00.173944 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8e40b61-1f16-47cf-9e5e-b260721f3ad9" containerName="extract-utilities" Nov 26 02:45:00 crc kubenswrapper[4766]: E1126 02:45:00.173982 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8e40b61-1f16-47cf-9e5e-b260721f3ad9" containerName="extract-content" Nov 26 02:45:00 crc kubenswrapper[4766]: I1126 02:45:00.173995 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8e40b61-1f16-47cf-9e5e-b260721f3ad9" containerName="extract-content" Nov 26 02:45:00 crc kubenswrapper[4766]: E1126 02:45:00.174024 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8e40b61-1f16-47cf-9e5e-b260721f3ad9" containerName="registry-server" Nov 26 02:45:00 crc kubenswrapper[4766]: I1126 02:45:00.174036 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8e40b61-1f16-47cf-9e5e-b260721f3ad9" containerName="registry-server" Nov 26 02:45:00 crc kubenswrapper[4766]: I1126 02:45:00.174482 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8e40b61-1f16-47cf-9e5e-b260721f3ad9" containerName="registry-server" Nov 26 02:45:00 crc kubenswrapper[4766]: I1126 02:45:00.175797 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402085-2gztd" Nov 26 02:45:00 crc kubenswrapper[4766]: I1126 02:45:00.178259 4766 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 26 02:45:00 crc kubenswrapper[4766]: I1126 02:45:00.180547 4766 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 26 02:45:00 crc kubenswrapper[4766]: I1126 02:45:00.193240 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402085-2gztd"] Nov 26 02:45:00 crc kubenswrapper[4766]: I1126 02:45:00.323320 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/97c08050-6f3c-4b4f-ae56-e5ef59213de1-secret-volume\") pod \"collect-profiles-29402085-2gztd\" (UID: \"97c08050-6f3c-4b4f-ae56-e5ef59213de1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402085-2gztd" Nov 26 02:45:00 crc kubenswrapper[4766]: I1126 02:45:00.323412 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rw9w\" (UniqueName: \"kubernetes.io/projected/97c08050-6f3c-4b4f-ae56-e5ef59213de1-kube-api-access-5rw9w\") pod \"collect-profiles-29402085-2gztd\" (UID: \"97c08050-6f3c-4b4f-ae56-e5ef59213de1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402085-2gztd" Nov 26 02:45:00 crc kubenswrapper[4766]: I1126 02:45:00.323688 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/97c08050-6f3c-4b4f-ae56-e5ef59213de1-config-volume\") pod \"collect-profiles-29402085-2gztd\" (UID: \"97c08050-6f3c-4b4f-ae56-e5ef59213de1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402085-2gztd" Nov 26 02:45:00 crc kubenswrapper[4766]: I1126 02:45:00.426008 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/97c08050-6f3c-4b4f-ae56-e5ef59213de1-config-volume\") pod \"collect-profiles-29402085-2gztd\" (UID: \"97c08050-6f3c-4b4f-ae56-e5ef59213de1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402085-2gztd" Nov 26 02:45:00 crc kubenswrapper[4766]: I1126 02:45:00.426542 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/97c08050-6f3c-4b4f-ae56-e5ef59213de1-secret-volume\") pod \"collect-profiles-29402085-2gztd\" (UID: \"97c08050-6f3c-4b4f-ae56-e5ef59213de1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402085-2gztd" Nov 26 02:45:00 crc kubenswrapper[4766]: I1126 02:45:00.426671 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rw9w\" (UniqueName: \"kubernetes.io/projected/97c08050-6f3c-4b4f-ae56-e5ef59213de1-kube-api-access-5rw9w\") pod \"collect-profiles-29402085-2gztd\" (UID: \"97c08050-6f3c-4b4f-ae56-e5ef59213de1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402085-2gztd" Nov 26 02:45:00 crc kubenswrapper[4766]: I1126 02:45:00.426969 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/97c08050-6f3c-4b4f-ae56-e5ef59213de1-config-volume\") pod \"collect-profiles-29402085-2gztd\" (UID: \"97c08050-6f3c-4b4f-ae56-e5ef59213de1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402085-2gztd" Nov 26 02:45:00 crc kubenswrapper[4766]: I1126 02:45:00.436866 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/97c08050-6f3c-4b4f-ae56-e5ef59213de1-secret-volume\") pod \"collect-profiles-29402085-2gztd\" (UID: \"97c08050-6f3c-4b4f-ae56-e5ef59213de1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402085-2gztd" Nov 26 02:45:00 crc kubenswrapper[4766]: I1126 02:45:00.468438 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rw9w\" (UniqueName: \"kubernetes.io/projected/97c08050-6f3c-4b4f-ae56-e5ef59213de1-kube-api-access-5rw9w\") pod \"collect-profiles-29402085-2gztd\" (UID: \"97c08050-6f3c-4b4f-ae56-e5ef59213de1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402085-2gztd" Nov 26 02:45:00 crc kubenswrapper[4766]: I1126 02:45:00.520597 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402085-2gztd" Nov 26 02:45:01 crc kubenswrapper[4766]: I1126 02:45:01.117891 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402085-2gztd"] Nov 26 02:45:01 crc kubenswrapper[4766]: I1126 02:45:01.262307 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402085-2gztd" event={"ID":"97c08050-6f3c-4b4f-ae56-e5ef59213de1","Type":"ContainerStarted","Data":"e5e0326045ce3a8f3f4834fc6d69880e0663844a714c3729fb106c4aa34cbb04"} Nov 26 02:45:02 crc kubenswrapper[4766]: I1126 02:45:02.285717 4766 generic.go:334] "Generic (PLEG): container finished" podID="97c08050-6f3c-4b4f-ae56-e5ef59213de1" containerID="fc071e325dcbaf76fd3d744d4d9376ca20b61e1a13adbbf20c1bdc8bb1c6ffa3" exitCode=0 Nov 26 02:45:02 crc kubenswrapper[4766]: I1126 02:45:02.285834 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402085-2gztd" event={"ID":"97c08050-6f3c-4b4f-ae56-e5ef59213de1","Type":"ContainerDied","Data":"fc071e325dcbaf76fd3d744d4d9376ca20b61e1a13adbbf20c1bdc8bb1c6ffa3"} Nov 26 02:45:03 crc kubenswrapper[4766]: I1126 02:45:03.743293 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402085-2gztd" Nov 26 02:45:03 crc kubenswrapper[4766]: I1126 02:45:03.906729 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rw9w\" (UniqueName: \"kubernetes.io/projected/97c08050-6f3c-4b4f-ae56-e5ef59213de1-kube-api-access-5rw9w\") pod \"97c08050-6f3c-4b4f-ae56-e5ef59213de1\" (UID: \"97c08050-6f3c-4b4f-ae56-e5ef59213de1\") " Nov 26 02:45:03 crc kubenswrapper[4766]: I1126 02:45:03.907129 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/97c08050-6f3c-4b4f-ae56-e5ef59213de1-config-volume\") pod \"97c08050-6f3c-4b4f-ae56-e5ef59213de1\" (UID: \"97c08050-6f3c-4b4f-ae56-e5ef59213de1\") " Nov 26 02:45:03 crc kubenswrapper[4766]: I1126 02:45:03.907281 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/97c08050-6f3c-4b4f-ae56-e5ef59213de1-secret-volume\") pod \"97c08050-6f3c-4b4f-ae56-e5ef59213de1\" (UID: \"97c08050-6f3c-4b4f-ae56-e5ef59213de1\") " Nov 26 02:45:03 crc kubenswrapper[4766]: I1126 02:45:03.907743 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97c08050-6f3c-4b4f-ae56-e5ef59213de1-config-volume" (OuterVolumeSpecName: "config-volume") pod "97c08050-6f3c-4b4f-ae56-e5ef59213de1" (UID: "97c08050-6f3c-4b4f-ae56-e5ef59213de1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 02:45:03 crc kubenswrapper[4766]: I1126 02:45:03.908317 4766 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/97c08050-6f3c-4b4f-ae56-e5ef59213de1-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 02:45:03 crc kubenswrapper[4766]: I1126 02:45:03.944804 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97c08050-6f3c-4b4f-ae56-e5ef59213de1-kube-api-access-5rw9w" (OuterVolumeSpecName: "kube-api-access-5rw9w") pod "97c08050-6f3c-4b4f-ae56-e5ef59213de1" (UID: "97c08050-6f3c-4b4f-ae56-e5ef59213de1"). InnerVolumeSpecName "kube-api-access-5rw9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:45:03 crc kubenswrapper[4766]: I1126 02:45:03.945542 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97c08050-6f3c-4b4f-ae56-e5ef59213de1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "97c08050-6f3c-4b4f-ae56-e5ef59213de1" (UID: "97c08050-6f3c-4b4f-ae56-e5ef59213de1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 02:45:04 crc kubenswrapper[4766]: I1126 02:45:04.010898 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rw9w\" (UniqueName: \"kubernetes.io/projected/97c08050-6f3c-4b4f-ae56-e5ef59213de1-kube-api-access-5rw9w\") on node \"crc\" DevicePath \"\"" Nov 26 02:45:04 crc kubenswrapper[4766]: I1126 02:45:04.010934 4766 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/97c08050-6f3c-4b4f-ae56-e5ef59213de1-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 26 02:45:04 crc kubenswrapper[4766]: I1126 02:45:04.316051 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402085-2gztd" event={"ID":"97c08050-6f3c-4b4f-ae56-e5ef59213de1","Type":"ContainerDied","Data":"e5e0326045ce3a8f3f4834fc6d69880e0663844a714c3729fb106c4aa34cbb04"} Nov 26 02:45:04 crc kubenswrapper[4766]: I1126 02:45:04.316422 4766 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5e0326045ce3a8f3f4834fc6d69880e0663844a714c3729fb106c4aa34cbb04" Nov 26 02:45:04 crc kubenswrapper[4766]: I1126 02:45:04.316786 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402085-2gztd" Nov 26 02:45:04 crc kubenswrapper[4766]: I1126 02:45:04.870587 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt"] Nov 26 02:45:04 crc kubenswrapper[4766]: I1126 02:45:04.887471 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402040-hg5nt"] Nov 26 02:45:05 crc kubenswrapper[4766]: I1126 02:45:05.843580 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79ce3c1d-945d-44d0-a263-de55f580b561" path="/var/lib/kubelet/pods/79ce3c1d-945d-44d0-a263-de55f580b561/volumes" Nov 26 02:45:07 crc kubenswrapper[4766]: I1126 02:45:07.600395 4766 scope.go:117] "RemoveContainer" containerID="c3fdaa8a53fca84445f7eb62a215407d07a46d3ead6bb9e962216858cfa6e30f" Nov 26 02:45:07 crc kubenswrapper[4766]: I1126 02:45:07.629229 4766 scope.go:117] "RemoveContainer" containerID="6ae52fd9463077d7b32b88151ac7223a081d7a6fae971789b68c1f4b947a9b13" Nov 26 02:45:09 crc kubenswrapper[4766]: I1126 02:45:09.526592 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ffdxw/must-gather-x859m"] Nov 26 02:45:09 crc kubenswrapper[4766]: I1126 02:45:09.527292 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-ffdxw/must-gather-x859m" podUID="e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb" containerName="copy" containerID="cri-o://dfe6aea51adb38a75d913e3b421827a39f2f9cd780d8595389a17b3f322ad9eb" gracePeriod=2 Nov 26 02:45:09 crc kubenswrapper[4766]: I1126 02:45:09.537829 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ffdxw/must-gather-x859m"] Nov 26 02:45:10 crc kubenswrapper[4766]: I1126 02:45:10.027184 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ffdxw_must-gather-x859m_e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb/copy/0.log" Nov 26 02:45:10 crc kubenswrapper[4766]: I1126 02:45:10.027884 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffdxw/must-gather-x859m" Nov 26 02:45:10 crc kubenswrapper[4766]: I1126 02:45:10.155365 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57ckq\" (UniqueName: \"kubernetes.io/projected/e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb-kube-api-access-57ckq\") pod \"e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb\" (UID: \"e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb\") " Nov 26 02:45:10 crc kubenswrapper[4766]: I1126 02:45:10.155796 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb-must-gather-output\") pod \"e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb\" (UID: \"e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb\") " Nov 26 02:45:10 crc kubenswrapper[4766]: I1126 02:45:10.176938 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb-kube-api-access-57ckq" (OuterVolumeSpecName: "kube-api-access-57ckq") pod "e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb" (UID: "e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb"). InnerVolumeSpecName "kube-api-access-57ckq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:45:10 crc kubenswrapper[4766]: I1126 02:45:10.259033 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57ckq\" (UniqueName: \"kubernetes.io/projected/e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb-kube-api-access-57ckq\") on node \"crc\" DevicePath \"\"" Nov 26 02:45:10 crc kubenswrapper[4766]: I1126 02:45:10.360562 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb" (UID: "e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:45:10 crc kubenswrapper[4766]: I1126 02:45:10.398850 4766 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ffdxw_must-gather-x859m_e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb/copy/0.log" Nov 26 02:45:10 crc kubenswrapper[4766]: I1126 02:45:10.399459 4766 generic.go:334] "Generic (PLEG): container finished" podID="e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb" containerID="dfe6aea51adb38a75d913e3b421827a39f2f9cd780d8595389a17b3f322ad9eb" exitCode=143 Nov 26 02:45:10 crc kubenswrapper[4766]: I1126 02:45:10.399518 4766 scope.go:117] "RemoveContainer" containerID="dfe6aea51adb38a75d913e3b421827a39f2f9cd780d8595389a17b3f322ad9eb" Nov 26 02:45:10 crc kubenswrapper[4766]: I1126 02:45:10.399519 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffdxw/must-gather-x859m" Nov 26 02:45:10 crc kubenswrapper[4766]: I1126 02:45:10.425052 4766 scope.go:117] "RemoveContainer" containerID="24e3e8da6468bd4fd80a5ae54b258044de46c878b291d6c281dd72a64b80b501" Nov 26 02:45:10 crc kubenswrapper[4766]: I1126 02:45:10.463464 4766 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 26 02:45:10 crc kubenswrapper[4766]: I1126 02:45:10.477441 4766 scope.go:117] "RemoveContainer" containerID="dfe6aea51adb38a75d913e3b421827a39f2f9cd780d8595389a17b3f322ad9eb" Nov 26 02:45:10 crc kubenswrapper[4766]: E1126 02:45:10.477858 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfe6aea51adb38a75d913e3b421827a39f2f9cd780d8595389a17b3f322ad9eb\": container with ID starting with dfe6aea51adb38a75d913e3b421827a39f2f9cd780d8595389a17b3f322ad9eb not found: ID does not exist" containerID="dfe6aea51adb38a75d913e3b421827a39f2f9cd780d8595389a17b3f322ad9eb" Nov 26 02:45:10 crc kubenswrapper[4766]: I1126 02:45:10.477883 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfe6aea51adb38a75d913e3b421827a39f2f9cd780d8595389a17b3f322ad9eb"} err="failed to get container status \"dfe6aea51adb38a75d913e3b421827a39f2f9cd780d8595389a17b3f322ad9eb\": rpc error: code = NotFound desc = could not find container \"dfe6aea51adb38a75d913e3b421827a39f2f9cd780d8595389a17b3f322ad9eb\": container with ID starting with dfe6aea51adb38a75d913e3b421827a39f2f9cd780d8595389a17b3f322ad9eb not found: ID does not exist" Nov 26 02:45:10 crc kubenswrapper[4766]: I1126 02:45:10.477904 4766 scope.go:117] "RemoveContainer" containerID="24e3e8da6468bd4fd80a5ae54b258044de46c878b291d6c281dd72a64b80b501" Nov 26 02:45:10 crc kubenswrapper[4766]: E1126 02:45:10.478115 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24e3e8da6468bd4fd80a5ae54b258044de46c878b291d6c281dd72a64b80b501\": container with ID starting with 24e3e8da6468bd4fd80a5ae54b258044de46c878b291d6c281dd72a64b80b501 not found: ID does not exist" containerID="24e3e8da6468bd4fd80a5ae54b258044de46c878b291d6c281dd72a64b80b501" Nov 26 02:45:10 crc kubenswrapper[4766]: I1126 02:45:10.478134 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24e3e8da6468bd4fd80a5ae54b258044de46c878b291d6c281dd72a64b80b501"} err="failed to get container status \"24e3e8da6468bd4fd80a5ae54b258044de46c878b291d6c281dd72a64b80b501\": rpc error: code = NotFound desc = could not find container \"24e3e8da6468bd4fd80a5ae54b258044de46c878b291d6c281dd72a64b80b501\": container with ID starting with 24e3e8da6468bd4fd80a5ae54b258044de46c878b291d6c281dd72a64b80b501 not found: ID does not exist" Nov 26 02:45:11 crc kubenswrapper[4766]: I1126 02:45:11.481719 4766 patch_prober.go:28] interesting pod/machine-config-daemon-wf9c2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 02:45:11 crc kubenswrapper[4766]: I1126 02:45:11.481822 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 02:45:11 crc kubenswrapper[4766]: I1126 02:45:11.481907 4766 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" Nov 26 02:45:11 crc kubenswrapper[4766]: I1126 02:45:11.483560 4766 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3"} pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 02:45:11 crc kubenswrapper[4766]: I1126 02:45:11.483871 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" containerName="machine-config-daemon" containerID="cri-o://fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3" gracePeriod=600 Nov 26 02:45:11 crc kubenswrapper[4766]: E1126 02:45:11.624424 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:45:11 crc kubenswrapper[4766]: I1126 02:45:11.846935 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb" path="/var/lib/kubelet/pods/e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb/volumes" Nov 26 02:45:12 crc kubenswrapper[4766]: I1126 02:45:12.454616 4766 generic.go:334] "Generic (PLEG): container finished" podID="63471884-61f7-41d5-8967-e3f81eba90d9" containerID="fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3" exitCode=0 Nov 26 02:45:12 crc kubenswrapper[4766]: I1126 02:45:12.454667 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" event={"ID":"63471884-61f7-41d5-8967-e3f81eba90d9","Type":"ContainerDied","Data":"fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3"} Nov 26 02:45:12 crc kubenswrapper[4766]: I1126 02:45:12.454704 4766 scope.go:117] "RemoveContainer" containerID="93d6a4a6ec497ff7bc4cee7766539719d24858cdc0bc941ecde4de3cf6ff02cd" Nov 26 02:45:12 crc kubenswrapper[4766]: I1126 02:45:12.455386 4766 scope.go:117] "RemoveContainer" containerID="fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3" Nov 26 02:45:12 crc kubenswrapper[4766]: E1126 02:45:12.455854 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:45:24 crc kubenswrapper[4766]: I1126 02:45:24.829052 4766 scope.go:117] "RemoveContainer" containerID="fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3" Nov 26 02:45:24 crc kubenswrapper[4766]: E1126 02:45:24.829799 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:45:35 crc kubenswrapper[4766]: I1126 02:45:35.827436 4766 scope.go:117] "RemoveContainer" containerID="fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3" Nov 26 02:45:35 crc kubenswrapper[4766]: E1126 02:45:35.830586 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:45:36 crc kubenswrapper[4766]: I1126 02:45:36.120577 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gn2hf"] Nov 26 02:45:36 crc kubenswrapper[4766]: E1126 02:45:36.121835 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb" containerName="gather" Nov 26 02:45:36 crc kubenswrapper[4766]: I1126 02:45:36.122017 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb" containerName="gather" Nov 26 02:45:36 crc kubenswrapper[4766]: E1126 02:45:36.122229 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97c08050-6f3c-4b4f-ae56-e5ef59213de1" containerName="collect-profiles" Nov 26 02:45:36 crc kubenswrapper[4766]: I1126 02:45:36.122400 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="97c08050-6f3c-4b4f-ae56-e5ef59213de1" containerName="collect-profiles" Nov 26 02:45:36 crc kubenswrapper[4766]: E1126 02:45:36.122585 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb" containerName="copy" Nov 26 02:45:36 crc kubenswrapper[4766]: I1126 02:45:36.122765 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb" containerName="copy" Nov 26 02:45:36 crc kubenswrapper[4766]: I1126 02:45:36.123307 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb" containerName="gather" Nov 26 02:45:36 crc kubenswrapper[4766]: I1126 02:45:36.123508 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="e62c4091-4e4c-4bf2-bf7d-bc5fa8cdc6eb" containerName="copy" Nov 26 02:45:36 crc kubenswrapper[4766]: I1126 02:45:36.123752 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="97c08050-6f3c-4b4f-ae56-e5ef59213de1" containerName="collect-profiles" Nov 26 02:45:36 crc kubenswrapper[4766]: I1126 02:45:36.127012 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gn2hf" Nov 26 02:45:36 crc kubenswrapper[4766]: I1126 02:45:36.138202 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gn2hf"] Nov 26 02:45:36 crc kubenswrapper[4766]: I1126 02:45:36.219181 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2029d62b-210b-45e0-b81b-e92d7de06b25-utilities\") pod \"certified-operators-gn2hf\" (UID: \"2029d62b-210b-45e0-b81b-e92d7de06b25\") " pod="openshift-marketplace/certified-operators-gn2hf" Nov 26 02:45:36 crc kubenswrapper[4766]: I1126 02:45:36.219269 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2029d62b-210b-45e0-b81b-e92d7de06b25-catalog-content\") pod \"certified-operators-gn2hf\" (UID: \"2029d62b-210b-45e0-b81b-e92d7de06b25\") " pod="openshift-marketplace/certified-operators-gn2hf" Nov 26 02:45:36 crc kubenswrapper[4766]: I1126 02:45:36.219333 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8znwh\" (UniqueName: \"kubernetes.io/projected/2029d62b-210b-45e0-b81b-e92d7de06b25-kube-api-access-8znwh\") pod \"certified-operators-gn2hf\" (UID: \"2029d62b-210b-45e0-b81b-e92d7de06b25\") " pod="openshift-marketplace/certified-operators-gn2hf" Nov 26 02:45:36 crc kubenswrapper[4766]: I1126 02:45:36.321451 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2029d62b-210b-45e0-b81b-e92d7de06b25-utilities\") pod \"certified-operators-gn2hf\" (UID: \"2029d62b-210b-45e0-b81b-e92d7de06b25\") " pod="openshift-marketplace/certified-operators-gn2hf" Nov 26 02:45:36 crc kubenswrapper[4766]: I1126 02:45:36.321563 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2029d62b-210b-45e0-b81b-e92d7de06b25-catalog-content\") pod \"certified-operators-gn2hf\" (UID: \"2029d62b-210b-45e0-b81b-e92d7de06b25\") " pod="openshift-marketplace/certified-operators-gn2hf" Nov 26 02:45:36 crc kubenswrapper[4766]: I1126 02:45:36.321626 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8znwh\" (UniqueName: \"kubernetes.io/projected/2029d62b-210b-45e0-b81b-e92d7de06b25-kube-api-access-8znwh\") pod \"certified-operators-gn2hf\" (UID: \"2029d62b-210b-45e0-b81b-e92d7de06b25\") " pod="openshift-marketplace/certified-operators-gn2hf" Nov 26 02:45:36 crc kubenswrapper[4766]: I1126 02:45:36.322623 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2029d62b-210b-45e0-b81b-e92d7de06b25-utilities\") pod \"certified-operators-gn2hf\" (UID: \"2029d62b-210b-45e0-b81b-e92d7de06b25\") " pod="openshift-marketplace/certified-operators-gn2hf" Nov 26 02:45:36 crc kubenswrapper[4766]: I1126 02:45:36.322687 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2029d62b-210b-45e0-b81b-e92d7de06b25-catalog-content\") pod \"certified-operators-gn2hf\" (UID: \"2029d62b-210b-45e0-b81b-e92d7de06b25\") " pod="openshift-marketplace/certified-operators-gn2hf" Nov 26 02:45:36 crc kubenswrapper[4766]: I1126 02:45:36.345368 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8znwh\" (UniqueName: \"kubernetes.io/projected/2029d62b-210b-45e0-b81b-e92d7de06b25-kube-api-access-8znwh\") pod \"certified-operators-gn2hf\" (UID: \"2029d62b-210b-45e0-b81b-e92d7de06b25\") " pod="openshift-marketplace/certified-operators-gn2hf" Nov 26 02:45:36 crc kubenswrapper[4766]: I1126 02:45:36.478311 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gn2hf" Nov 26 02:45:36 crc kubenswrapper[4766]: I1126 02:45:36.989196 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gn2hf"] Nov 26 02:45:37 crc kubenswrapper[4766]: I1126 02:45:37.793320 4766 generic.go:334] "Generic (PLEG): container finished" podID="2029d62b-210b-45e0-b81b-e92d7de06b25" containerID="2d0a52fe649c78104188aee7084b95bec16cdfd3e95b46ca4ca24549c6f5d816" exitCode=0 Nov 26 02:45:37 crc kubenswrapper[4766]: I1126 02:45:37.793384 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gn2hf" event={"ID":"2029d62b-210b-45e0-b81b-e92d7de06b25","Type":"ContainerDied","Data":"2d0a52fe649c78104188aee7084b95bec16cdfd3e95b46ca4ca24549c6f5d816"} Nov 26 02:45:37 crc kubenswrapper[4766]: I1126 02:45:37.793752 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gn2hf" event={"ID":"2029d62b-210b-45e0-b81b-e92d7de06b25","Type":"ContainerStarted","Data":"047ed224c90b6899679f34f5e9c80a6fbadfa7ff24a5e1b99789b801c4c96f73"} Nov 26 02:45:37 crc kubenswrapper[4766]: I1126 02:45:37.795600 4766 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 02:45:39 crc kubenswrapper[4766]: I1126 02:45:39.818638 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gn2hf" event={"ID":"2029d62b-210b-45e0-b81b-e92d7de06b25","Type":"ContainerStarted","Data":"dcdf480b90136f6f79fd0f179fe8c7c691b791df907743a4be563abf1ae93d63"} Nov 26 02:45:40 crc kubenswrapper[4766]: I1126 02:45:40.845547 4766 generic.go:334] "Generic (PLEG): container finished" podID="2029d62b-210b-45e0-b81b-e92d7de06b25" containerID="dcdf480b90136f6f79fd0f179fe8c7c691b791df907743a4be563abf1ae93d63" exitCode=0 Nov 26 02:45:40 crc kubenswrapper[4766]: I1126 02:45:40.845599 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gn2hf" event={"ID":"2029d62b-210b-45e0-b81b-e92d7de06b25","Type":"ContainerDied","Data":"dcdf480b90136f6f79fd0f179fe8c7c691b791df907743a4be563abf1ae93d63"} Nov 26 02:45:41 crc kubenswrapper[4766]: I1126 02:45:41.859062 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gn2hf" event={"ID":"2029d62b-210b-45e0-b81b-e92d7de06b25","Type":"ContainerStarted","Data":"c31ddef8a5d2ee2a3d0e1c80e5da0a3ab2db9bc84b141073f207b7fdbd5ea854"} Nov 26 02:45:41 crc kubenswrapper[4766]: I1126 02:45:41.884568 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gn2hf" podStartSLOduration=2.3253935820000002 podStartE2EDuration="5.884548337s" podCreationTimestamp="2025-11-26 02:45:36 +0000 UTC" firstStartedPulling="2025-11-26 02:45:37.795412616 +0000 UTC m=+8518.644183046" lastFinishedPulling="2025-11-26 02:45:41.354567361 +0000 UTC m=+8522.203337801" observedRunningTime="2025-11-26 02:45:41.87894963 +0000 UTC m=+8522.727720070" watchObservedRunningTime="2025-11-26 02:45:41.884548337 +0000 UTC m=+8522.733318777" Nov 26 02:45:46 crc kubenswrapper[4766]: I1126 02:45:46.479401 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gn2hf" Nov 26 02:45:46 crc kubenswrapper[4766]: I1126 02:45:46.481513 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gn2hf" Nov 26 02:45:46 crc kubenswrapper[4766]: I1126 02:45:46.576170 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gn2hf" Nov 26 02:45:47 crc kubenswrapper[4766]: I1126 02:45:47.007502 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gn2hf" Nov 26 02:45:47 crc kubenswrapper[4766]: I1126 02:45:47.068773 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gn2hf"] Nov 26 02:45:48 crc kubenswrapper[4766]: I1126 02:45:48.948886 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gn2hf" podUID="2029d62b-210b-45e0-b81b-e92d7de06b25" containerName="registry-server" containerID="cri-o://c31ddef8a5d2ee2a3d0e1c80e5da0a3ab2db9bc84b141073f207b7fdbd5ea854" gracePeriod=2 Nov 26 02:45:49 crc kubenswrapper[4766]: I1126 02:45:49.636523 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gn2hf" Nov 26 02:45:49 crc kubenswrapper[4766]: I1126 02:45:49.774911 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2029d62b-210b-45e0-b81b-e92d7de06b25-catalog-content\") pod \"2029d62b-210b-45e0-b81b-e92d7de06b25\" (UID: \"2029d62b-210b-45e0-b81b-e92d7de06b25\") " Nov 26 02:45:49 crc kubenswrapper[4766]: I1126 02:45:49.775123 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2029d62b-210b-45e0-b81b-e92d7de06b25-utilities\") pod \"2029d62b-210b-45e0-b81b-e92d7de06b25\" (UID: \"2029d62b-210b-45e0-b81b-e92d7de06b25\") " Nov 26 02:45:49 crc kubenswrapper[4766]: I1126 02:45:49.775197 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8znwh\" (UniqueName: \"kubernetes.io/projected/2029d62b-210b-45e0-b81b-e92d7de06b25-kube-api-access-8znwh\") pod \"2029d62b-210b-45e0-b81b-e92d7de06b25\" (UID: \"2029d62b-210b-45e0-b81b-e92d7de06b25\") " Nov 26 02:45:49 crc kubenswrapper[4766]: I1126 02:45:49.777399 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2029d62b-210b-45e0-b81b-e92d7de06b25-utilities" (OuterVolumeSpecName: "utilities") pod "2029d62b-210b-45e0-b81b-e92d7de06b25" (UID: "2029d62b-210b-45e0-b81b-e92d7de06b25"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:45:49 crc kubenswrapper[4766]: I1126 02:45:49.783540 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2029d62b-210b-45e0-b81b-e92d7de06b25-kube-api-access-8znwh" (OuterVolumeSpecName: "kube-api-access-8znwh") pod "2029d62b-210b-45e0-b81b-e92d7de06b25" (UID: "2029d62b-210b-45e0-b81b-e92d7de06b25"). InnerVolumeSpecName "kube-api-access-8znwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:45:49 crc kubenswrapper[4766]: I1126 02:45:49.835190 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2029d62b-210b-45e0-b81b-e92d7de06b25-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2029d62b-210b-45e0-b81b-e92d7de06b25" (UID: "2029d62b-210b-45e0-b81b-e92d7de06b25"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:45:49 crc kubenswrapper[4766]: I1126 02:45:49.877795 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2029d62b-210b-45e0-b81b-e92d7de06b25-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 02:45:49 crc kubenswrapper[4766]: I1126 02:45:49.877834 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8znwh\" (UniqueName: \"kubernetes.io/projected/2029d62b-210b-45e0-b81b-e92d7de06b25-kube-api-access-8znwh\") on node \"crc\" DevicePath \"\"" Nov 26 02:45:49 crc kubenswrapper[4766]: I1126 02:45:49.877849 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2029d62b-210b-45e0-b81b-e92d7de06b25-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 02:45:49 crc kubenswrapper[4766]: I1126 02:45:49.963311 4766 generic.go:334] "Generic (PLEG): container finished" podID="2029d62b-210b-45e0-b81b-e92d7de06b25" containerID="c31ddef8a5d2ee2a3d0e1c80e5da0a3ab2db9bc84b141073f207b7fdbd5ea854" exitCode=0 Nov 26 02:45:49 crc kubenswrapper[4766]: I1126 02:45:49.963350 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gn2hf" event={"ID":"2029d62b-210b-45e0-b81b-e92d7de06b25","Type":"ContainerDied","Data":"c31ddef8a5d2ee2a3d0e1c80e5da0a3ab2db9bc84b141073f207b7fdbd5ea854"} Nov 26 02:45:49 crc kubenswrapper[4766]: I1126 02:45:49.963381 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gn2hf" event={"ID":"2029d62b-210b-45e0-b81b-e92d7de06b25","Type":"ContainerDied","Data":"047ed224c90b6899679f34f5e9c80a6fbadfa7ff24a5e1b99789b801c4c96f73"} Nov 26 02:45:49 crc kubenswrapper[4766]: I1126 02:45:49.963399 4766 scope.go:117] "RemoveContainer" containerID="c31ddef8a5d2ee2a3d0e1c80e5da0a3ab2db9bc84b141073f207b7fdbd5ea854" Nov 26 02:45:49 crc kubenswrapper[4766]: I1126 02:45:49.964058 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gn2hf" Nov 26 02:45:50 crc kubenswrapper[4766]: I1126 02:45:50.001214 4766 scope.go:117] "RemoveContainer" containerID="dcdf480b90136f6f79fd0f179fe8c7c691b791df907743a4be563abf1ae93d63" Nov 26 02:45:50 crc kubenswrapper[4766]: I1126 02:45:50.010709 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gn2hf"] Nov 26 02:45:50 crc kubenswrapper[4766]: I1126 02:45:50.025612 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gn2hf"] Nov 26 02:45:50 crc kubenswrapper[4766]: I1126 02:45:50.033947 4766 scope.go:117] "RemoveContainer" containerID="2d0a52fe649c78104188aee7084b95bec16cdfd3e95b46ca4ca24549c6f5d816" Nov 26 02:45:50 crc kubenswrapper[4766]: I1126 02:45:50.096610 4766 scope.go:117] "RemoveContainer" containerID="c31ddef8a5d2ee2a3d0e1c80e5da0a3ab2db9bc84b141073f207b7fdbd5ea854" Nov 26 02:45:50 crc kubenswrapper[4766]: E1126 02:45:50.097300 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c31ddef8a5d2ee2a3d0e1c80e5da0a3ab2db9bc84b141073f207b7fdbd5ea854\": container with ID starting with c31ddef8a5d2ee2a3d0e1c80e5da0a3ab2db9bc84b141073f207b7fdbd5ea854 not found: ID does not exist" containerID="c31ddef8a5d2ee2a3d0e1c80e5da0a3ab2db9bc84b141073f207b7fdbd5ea854" Nov 26 02:45:50 crc kubenswrapper[4766]: I1126 02:45:50.097356 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c31ddef8a5d2ee2a3d0e1c80e5da0a3ab2db9bc84b141073f207b7fdbd5ea854"} err="failed to get container status \"c31ddef8a5d2ee2a3d0e1c80e5da0a3ab2db9bc84b141073f207b7fdbd5ea854\": rpc error: code = NotFound desc = could not find container \"c31ddef8a5d2ee2a3d0e1c80e5da0a3ab2db9bc84b141073f207b7fdbd5ea854\": container with ID starting with c31ddef8a5d2ee2a3d0e1c80e5da0a3ab2db9bc84b141073f207b7fdbd5ea854 not found: ID does not exist" Nov 26 02:45:50 crc kubenswrapper[4766]: I1126 02:45:50.097390 4766 scope.go:117] "RemoveContainer" containerID="dcdf480b90136f6f79fd0f179fe8c7c691b791df907743a4be563abf1ae93d63" Nov 26 02:45:50 crc kubenswrapper[4766]: E1126 02:45:50.097833 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcdf480b90136f6f79fd0f179fe8c7c691b791df907743a4be563abf1ae93d63\": container with ID starting with dcdf480b90136f6f79fd0f179fe8c7c691b791df907743a4be563abf1ae93d63 not found: ID does not exist" containerID="dcdf480b90136f6f79fd0f179fe8c7c691b791df907743a4be563abf1ae93d63" Nov 26 02:45:50 crc kubenswrapper[4766]: I1126 02:45:50.097866 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcdf480b90136f6f79fd0f179fe8c7c691b791df907743a4be563abf1ae93d63"} err="failed to get container status \"dcdf480b90136f6f79fd0f179fe8c7c691b791df907743a4be563abf1ae93d63\": rpc error: code = NotFound desc = could not find container \"dcdf480b90136f6f79fd0f179fe8c7c691b791df907743a4be563abf1ae93d63\": container with ID starting with dcdf480b90136f6f79fd0f179fe8c7c691b791df907743a4be563abf1ae93d63 not found: ID does not exist" Nov 26 02:45:50 crc kubenswrapper[4766]: I1126 02:45:50.097884 4766 scope.go:117] "RemoveContainer" containerID="2d0a52fe649c78104188aee7084b95bec16cdfd3e95b46ca4ca24549c6f5d816" Nov 26 02:45:50 crc kubenswrapper[4766]: E1126 02:45:50.098266 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d0a52fe649c78104188aee7084b95bec16cdfd3e95b46ca4ca24549c6f5d816\": container with ID starting with 2d0a52fe649c78104188aee7084b95bec16cdfd3e95b46ca4ca24549c6f5d816 not found: ID does not exist" containerID="2d0a52fe649c78104188aee7084b95bec16cdfd3e95b46ca4ca24549c6f5d816" Nov 26 02:45:50 crc kubenswrapper[4766]: I1126 02:45:50.098304 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d0a52fe649c78104188aee7084b95bec16cdfd3e95b46ca4ca24549c6f5d816"} err="failed to get container status \"2d0a52fe649c78104188aee7084b95bec16cdfd3e95b46ca4ca24549c6f5d816\": rpc error: code = NotFound desc = could not find container \"2d0a52fe649c78104188aee7084b95bec16cdfd3e95b46ca4ca24549c6f5d816\": container with ID starting with 2d0a52fe649c78104188aee7084b95bec16cdfd3e95b46ca4ca24549c6f5d816 not found: ID does not exist" Nov 26 02:45:50 crc kubenswrapper[4766]: I1126 02:45:50.827772 4766 scope.go:117] "RemoveContainer" containerID="fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3" Nov 26 02:45:50 crc kubenswrapper[4766]: E1126 02:45:50.828388 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:45:51 crc kubenswrapper[4766]: I1126 02:45:51.849445 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2029d62b-210b-45e0-b81b-e92d7de06b25" path="/var/lib/kubelet/pods/2029d62b-210b-45e0-b81b-e92d7de06b25/volumes" Nov 26 02:46:03 crc kubenswrapper[4766]: I1126 02:46:03.831636 4766 scope.go:117] "RemoveContainer" containerID="fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3" Nov 26 02:46:03 crc kubenswrapper[4766]: E1126 02:46:03.856152 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:46:16 crc kubenswrapper[4766]: I1126 02:46:16.828487 4766 scope.go:117] "RemoveContainer" containerID="fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3" Nov 26 02:46:16 crc kubenswrapper[4766]: E1126 02:46:16.830481 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:46:30 crc kubenswrapper[4766]: I1126 02:46:30.827811 4766 scope.go:117] "RemoveContainer" containerID="fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3" Nov 26 02:46:30 crc kubenswrapper[4766]: E1126 02:46:30.829213 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:46:45 crc kubenswrapper[4766]: I1126 02:46:45.827813 4766 scope.go:117] "RemoveContainer" containerID="fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3" Nov 26 02:46:45 crc kubenswrapper[4766]: E1126 02:46:45.828991 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:46:57 crc kubenswrapper[4766]: I1126 02:46:57.827563 4766 scope.go:117] "RemoveContainer" containerID="fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3" Nov 26 02:46:57 crc kubenswrapper[4766]: E1126 02:46:57.828931 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:47:08 crc kubenswrapper[4766]: I1126 02:47:08.296005 4766 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-74d678b987-q89lp" podUID="3d4e6997-4bbc-4cb2-a0f2-abf3b940090c" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Nov 26 02:47:08 crc kubenswrapper[4766]: I1126 02:47:08.829142 4766 scope.go:117] "RemoveContainer" containerID="fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3" Nov 26 02:47:08 crc kubenswrapper[4766]: E1126 02:47:08.829934 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:47:17 crc kubenswrapper[4766]: E1126 02:47:17.163689 4766 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.192s" Nov 26 02:47:21 crc kubenswrapper[4766]: I1126 02:47:21.828310 4766 scope.go:117] "RemoveContainer" containerID="fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3" Nov 26 02:47:21 crc kubenswrapper[4766]: E1126 02:47:21.829600 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:47:36 crc kubenswrapper[4766]: I1126 02:47:36.829136 4766 scope.go:117] "RemoveContainer" containerID="fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3" Nov 26 02:47:36 crc kubenswrapper[4766]: E1126 02:47:36.829901 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:47:50 crc kubenswrapper[4766]: I1126 02:47:50.909158 4766 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k2vfm"] Nov 26 02:47:50 crc kubenswrapper[4766]: E1126 02:47:50.910121 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2029d62b-210b-45e0-b81b-e92d7de06b25" containerName="extract-utilities" Nov 26 02:47:50 crc kubenswrapper[4766]: I1126 02:47:50.910134 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="2029d62b-210b-45e0-b81b-e92d7de06b25" containerName="extract-utilities" Nov 26 02:47:50 crc kubenswrapper[4766]: E1126 02:47:50.910149 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2029d62b-210b-45e0-b81b-e92d7de06b25" containerName="extract-content" Nov 26 02:47:50 crc kubenswrapper[4766]: I1126 02:47:50.910155 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="2029d62b-210b-45e0-b81b-e92d7de06b25" containerName="extract-content" Nov 26 02:47:50 crc kubenswrapper[4766]: E1126 02:47:50.910183 4766 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2029d62b-210b-45e0-b81b-e92d7de06b25" containerName="registry-server" Nov 26 02:47:50 crc kubenswrapper[4766]: I1126 02:47:50.910189 4766 state_mem.go:107] "Deleted CPUSet assignment" podUID="2029d62b-210b-45e0-b81b-e92d7de06b25" containerName="registry-server" Nov 26 02:47:50 crc kubenswrapper[4766]: I1126 02:47:50.910420 4766 memory_manager.go:354] "RemoveStaleState removing state" podUID="2029d62b-210b-45e0-b81b-e92d7de06b25" containerName="registry-server" Nov 26 02:47:50 crc kubenswrapper[4766]: I1126 02:47:50.912143 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2vfm" Nov 26 02:47:50 crc kubenswrapper[4766]: I1126 02:47:50.938358 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k2vfm"] Nov 26 02:47:51 crc kubenswrapper[4766]: I1126 02:47:51.062178 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c90a91f-c037-43fb-844e-bc4b95b37dee-catalog-content\") pod \"community-operators-k2vfm\" (UID: \"0c90a91f-c037-43fb-844e-bc4b95b37dee\") " pod="openshift-marketplace/community-operators-k2vfm" Nov 26 02:47:51 crc kubenswrapper[4766]: I1126 02:47:51.062248 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c90a91f-c037-43fb-844e-bc4b95b37dee-utilities\") pod \"community-operators-k2vfm\" (UID: \"0c90a91f-c037-43fb-844e-bc4b95b37dee\") " pod="openshift-marketplace/community-operators-k2vfm" Nov 26 02:47:51 crc kubenswrapper[4766]: I1126 02:47:51.062274 4766 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnr2r\" (UniqueName: \"kubernetes.io/projected/0c90a91f-c037-43fb-844e-bc4b95b37dee-kube-api-access-mnr2r\") pod \"community-operators-k2vfm\" (UID: \"0c90a91f-c037-43fb-844e-bc4b95b37dee\") " pod="openshift-marketplace/community-operators-k2vfm" Nov 26 02:47:51 crc kubenswrapper[4766]: I1126 02:47:51.163827 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c90a91f-c037-43fb-844e-bc4b95b37dee-catalog-content\") pod \"community-operators-k2vfm\" (UID: \"0c90a91f-c037-43fb-844e-bc4b95b37dee\") " pod="openshift-marketplace/community-operators-k2vfm" Nov 26 02:47:51 crc kubenswrapper[4766]: I1126 02:47:51.163875 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c90a91f-c037-43fb-844e-bc4b95b37dee-utilities\") pod \"community-operators-k2vfm\" (UID: \"0c90a91f-c037-43fb-844e-bc4b95b37dee\") " pod="openshift-marketplace/community-operators-k2vfm" Nov 26 02:47:51 crc kubenswrapper[4766]: I1126 02:47:51.163905 4766 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnr2r\" (UniqueName: \"kubernetes.io/projected/0c90a91f-c037-43fb-844e-bc4b95b37dee-kube-api-access-mnr2r\") pod \"community-operators-k2vfm\" (UID: \"0c90a91f-c037-43fb-844e-bc4b95b37dee\") " pod="openshift-marketplace/community-operators-k2vfm" Nov 26 02:47:51 crc kubenswrapper[4766]: I1126 02:47:51.164271 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c90a91f-c037-43fb-844e-bc4b95b37dee-utilities\") pod \"community-operators-k2vfm\" (UID: \"0c90a91f-c037-43fb-844e-bc4b95b37dee\") " pod="openshift-marketplace/community-operators-k2vfm" Nov 26 02:47:51 crc kubenswrapper[4766]: I1126 02:47:51.164395 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c90a91f-c037-43fb-844e-bc4b95b37dee-catalog-content\") pod \"community-operators-k2vfm\" (UID: \"0c90a91f-c037-43fb-844e-bc4b95b37dee\") " pod="openshift-marketplace/community-operators-k2vfm" Nov 26 02:47:51 crc kubenswrapper[4766]: I1126 02:47:51.187602 4766 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnr2r\" (UniqueName: \"kubernetes.io/projected/0c90a91f-c037-43fb-844e-bc4b95b37dee-kube-api-access-mnr2r\") pod \"community-operators-k2vfm\" (UID: \"0c90a91f-c037-43fb-844e-bc4b95b37dee\") " pod="openshift-marketplace/community-operators-k2vfm" Nov 26 02:47:51 crc kubenswrapper[4766]: I1126 02:47:51.263241 4766 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2vfm" Nov 26 02:47:51 crc kubenswrapper[4766]: I1126 02:47:51.788141 4766 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k2vfm"] Nov 26 02:47:51 crc kubenswrapper[4766]: I1126 02:47:51.827178 4766 scope.go:117] "RemoveContainer" containerID="fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3" Nov 26 02:47:51 crc kubenswrapper[4766]: E1126 02:47:51.827549 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:47:52 crc kubenswrapper[4766]: I1126 02:47:52.616737 4766 generic.go:334] "Generic (PLEG): container finished" podID="0c90a91f-c037-43fb-844e-bc4b95b37dee" containerID="cd926fea5869f1b15e84c2052a675ad2d2cd38d79910f238c75718aa30120a68" exitCode=0 Nov 26 02:47:52 crc kubenswrapper[4766]: I1126 02:47:52.616845 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2vfm" event={"ID":"0c90a91f-c037-43fb-844e-bc4b95b37dee","Type":"ContainerDied","Data":"cd926fea5869f1b15e84c2052a675ad2d2cd38d79910f238c75718aa30120a68"} Nov 26 02:47:52 crc kubenswrapper[4766]: I1126 02:47:52.617076 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2vfm" event={"ID":"0c90a91f-c037-43fb-844e-bc4b95b37dee","Type":"ContainerStarted","Data":"83f306d56c95c9186ae15daa8d7c96f7fd10509c6bd0de9e8afeed82bcd3ddd0"} Nov 26 02:47:53 crc kubenswrapper[4766]: I1126 02:47:53.628287 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2vfm" event={"ID":"0c90a91f-c037-43fb-844e-bc4b95b37dee","Type":"ContainerStarted","Data":"ffceb6ebeba097080ddce9b6e8f266c4509276c0bd8cd398f46278a1fb09065a"} Nov 26 02:47:55 crc kubenswrapper[4766]: I1126 02:47:55.650044 4766 generic.go:334] "Generic (PLEG): container finished" podID="0c90a91f-c037-43fb-844e-bc4b95b37dee" containerID="ffceb6ebeba097080ddce9b6e8f266c4509276c0bd8cd398f46278a1fb09065a" exitCode=0 Nov 26 02:47:55 crc kubenswrapper[4766]: I1126 02:47:55.650162 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2vfm" event={"ID":"0c90a91f-c037-43fb-844e-bc4b95b37dee","Type":"ContainerDied","Data":"ffceb6ebeba097080ddce9b6e8f266c4509276c0bd8cd398f46278a1fb09065a"} Nov 26 02:47:56 crc kubenswrapper[4766]: I1126 02:47:56.664119 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2vfm" event={"ID":"0c90a91f-c037-43fb-844e-bc4b95b37dee","Type":"ContainerStarted","Data":"e6d132daeb59c565eed45e9f82be2700739e00c10cd5046b09945937494727cd"} Nov 26 02:47:56 crc kubenswrapper[4766]: I1126 02:47:56.696188 4766 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k2vfm" podStartSLOduration=3.206066948 podStartE2EDuration="6.696168544s" podCreationTimestamp="2025-11-26 02:47:50 +0000 UTC" firstStartedPulling="2025-11-26 02:47:52.618613026 +0000 UTC m=+8653.467383456" lastFinishedPulling="2025-11-26 02:47:56.108714602 +0000 UTC m=+8656.957485052" observedRunningTime="2025-11-26 02:47:56.684455118 +0000 UTC m=+8657.533225568" watchObservedRunningTime="2025-11-26 02:47:56.696168544 +0000 UTC m=+8657.544938984" Nov 26 02:48:01 crc kubenswrapper[4766]: I1126 02:48:01.263397 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k2vfm" Nov 26 02:48:01 crc kubenswrapper[4766]: I1126 02:48:01.263877 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k2vfm" Nov 26 02:48:01 crc kubenswrapper[4766]: I1126 02:48:01.327373 4766 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k2vfm" Nov 26 02:48:01 crc kubenswrapper[4766]: I1126 02:48:01.802323 4766 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k2vfm" Nov 26 02:48:01 crc kubenswrapper[4766]: I1126 02:48:01.888701 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k2vfm"] Nov 26 02:48:03 crc kubenswrapper[4766]: I1126 02:48:03.752022 4766 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k2vfm" podUID="0c90a91f-c037-43fb-844e-bc4b95b37dee" containerName="registry-server" containerID="cri-o://e6d132daeb59c565eed45e9f82be2700739e00c10cd5046b09945937494727cd" gracePeriod=2 Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.316543 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2vfm" Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.449696 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c90a91f-c037-43fb-844e-bc4b95b37dee-catalog-content\") pod \"0c90a91f-c037-43fb-844e-bc4b95b37dee\" (UID: \"0c90a91f-c037-43fb-844e-bc4b95b37dee\") " Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.449817 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnr2r\" (UniqueName: \"kubernetes.io/projected/0c90a91f-c037-43fb-844e-bc4b95b37dee-kube-api-access-mnr2r\") pod \"0c90a91f-c037-43fb-844e-bc4b95b37dee\" (UID: \"0c90a91f-c037-43fb-844e-bc4b95b37dee\") " Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.450033 4766 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c90a91f-c037-43fb-844e-bc4b95b37dee-utilities\") pod \"0c90a91f-c037-43fb-844e-bc4b95b37dee\" (UID: \"0c90a91f-c037-43fb-844e-bc4b95b37dee\") " Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.450789 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c90a91f-c037-43fb-844e-bc4b95b37dee-utilities" (OuterVolumeSpecName: "utilities") pod "0c90a91f-c037-43fb-844e-bc4b95b37dee" (UID: "0c90a91f-c037-43fb-844e-bc4b95b37dee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.455855 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c90a91f-c037-43fb-844e-bc4b95b37dee-kube-api-access-mnr2r" (OuterVolumeSpecName: "kube-api-access-mnr2r") pod "0c90a91f-c037-43fb-844e-bc4b95b37dee" (UID: "0c90a91f-c037-43fb-844e-bc4b95b37dee"). InnerVolumeSpecName "kube-api-access-mnr2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.511859 4766 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c90a91f-c037-43fb-844e-bc4b95b37dee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c90a91f-c037-43fb-844e-bc4b95b37dee" (UID: "0c90a91f-c037-43fb-844e-bc4b95b37dee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.552310 4766 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c90a91f-c037-43fb-844e-bc4b95b37dee-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.552536 4766 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c90a91f-c037-43fb-844e-bc4b95b37dee-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.552609 4766 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnr2r\" (UniqueName: \"kubernetes.io/projected/0c90a91f-c037-43fb-844e-bc4b95b37dee-kube-api-access-mnr2r\") on node \"crc\" DevicePath \"\"" Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.762670 4766 generic.go:334] "Generic (PLEG): container finished" podID="0c90a91f-c037-43fb-844e-bc4b95b37dee" containerID="e6d132daeb59c565eed45e9f82be2700739e00c10cd5046b09945937494727cd" exitCode=0 Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.762710 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2vfm" event={"ID":"0c90a91f-c037-43fb-844e-bc4b95b37dee","Type":"ContainerDied","Data":"e6d132daeb59c565eed45e9f82be2700739e00c10cd5046b09945937494727cd"} Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.762734 4766 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2vfm" event={"ID":"0c90a91f-c037-43fb-844e-bc4b95b37dee","Type":"ContainerDied","Data":"83f306d56c95c9186ae15daa8d7c96f7fd10509c6bd0de9e8afeed82bcd3ddd0"} Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.762736 4766 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2vfm" Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.762750 4766 scope.go:117] "RemoveContainer" containerID="e6d132daeb59c565eed45e9f82be2700739e00c10cd5046b09945937494727cd" Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.791832 4766 scope.go:117] "RemoveContainer" containerID="ffceb6ebeba097080ddce9b6e8f266c4509276c0bd8cd398f46278a1fb09065a" Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.812142 4766 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k2vfm"] Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.833899 4766 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k2vfm"] Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.845216 4766 scope.go:117] "RemoveContainer" containerID="cd926fea5869f1b15e84c2052a675ad2d2cd38d79910f238c75718aa30120a68" Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.902233 4766 scope.go:117] "RemoveContainer" containerID="e6d132daeb59c565eed45e9f82be2700739e00c10cd5046b09945937494727cd" Nov 26 02:48:04 crc kubenswrapper[4766]: E1126 02:48:04.902630 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6d132daeb59c565eed45e9f82be2700739e00c10cd5046b09945937494727cd\": container with ID starting with e6d132daeb59c565eed45e9f82be2700739e00c10cd5046b09945937494727cd not found: ID does not exist" containerID="e6d132daeb59c565eed45e9f82be2700739e00c10cd5046b09945937494727cd" Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.902693 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6d132daeb59c565eed45e9f82be2700739e00c10cd5046b09945937494727cd"} err="failed to get container status \"e6d132daeb59c565eed45e9f82be2700739e00c10cd5046b09945937494727cd\": rpc error: code = NotFound desc = could not find container \"e6d132daeb59c565eed45e9f82be2700739e00c10cd5046b09945937494727cd\": container with ID starting with e6d132daeb59c565eed45e9f82be2700739e00c10cd5046b09945937494727cd not found: ID does not exist" Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.902724 4766 scope.go:117] "RemoveContainer" containerID="ffceb6ebeba097080ddce9b6e8f266c4509276c0bd8cd398f46278a1fb09065a" Nov 26 02:48:04 crc kubenswrapper[4766]: E1126 02:48:04.903034 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffceb6ebeba097080ddce9b6e8f266c4509276c0bd8cd398f46278a1fb09065a\": container with ID starting with ffceb6ebeba097080ddce9b6e8f266c4509276c0bd8cd398f46278a1fb09065a not found: ID does not exist" containerID="ffceb6ebeba097080ddce9b6e8f266c4509276c0bd8cd398f46278a1fb09065a" Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.903091 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffceb6ebeba097080ddce9b6e8f266c4509276c0bd8cd398f46278a1fb09065a"} err="failed to get container status \"ffceb6ebeba097080ddce9b6e8f266c4509276c0bd8cd398f46278a1fb09065a\": rpc error: code = NotFound desc = could not find container \"ffceb6ebeba097080ddce9b6e8f266c4509276c0bd8cd398f46278a1fb09065a\": container with ID starting with ffceb6ebeba097080ddce9b6e8f266c4509276c0bd8cd398f46278a1fb09065a not found: ID does not exist" Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.903118 4766 scope.go:117] "RemoveContainer" containerID="cd926fea5869f1b15e84c2052a675ad2d2cd38d79910f238c75718aa30120a68" Nov 26 02:48:04 crc kubenswrapper[4766]: E1126 02:48:04.903400 4766 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd926fea5869f1b15e84c2052a675ad2d2cd38d79910f238c75718aa30120a68\": container with ID starting with cd926fea5869f1b15e84c2052a675ad2d2cd38d79910f238c75718aa30120a68 not found: ID does not exist" containerID="cd926fea5869f1b15e84c2052a675ad2d2cd38d79910f238c75718aa30120a68" Nov 26 02:48:04 crc kubenswrapper[4766]: I1126 02:48:04.903429 4766 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd926fea5869f1b15e84c2052a675ad2d2cd38d79910f238c75718aa30120a68"} err="failed to get container status \"cd926fea5869f1b15e84c2052a675ad2d2cd38d79910f238c75718aa30120a68\": rpc error: code = NotFound desc = could not find container \"cd926fea5869f1b15e84c2052a675ad2d2cd38d79910f238c75718aa30120a68\": container with ID starting with cd926fea5869f1b15e84c2052a675ad2d2cd38d79910f238c75718aa30120a68 not found: ID does not exist" Nov 26 02:48:05 crc kubenswrapper[4766]: I1126 02:48:05.826842 4766 scope.go:117] "RemoveContainer" containerID="fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3" Nov 26 02:48:05 crc kubenswrapper[4766]: E1126 02:48:05.827562 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:48:05 crc kubenswrapper[4766]: I1126 02:48:05.839384 4766 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c90a91f-c037-43fb-844e-bc4b95b37dee" path="/var/lib/kubelet/pods/0c90a91f-c037-43fb-844e-bc4b95b37dee/volumes" Nov 26 02:48:19 crc kubenswrapper[4766]: I1126 02:48:19.850039 4766 scope.go:117] "RemoveContainer" containerID="fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3" Nov 26 02:48:19 crc kubenswrapper[4766]: E1126 02:48:19.851265 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:48:32 crc kubenswrapper[4766]: I1126 02:48:32.827860 4766 scope.go:117] "RemoveContainer" containerID="fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3" Nov 26 02:48:32 crc kubenswrapper[4766]: E1126 02:48:32.828829 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:48:44 crc kubenswrapper[4766]: I1126 02:48:44.828081 4766 scope.go:117] "RemoveContainer" containerID="fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3" Nov 26 02:48:44 crc kubenswrapper[4766]: E1126 02:48:44.829399 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" Nov 26 02:48:55 crc kubenswrapper[4766]: I1126 02:48:55.828416 4766 scope.go:117] "RemoveContainer" containerID="fd686bb1f8de67cbdac4be2a8788df0188da2c5fe3043c4a859cd77feb8b3af3" Nov 26 02:48:55 crc kubenswrapper[4766]: E1126 02:48:55.829538 4766 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wf9c2_openshift-machine-config-operator(63471884-61f7-41d5-8967-e3f81eba90d9)\"" pod="openshift-machine-config-operator/machine-config-daemon-wf9c2" podUID="63471884-61f7-41d5-8967-e3f81eba90d9" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111465251024446 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111465252017364 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111443701016502 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111443701015452 5ustar corecore